Add GGUF internal file structure
Browse files- scores/Watt-Tool-8B-F16.md +978 -0
- scores/Watt-Tool-8B-IQ3_M.md +982 -0
- scores/Watt-Tool-8B-IQ3_S.md +982 -0
- scores/Watt-Tool-8B-IQ4_NL.md +982 -0
- scores/Watt-Tool-8B-Q3_K_L.md +982 -0
- scores/Watt-Tool-8B-Q3_K_M.md +982 -0
- scores/Watt-Tool-8B-Q3_K_S.md +982 -0
- scores/Watt-Tool-8B-Q4_K_M.md +982 -0
- scores/Watt-Tool-8B-Q4_K_S.md +982 -0
- scores/Watt-Tool-8B-Q5_K_M.md +982 -0
- scores/Watt-Tool-8B-Q5_K_S.md +982 -0
- scores/Watt-Tool-8B-Q6_K.md +982 -0
- scores/Watt-Tool-8B-Q8_0.md +982 -0
scores/Watt-Tool-8B-F16.md
ADDED
|
@@ -0,0 +1,978 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Watt-Tool-8B-F16.gguf - GGUF Internal File Dump
|
| 2 |
+
|
| 3 |
+
- Endian: LITTLE endian
|
| 4 |
+
|
| 5 |
+
## Key Value Metadata Store
|
| 6 |
+
|
| 7 |
+
There are 39 key-value pairs in this file
|
| 8 |
+
|
| 9 |
+
| POS | TYPE | Count | Key | Value |
|
| 10 |
+
|----:|:---------|-------:|:---------------------------------------|:--------------------------------------------------------------------|
|
| 11 |
+
| 1 | UINT32 | 1 | GGUF.version | 3 |
|
| 12 |
+
| 2 | UINT64 | 1 | GGUF.tensor_count | 292 |
|
| 13 |
+
| 3 | UINT64 | 1 | GGUF.kv_count | 36 |
|
| 14 |
+
| 4 | STRING | 1 | general.architecture | `llama` |
|
| 15 |
+
| 5 | STRING | 1 | general.type | `model` |
|
| 16 |
+
| 6 | STRING | 1 | general.name | `Watt Tool 8B GGUF` |
|
| 17 |
+
| 7 | STRING | 1 | general.finetune | `GGUF` |
|
| 18 |
+
| 8 | STRING | 1 | general.basename | `Watt-Tool` |
|
| 19 |
+
| 9 | STRING | 1 | general.size_label | `8B` |
|
| 20 |
+
| 10 | STRING | 1 | general.license | `apache-2.0` |
|
| 21 |
+
| 11 | UINT32 | 1 | general.base_model.count | 1 |
|
| 22 |
+
| 12 | STRING | 1 | general.base_model.0.name | `Llama 3.1 8B Instruct` |
|
| 23 |
+
| 13 | STRING | 1 | general.base_model.0.organization | `Meta Llama` |
|
| 24 |
+
| 14 | STRING | 1 | general.base_model.0.repo_url | `https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct` |
|
| 25 |
+
| 15 | [STRING] | 4 | general.tags | [ `function-calling`, `tool-use`, `llama`, `bfcl` ] |
|
| 26 |
+
| 16 | [STRING] | 1 | general.languages | [ `en` ] |
|
| 27 |
+
| 17 | UINT32 | 1 | llama.block_count | 32 |
|
| 28 |
+
| 18 | UINT32 | 1 | llama.context_length | 131072 |
|
| 29 |
+
| 19 | UINT32 | 1 | llama.embedding_length | 4096 |
|
| 30 |
+
| 20 | UINT32 | 1 | llama.feed_forward_length | 14336 |
|
| 31 |
+
| 21 | UINT32 | 1 | llama.attention.head_count | 32 |
|
| 32 |
+
| 22 | UINT32 | 1 | llama.attention.head_count_kv | 8 |
|
| 33 |
+
| 23 | FLOAT32 | 1 | llama.rope.freq_base | 500000.0 |
|
| 34 |
+
| 24 | FLOAT32 | 1 | llama.attention.layer_norm_rms_epsilon | 1e-05 |
|
| 35 |
+
| 25 | UINT32 | 1 | llama.attention.key_length | 128 |
|
| 36 |
+
| 26 | UINT32 | 1 | llama.attention.value_length | 128 |
|
| 37 |
+
| 27 | UINT32 | 1 | general.file_type | 1 |
|
| 38 |
+
| 28 | UINT32 | 1 | llama.vocab_size | 128256 |
|
| 39 |
+
| 29 | UINT32 | 1 | llama.rope.dimension_count | 128 |
|
| 40 |
+
| 30 | STRING | 1 | tokenizer.ggml.model | `gpt2` |
|
| 41 |
+
| 31 | STRING | 1 | tokenizer.ggml.pre | `llama-bpe` |
|
| 42 |
+
| 32 | [STRING] | 128256 | tokenizer.ggml.tokens | [ `!`, `"`, `#`, `$`, `%`, ... ] |
|
| 43 |
+
| 33 | [INT32] | 128256 | tokenizer.ggml.token_type | [ 1, 1, 1, 1, 1, 1, 1, ... ] |
|
| 44 |
+
| 34 | [STRING] | 280147 | tokenizer.ggml.merges | [ `Ġ Ġ`, `Ġ ĠĠĠ`, `ĠĠ ĠĠ`, `ĠĠĠ Ġ`, `i n`, ... ] |
|
| 45 |
+
| 35 | UINT32 | 1 | tokenizer.ggml.bos_token_id | 128000 |
|
| 46 |
+
| 36 | UINT32 | 1 | tokenizer.ggml.eos_token_id | 128009 |
|
| 47 |
+
| 37 | UINT32 | 1 | tokenizer.ggml.padding_token_id | 128009 |
|
| 48 |
+
| 38 | STRING | 1 | tokenizer.chat_template | `{{ '<|begin_of_text|>' }}{% if`...`d|>' }}{% endif %}{% endfor %}` |
|
| 49 |
+
| 39 | UINT32 | 1 | general.quantization_version | 2 |
|
| 50 |
+
|
| 51 |
+
## Tensors Overview ~8B Elements
|
| 52 |
+
|
| 53 |
+
Total number of elements in all tensors: 8030261312 Elements
|
| 54 |
+
|
| 55 |
+
- [Watt-Tool-8B-F16.gguf - GGUF Internal File Dump](#watt-tool-8b-f16gguf---gguf-internal-file-dump)
|
| 56 |
+
- [Key Value Metadata Store](#key-value-metadata-store)
|
| 57 |
+
- [Tensors Overview ~8B Elements](#tensors-overview-8b-elements)
|
| 58 |
+
- [Tensor Data Offset](#tensor-data-offset)
|
| 59 |
+
- [Base Tensor Group : ~1B Elements](#base-tensor-group--1b-elements)
|
| 60 |
+
- [Block 0 Tensor Group : ~218M Elements](#block-0-tensor-group--218m-elements)
|
| 61 |
+
- [Block 1 Tensor Group : ~218M Elements](#block-1-tensor-group--218m-elements)
|
| 62 |
+
- [Block 10 Tensor Group : ~218M Elements](#block-10-tensor-group--218m-elements)
|
| 63 |
+
- [Block 11 Tensor Group : ~218M Elements](#block-11-tensor-group--218m-elements)
|
| 64 |
+
- [Block 12 Tensor Group : ~218M Elements](#block-12-tensor-group--218m-elements)
|
| 65 |
+
- [Block 13 Tensor Group : ~218M Elements](#block-13-tensor-group--218m-elements)
|
| 66 |
+
- [Block 14 Tensor Group : ~218M Elements](#block-14-tensor-group--218m-elements)
|
| 67 |
+
- [Block 15 Tensor Group : ~218M Elements](#block-15-tensor-group--218m-elements)
|
| 68 |
+
- [Block 16 Tensor Group : ~218M Elements](#block-16-tensor-group--218m-elements)
|
| 69 |
+
- [Block 17 Tensor Group : ~218M Elements](#block-17-tensor-group--218m-elements)
|
| 70 |
+
- [Block 18 Tensor Group : ~218M Elements](#block-18-tensor-group--218m-elements)
|
| 71 |
+
- [Block 19 Tensor Group : ~218M Elements](#block-19-tensor-group--218m-elements)
|
| 72 |
+
- [Block 2 Tensor Group : ~218M Elements](#block-2-tensor-group--218m-elements)
|
| 73 |
+
- [Block 20 Tensor Group : ~218M Elements](#block-20-tensor-group--218m-elements)
|
| 74 |
+
- [Block 21 Tensor Group : ~218M Elements](#block-21-tensor-group--218m-elements)
|
| 75 |
+
- [Block 22 Tensor Group : ~218M Elements](#block-22-tensor-group--218m-elements)
|
| 76 |
+
- [Block 23 Tensor Group : ~218M Elements](#block-23-tensor-group--218m-elements)
|
| 77 |
+
- [Block 24 Tensor Group : ~218M Elements](#block-24-tensor-group--218m-elements)
|
| 78 |
+
- [Block 25 Tensor Group : ~218M Elements](#block-25-tensor-group--218m-elements)
|
| 79 |
+
- [Block 26 Tensor Group : ~218M Elements](#block-26-tensor-group--218m-elements)
|
| 80 |
+
- [Block 27 Tensor Group : ~218M Elements](#block-27-tensor-group--218m-elements)
|
| 81 |
+
- [Block 28 Tensor Group : ~218M Elements](#block-28-tensor-group--218m-elements)
|
| 82 |
+
- [Block 29 Tensor Group : ~218M Elements](#block-29-tensor-group--218m-elements)
|
| 83 |
+
- [Block 3 Tensor Group : ~218M Elements](#block-3-tensor-group--218m-elements)
|
| 84 |
+
- [Block 30 Tensor Group : ~218M Elements](#block-30-tensor-group--218m-elements)
|
| 85 |
+
- [Block 31 Tensor Group : ~218M Elements](#block-31-tensor-group--218m-elements)
|
| 86 |
+
- [Block 4 Tensor Group : ~218M Elements](#block-4-tensor-group--218m-elements)
|
| 87 |
+
- [Block 5 Tensor Group : ~218M Elements](#block-5-tensor-group--218m-elements)
|
| 88 |
+
- [Block 6 Tensor Group : ~218M Elements](#block-6-tensor-group--218m-elements)
|
| 89 |
+
- [Block 7 Tensor Group : ~218M Elements](#block-7-tensor-group--218m-elements)
|
| 90 |
+
- [Block 8 Tensor Group : ~218M Elements](#block-8-tensor-group--218m-elements)
|
| 91 |
+
- [Block 9 Tensor Group : ~218M Elements](#block-9-tensor-group--218m-elements)
|
| 92 |
+
|
| 93 |
+
### Tensor Data Offset
|
| 94 |
+
|
| 95 |
+
This table contains the offset and data segment relative to start of file
|
| 96 |
+
|
| 97 |
+
| T_ID | Tensor Layer Name | Data Offset (B) | Data Size (B) |
|
| 98 |
+
|-----:|:--------------------------|-----------------:|-----------------:|
|
| 99 |
+
| 0 | rope_freqs.weight | 0x779520 | 0x100 |
|
| 100 |
+
| 1 | output.weight | 0x779620 | 0x3ea00000 |
|
| 101 |
+
| 2 | token_embd.weight | 0x3f179620 | 0x3ea00000 |
|
| 102 |
+
| 3 | blk.0.attn_norm.weight | 0x7db79620 | 0x4000 |
|
| 103 |
+
| 4 | blk.0.ffn_down.weight | 0x7db7d620 | 0x7000000 |
|
| 104 |
+
| 5 | blk.0.ffn_gate.weight | 0x84b7d620 | 0x7000000 |
|
| 105 |
+
| 6 | blk.0.ffn_up.weight | 0x8bb7d620 | 0x7000000 |
|
| 106 |
+
| 7 | blk.0.ffn_norm.weight | 0x92b7d620 | 0x4000 |
|
| 107 |
+
| 8 | blk.0.attn_k.weight | 0x92b81620 | 0x800000 |
|
| 108 |
+
| 9 | blk.0.attn_output.weight | 0x93381620 | 0x2000000 |
|
| 109 |
+
| 10 | blk.0.attn_q.weight | 0x95381620 | 0x2000000 |
|
| 110 |
+
| 11 | blk.0.attn_v.weight | 0x97381620 | 0x800000 |
|
| 111 |
+
| 12 | blk.1.attn_norm.weight | 0x97b81620 | 0x4000 |
|
| 112 |
+
| 13 | blk.1.ffn_down.weight | 0x97b85620 | 0x7000000 |
|
| 113 |
+
| 14 | blk.1.ffn_gate.weight | 0x9eb85620 | 0x7000000 |
|
| 114 |
+
| 15 | blk.1.ffn_up.weight | 0xa5b85620 | 0x7000000 |
|
| 115 |
+
| 16 | blk.1.ffn_norm.weight | 0xacb85620 | 0x4000 |
|
| 116 |
+
| 17 | blk.1.attn_k.weight | 0xacb89620 | 0x800000 |
|
| 117 |
+
| 18 | blk.1.attn_output.weight | 0xad389620 | 0x2000000 |
|
| 118 |
+
| 19 | blk.1.attn_q.weight | 0xaf389620 | 0x2000000 |
|
| 119 |
+
| 20 | blk.1.attn_v.weight | 0xb1389620 | 0x800000 |
|
| 120 |
+
| 21 | blk.10.attn_norm.weight | 0xb1b89620 | 0x4000 |
|
| 121 |
+
| 22 | blk.10.ffn_down.weight | 0xb1b8d620 | 0x7000000 |
|
| 122 |
+
| 23 | blk.10.ffn_gate.weight | 0xb8b8d620 | 0x7000000 |
|
| 123 |
+
| 24 | blk.10.ffn_up.weight | 0xbfb8d620 | 0x7000000 |
|
| 124 |
+
| 25 | blk.10.ffn_norm.weight | 0xc6b8d620 | 0x4000 |
|
| 125 |
+
| 26 | blk.10.attn_k.weight | 0xc6b91620 | 0x800000 |
|
| 126 |
+
| 27 | blk.10.attn_output.weight | 0xc7391620 | 0x2000000 |
|
| 127 |
+
| 28 | blk.10.attn_q.weight | 0xc9391620 | 0x2000000 |
|
| 128 |
+
| 29 | blk.10.attn_v.weight | 0xcb391620 | 0x800000 |
|
| 129 |
+
| 30 | blk.11.attn_norm.weight | 0xcbb91620 | 0x4000 |
|
| 130 |
+
| 31 | blk.11.ffn_down.weight | 0xcbb95620 | 0x7000000 |
|
| 131 |
+
| 32 | blk.11.ffn_gate.weight | 0xd2b95620 | 0x7000000 |
|
| 132 |
+
| 33 | blk.11.ffn_up.weight | 0xd9b95620 | 0x7000000 |
|
| 133 |
+
| 34 | blk.11.ffn_norm.weight | 0xe0b95620 | 0x4000 |
|
| 134 |
+
| 35 | blk.11.attn_k.weight | 0xe0b99620 | 0x800000 |
|
| 135 |
+
| 36 | blk.11.attn_output.weight | 0xe1399620 | 0x2000000 |
|
| 136 |
+
| 37 | blk.11.attn_q.weight | 0xe3399620 | 0x2000000 |
|
| 137 |
+
| 38 | blk.11.attn_v.weight | 0xe5399620 | 0x800000 |
|
| 138 |
+
| 39 | blk.12.attn_norm.weight | 0xe5b99620 | 0x4000 |
|
| 139 |
+
| 40 | blk.12.ffn_down.weight | 0xe5b9d620 | 0x7000000 |
|
| 140 |
+
| 41 | blk.12.ffn_gate.weight | 0xecb9d620 | 0x7000000 |
|
| 141 |
+
| 42 | blk.12.ffn_up.weight | 0xf3b9d620 | 0x7000000 |
|
| 142 |
+
| 43 | blk.12.ffn_norm.weight | 0xfab9d620 | 0x4000 |
|
| 143 |
+
| 44 | blk.12.attn_k.weight | 0xfaba1620 | 0x800000 |
|
| 144 |
+
| 45 | blk.12.attn_output.weight | 0xfb3a1620 | 0x2000000 |
|
| 145 |
+
| 46 | blk.12.attn_q.weight | 0xfd3a1620 | 0x2000000 |
|
| 146 |
+
| 47 | blk.12.attn_v.weight | 0xff3a1620 | 0x800000 |
|
| 147 |
+
| 48 | blk.13.attn_norm.weight | 0xffba1620 | 0x4000 |
|
| 148 |
+
| 49 | blk.13.ffn_down.weight | 0xffba5620 | 0x7000000 |
|
| 149 |
+
| 50 | blk.13.ffn_gate.weight | 0x106ba5620 | 0x7000000 |
|
| 150 |
+
| 51 | blk.13.ffn_up.weight | 0x10dba5620 | 0x7000000 |
|
| 151 |
+
| 52 | blk.13.ffn_norm.weight | 0x114ba5620 | 0x4000 |
|
| 152 |
+
| 53 | blk.13.attn_k.weight | 0x114ba9620 | 0x800000 |
|
| 153 |
+
| 54 | blk.13.attn_output.weight | 0x1153a9620 | 0x2000000 |
|
| 154 |
+
| 55 | blk.13.attn_q.weight | 0x1173a9620 | 0x2000000 |
|
| 155 |
+
| 56 | blk.13.attn_v.weight | 0x1193a9620 | 0x800000 |
|
| 156 |
+
| 57 | blk.14.attn_norm.weight | 0x119ba9620 | 0x4000 |
|
| 157 |
+
| 58 | blk.14.ffn_down.weight | 0x119bad620 | 0x7000000 |
|
| 158 |
+
| 59 | blk.14.ffn_gate.weight | 0x120bad620 | 0x7000000 |
|
| 159 |
+
| 60 | blk.14.ffn_up.weight | 0x127bad620 | 0x7000000 |
|
| 160 |
+
| 61 | blk.14.ffn_norm.weight | 0x12ebad620 | 0x4000 |
|
| 161 |
+
| 62 | blk.14.attn_k.weight | 0x12ebb1620 | 0x800000 |
|
| 162 |
+
| 63 | blk.14.attn_output.weight | 0x12f3b1620 | 0x2000000 |
|
| 163 |
+
| 64 | blk.14.attn_q.weight | 0x1313b1620 | 0x2000000 |
|
| 164 |
+
| 65 | blk.14.attn_v.weight | 0x1333b1620 | 0x800000 |
|
| 165 |
+
| 66 | blk.15.attn_norm.weight | 0x133bb1620 | 0x4000 |
|
| 166 |
+
| 67 | blk.15.ffn_down.weight | 0x133bb5620 | 0x7000000 |
|
| 167 |
+
| 68 | blk.15.ffn_gate.weight | 0x13abb5620 | 0x7000000 |
|
| 168 |
+
| 69 | blk.15.ffn_up.weight | 0x141bb5620 | 0x7000000 |
|
| 169 |
+
| 70 | blk.15.ffn_norm.weight | 0x148bb5620 | 0x4000 |
|
| 170 |
+
| 71 | blk.15.attn_k.weight | 0x148bb9620 | 0x800000 |
|
| 171 |
+
| 72 | blk.15.attn_output.weight | 0x1493b9620 | 0x2000000 |
|
| 172 |
+
| 73 | blk.15.attn_q.weight | 0x14b3b9620 | 0x2000000 |
|
| 173 |
+
| 74 | blk.15.attn_v.weight | 0x14d3b9620 | 0x800000 |
|
| 174 |
+
| 75 | blk.16.attn_norm.weight | 0x14dbb9620 | 0x4000 |
|
| 175 |
+
| 76 | blk.16.ffn_down.weight | 0x14dbbd620 | 0x7000000 |
|
| 176 |
+
| 77 | blk.16.ffn_gate.weight | 0x154bbd620 | 0x7000000 |
|
| 177 |
+
| 78 | blk.16.ffn_up.weight | 0x15bbbd620 | 0x7000000 |
|
| 178 |
+
| 79 | blk.16.ffn_norm.weight | 0x162bbd620 | 0x4000 |
|
| 179 |
+
| 80 | blk.16.attn_k.weight | 0x162bc1620 | 0x800000 |
|
| 180 |
+
| 81 | blk.16.attn_output.weight | 0x1633c1620 | 0x2000000 |
|
| 181 |
+
| 82 | blk.16.attn_q.weight | 0x1653c1620 | 0x2000000 |
|
| 182 |
+
| 83 | blk.16.attn_v.weight | 0x1673c1620 | 0x800000 |
|
| 183 |
+
| 84 | blk.17.attn_norm.weight | 0x167bc1620 | 0x4000 |
|
| 184 |
+
| 85 | blk.17.ffn_down.weight | 0x167bc5620 | 0x7000000 |
|
| 185 |
+
| 86 | blk.17.ffn_gate.weight | 0x16ebc5620 | 0x7000000 |
|
| 186 |
+
| 87 | blk.17.ffn_up.weight | 0x175bc5620 | 0x7000000 |
|
| 187 |
+
| 88 | blk.17.ffn_norm.weight | 0x17cbc5620 | 0x4000 |
|
| 188 |
+
| 89 | blk.17.attn_k.weight | 0x17cbc9620 | 0x800000 |
|
| 189 |
+
| 90 | blk.17.attn_output.weight | 0x17d3c9620 | 0x2000000 |
|
| 190 |
+
| 91 | blk.17.attn_q.weight | 0x17f3c9620 | 0x2000000 |
|
| 191 |
+
| 92 | blk.17.attn_v.weight | 0x1813c9620 | 0x800000 |
|
| 192 |
+
| 93 | blk.18.attn_norm.weight | 0x181bc9620 | 0x4000 |
|
| 193 |
+
| 94 | blk.18.ffn_down.weight | 0x181bcd620 | 0x7000000 |
|
| 194 |
+
| 95 | blk.18.ffn_gate.weight | 0x188bcd620 | 0x7000000 |
|
| 195 |
+
| 96 | blk.18.ffn_up.weight | 0x18fbcd620 | 0x7000000 |
|
| 196 |
+
| 97 | blk.18.ffn_norm.weight | 0x196bcd620 | 0x4000 |
|
| 197 |
+
| 98 | blk.18.attn_k.weight | 0x196bd1620 | 0x800000 |
|
| 198 |
+
| 99 | blk.18.attn_output.weight | 0x1973d1620 | 0x2000000 |
|
| 199 |
+
| 100 | blk.18.attn_q.weight | 0x1993d1620 | 0x2000000 |
|
| 200 |
+
| 101 | blk.18.attn_v.weight | 0x19b3d1620 | 0x800000 |
|
| 201 |
+
| 102 | blk.19.attn_norm.weight | 0x19bbd1620 | 0x4000 |
|
| 202 |
+
| 103 | blk.19.ffn_down.weight | 0x19bbd5620 | 0x7000000 |
|
| 203 |
+
| 104 | blk.19.ffn_gate.weight | 0x1a2bd5620 | 0x7000000 |
|
| 204 |
+
| 105 | blk.19.ffn_up.weight | 0x1a9bd5620 | 0x7000000 |
|
| 205 |
+
| 106 | blk.19.ffn_norm.weight | 0x1b0bd5620 | 0x4000 |
|
| 206 |
+
| 107 | blk.19.attn_k.weight | 0x1b0bd9620 | 0x800000 |
|
| 207 |
+
| 108 | blk.19.attn_output.weight | 0x1b13d9620 | 0x2000000 |
|
| 208 |
+
| 109 | blk.19.attn_q.weight | 0x1b33d9620 | 0x2000000 |
|
| 209 |
+
| 110 | blk.19.attn_v.weight | 0x1b53d9620 | 0x800000 |
|
| 210 |
+
| 111 | blk.2.attn_norm.weight | 0x1b5bd9620 | 0x4000 |
|
| 211 |
+
| 112 | blk.2.ffn_down.weight | 0x1b5bdd620 | 0x7000000 |
|
| 212 |
+
| 113 | blk.2.ffn_gate.weight | 0x1bcbdd620 | 0x7000000 |
|
| 213 |
+
| 114 | blk.2.ffn_up.weight | 0x1c3bdd620 | 0x7000000 |
|
| 214 |
+
| 115 | blk.2.ffn_norm.weight | 0x1cabdd620 | 0x4000 |
|
| 215 |
+
| 116 | blk.2.attn_k.weight | 0x1cabe1620 | 0x800000 |
|
| 216 |
+
| 117 | blk.2.attn_output.weight | 0x1cb3e1620 | 0x2000000 |
|
| 217 |
+
| 118 | blk.2.attn_q.weight | 0x1cd3e1620 | 0x2000000 |
|
| 218 |
+
| 119 | blk.2.attn_v.weight | 0x1cf3e1620 | 0x800000 |
|
| 219 |
+
| 120 | blk.20.attn_norm.weight | 0x1cfbe1620 | 0x4000 |
|
| 220 |
+
| 121 | blk.20.ffn_down.weight | 0x1cfbe5620 | 0x7000000 |
|
| 221 |
+
| 122 | blk.20.ffn_gate.weight | 0x1d6be5620 | 0x7000000 |
|
| 222 |
+
| 123 | blk.20.ffn_up.weight | 0x1ddbe5620 | 0x7000000 |
|
| 223 |
+
| 124 | blk.20.ffn_norm.weight | 0x1e4be5620 | 0x4000 |
|
| 224 |
+
| 125 | blk.20.attn_k.weight | 0x1e4be9620 | 0x800000 |
|
| 225 |
+
| 126 | blk.20.attn_output.weight | 0x1e53e9620 | 0x2000000 |
|
| 226 |
+
| 127 | blk.20.attn_q.weight | 0x1e73e9620 | 0x2000000 |
|
| 227 |
+
| 128 | blk.20.attn_v.weight | 0x1e93e9620 | 0x800000 |
|
| 228 |
+
| 129 | blk.21.attn_norm.weight | 0x1e9be9620 | 0x4000 |
|
| 229 |
+
| 130 | blk.21.ffn_down.weight | 0x1e9bed620 | 0x7000000 |
|
| 230 |
+
| 131 | blk.21.ffn_gate.weight | 0x1f0bed620 | 0x7000000 |
|
| 231 |
+
| 132 | blk.21.ffn_up.weight | 0x1f7bed620 | 0x7000000 |
|
| 232 |
+
| 133 | blk.21.ffn_norm.weight | 0x1febed620 | 0x4000 |
|
| 233 |
+
| 134 | blk.21.attn_k.weight | 0x1febf1620 | 0x800000 |
|
| 234 |
+
| 135 | blk.21.attn_output.weight | 0x1ff3f1620 | 0x2000000 |
|
| 235 |
+
| 136 | blk.21.attn_q.weight | 0x2013f1620 | 0x2000000 |
|
| 236 |
+
| 137 | blk.21.attn_v.weight | 0x2033f1620 | 0x800000 |
|
| 237 |
+
| 138 | blk.22.attn_norm.weight | 0x203bf1620 | 0x4000 |
|
| 238 |
+
| 139 | blk.22.ffn_down.weight | 0x203bf5620 | 0x7000000 |
|
| 239 |
+
| 140 | blk.22.ffn_gate.weight | 0x20abf5620 | 0x7000000 |
|
| 240 |
+
| 141 | blk.22.ffn_up.weight | 0x211bf5620 | 0x7000000 |
|
| 241 |
+
| 142 | blk.22.ffn_norm.weight | 0x218bf5620 | 0x4000 |
|
| 242 |
+
| 143 | blk.22.attn_k.weight | 0x218bf9620 | 0x800000 |
|
| 243 |
+
| 144 | blk.22.attn_output.weight | 0x2193f9620 | 0x2000000 |
|
| 244 |
+
| 145 | blk.22.attn_q.weight | 0x21b3f9620 | 0x2000000 |
|
| 245 |
+
| 146 | blk.22.attn_v.weight | 0x21d3f9620 | 0x800000 |
|
| 246 |
+
| 147 | blk.23.attn_norm.weight | 0x21dbf9620 | 0x4000 |
|
| 247 |
+
| 148 | blk.23.ffn_down.weight | 0x21dbfd620 | 0x7000000 |
|
| 248 |
+
| 149 | blk.23.ffn_gate.weight | 0x224bfd620 | 0x7000000 |
|
| 249 |
+
| 150 | blk.23.ffn_up.weight | 0x22bbfd620 | 0x7000000 |
|
| 250 |
+
| 151 | blk.23.ffn_norm.weight | 0x232bfd620 | 0x4000 |
|
| 251 |
+
| 152 | blk.23.attn_k.weight | 0x232c01620 | 0x800000 |
|
| 252 |
+
| 153 | blk.23.attn_output.weight | 0x233401620 | 0x2000000 |
|
| 253 |
+
| 154 | blk.23.attn_q.weight | 0x235401620 | 0x2000000 |
|
| 254 |
+
| 155 | blk.23.attn_v.weight | 0x237401620 | 0x800000 |
|
| 255 |
+
| 156 | blk.24.attn_norm.weight | 0x237c01620 | 0x4000 |
|
| 256 |
+
| 157 | blk.24.ffn_down.weight | 0x237c05620 | 0x7000000 |
|
| 257 |
+
| 158 | blk.24.ffn_gate.weight | 0x23ec05620 | 0x7000000 |
|
| 258 |
+
| 159 | blk.24.ffn_up.weight | 0x245c05620 | 0x7000000 |
|
| 259 |
+
| 160 | blk.24.ffn_norm.weight | 0x24cc05620 | 0x4000 |
|
| 260 |
+
| 161 | blk.24.attn_k.weight | 0x24cc09620 | 0x800000 |
|
| 261 |
+
| 162 | blk.24.attn_output.weight | 0x24d409620 | 0x2000000 |
|
| 262 |
+
| 163 | blk.24.attn_q.weight | 0x24f409620 | 0x2000000 |
|
| 263 |
+
| 164 | blk.24.attn_v.weight | 0x251409620 | 0x800000 |
|
| 264 |
+
| 165 | blk.25.attn_norm.weight | 0x251c09620 | 0x4000 |
|
| 265 |
+
| 166 | blk.25.ffn_down.weight | 0x251c0d620 | 0x7000000 |
|
| 266 |
+
| 167 | blk.25.ffn_gate.weight | 0x258c0d620 | 0x7000000 |
|
| 267 |
+
| 168 | blk.25.ffn_up.weight | 0x25fc0d620 | 0x7000000 |
|
| 268 |
+
| 169 | blk.25.ffn_norm.weight | 0x266c0d620 | 0x4000 |
|
| 269 |
+
| 170 | blk.25.attn_k.weight | 0x266c11620 | 0x800000 |
|
| 270 |
+
| 171 | blk.25.attn_output.weight | 0x267411620 | 0x2000000 |
|
| 271 |
+
| 172 | blk.25.attn_q.weight | 0x269411620 | 0x2000000 |
|
| 272 |
+
| 173 | blk.25.attn_v.weight | 0x26b411620 | 0x800000 |
|
| 273 |
+
| 174 | blk.26.attn_norm.weight | 0x26bc11620 | 0x4000 |
|
| 274 |
+
| 175 | blk.26.ffn_down.weight | 0x26bc15620 | 0x7000000 |
|
| 275 |
+
| 176 | blk.26.ffn_gate.weight | 0x272c15620 | 0x7000000 |
|
| 276 |
+
| 177 | blk.26.ffn_up.weight | 0x279c15620 | 0x7000000 |
|
| 277 |
+
| 178 | blk.26.ffn_norm.weight | 0x280c15620 | 0x4000 |
|
| 278 |
+
| 179 | blk.26.attn_k.weight | 0x280c19620 | 0x800000 |
|
| 279 |
+
| 180 | blk.26.attn_output.weight | 0x281419620 | 0x2000000 |
|
| 280 |
+
| 181 | blk.26.attn_q.weight | 0x283419620 | 0x2000000 |
|
| 281 |
+
| 182 | blk.26.attn_v.weight | 0x285419620 | 0x800000 |
|
| 282 |
+
| 183 | blk.27.attn_norm.weight | 0x285c19620 | 0x4000 |
|
| 283 |
+
| 184 | blk.27.ffn_down.weight | 0x285c1d620 | 0x7000000 |
|
| 284 |
+
| 185 | blk.27.ffn_gate.weight | 0x28cc1d620 | 0x7000000 |
|
| 285 |
+
| 186 | blk.27.ffn_up.weight | 0x293c1d620 | 0x7000000 |
|
| 286 |
+
| 187 | blk.27.ffn_norm.weight | 0x29ac1d620 | 0x4000 |
|
| 287 |
+
| 188 | blk.27.attn_k.weight | 0x29ac21620 | 0x800000 |
|
| 288 |
+
| 189 | blk.27.attn_output.weight | 0x29b421620 | 0x2000000 |
|
| 289 |
+
| 190 | blk.27.attn_q.weight | 0x29d421620 | 0x2000000 |
|
| 290 |
+
| 191 | blk.27.attn_v.weight | 0x29f421620 | 0x800000 |
|
| 291 |
+
| 192 | blk.28.attn_norm.weight | 0x29fc21620 | 0x4000 |
|
| 292 |
+
| 193 | blk.28.ffn_down.weight | 0x29fc25620 | 0x7000000 |
|
| 293 |
+
| 194 | blk.28.ffn_gate.weight | 0x2a6c25620 | 0x7000000 |
|
| 294 |
+
| 195 | blk.28.ffn_up.weight | 0x2adc25620 | 0x7000000 |
|
| 295 |
+
| 196 | blk.28.ffn_norm.weight | 0x2b4c25620 | 0x4000 |
|
| 296 |
+
| 197 | blk.28.attn_k.weight | 0x2b4c29620 | 0x800000 |
|
| 297 |
+
| 198 | blk.28.attn_output.weight | 0x2b5429620 | 0x2000000 |
|
| 298 |
+
| 199 | blk.28.attn_q.weight | 0x2b7429620 | 0x2000000 |
|
| 299 |
+
| 200 | blk.28.attn_v.weight | 0x2b9429620 | 0x800000 |
|
| 300 |
+
| 201 | blk.29.attn_norm.weight | 0x2b9c29620 | 0x4000 |
|
| 301 |
+
| 202 | blk.29.ffn_down.weight | 0x2b9c2d620 | 0x7000000 |
|
| 302 |
+
| 203 | blk.29.ffn_gate.weight | 0x2c0c2d620 | 0x7000000 |
|
| 303 |
+
| 204 | blk.29.ffn_up.weight | 0x2c7c2d620 | 0x7000000 |
|
| 304 |
+
| 205 | blk.29.ffn_norm.weight | 0x2cec2d620 | 0x4000 |
|
| 305 |
+
| 206 | blk.29.attn_k.weight | 0x2cec31620 | 0x800000 |
|
| 306 |
+
| 207 | blk.29.attn_output.weight | 0x2cf431620 | 0x2000000 |
|
| 307 |
+
| 208 | blk.29.attn_q.weight | 0x2d1431620 | 0x2000000 |
|
| 308 |
+
| 209 | blk.29.attn_v.weight | 0x2d3431620 | 0x800000 |
|
| 309 |
+
| 210 | blk.3.attn_norm.weight | 0x2d3c31620 | 0x4000 |
|
| 310 |
+
| 211 | blk.3.ffn_down.weight | 0x2d3c35620 | 0x7000000 |
|
| 311 |
+
| 212 | blk.3.ffn_gate.weight | 0x2dac35620 | 0x7000000 |
|
| 312 |
+
| 213 | blk.3.ffn_up.weight | 0x2e1c35620 | 0x7000000 |
|
| 313 |
+
| 214 | blk.3.ffn_norm.weight | 0x2e8c35620 | 0x4000 |
|
| 314 |
+
| 215 | blk.3.attn_k.weight | 0x2e8c39620 | 0x800000 |
|
| 315 |
+
| 216 | blk.3.attn_output.weight | 0x2e9439620 | 0x2000000 |
|
| 316 |
+
| 217 | blk.3.attn_q.weight | 0x2eb439620 | 0x2000000 |
|
| 317 |
+
| 218 | blk.3.attn_v.weight | 0x2ed439620 | 0x800000 |
|
| 318 |
+
| 219 | blk.30.attn_norm.weight | 0x2edc39620 | 0x4000 |
|
| 319 |
+
| 220 | blk.30.ffn_down.weight | 0x2edc3d620 | 0x7000000 |
|
| 320 |
+
| 221 | blk.30.ffn_gate.weight | 0x2f4c3d620 | 0x7000000 |
|
| 321 |
+
| 222 | blk.30.ffn_up.weight | 0x2fbc3d620 | 0x7000000 |
|
| 322 |
+
| 223 | blk.30.ffn_norm.weight | 0x302c3d620 | 0x4000 |
|
| 323 |
+
| 224 | blk.30.attn_k.weight | 0x302c41620 | 0x800000 |
|
| 324 |
+
| 225 | blk.30.attn_output.weight | 0x303441620 | 0x2000000 |
|
| 325 |
+
| 226 | blk.30.attn_q.weight | 0x305441620 | 0x2000000 |
|
| 326 |
+
| 227 | blk.30.attn_v.weight | 0x307441620 | 0x800000 |
|
| 327 |
+
| 228 | blk.31.attn_norm.weight | 0x307c41620 | 0x4000 |
|
| 328 |
+
| 229 | blk.31.ffn_down.weight | 0x307c45620 | 0x7000000 |
|
| 329 |
+
| 230 | blk.31.ffn_gate.weight | 0x30ec45620 | 0x7000000 |
|
| 330 |
+
| 231 | blk.31.ffn_up.weight | 0x315c45620 | 0x7000000 |
|
| 331 |
+
| 232 | blk.31.ffn_norm.weight | 0x31cc45620 | 0x4000 |
|
| 332 |
+
| 233 | blk.31.attn_k.weight | 0x31cc49620 | 0x800000 |
|
| 333 |
+
| 234 | blk.31.attn_output.weight | 0x31d449620 | 0x2000000 |
|
| 334 |
+
| 235 | blk.31.attn_q.weight | 0x31f449620 | 0x2000000 |
|
| 335 |
+
| 236 | blk.31.attn_v.weight | 0x321449620 | 0x800000 |
|
| 336 |
+
| 237 | blk.4.attn_norm.weight | 0x321c49620 | 0x4000 |
|
| 337 |
+
| 238 | blk.4.ffn_down.weight | 0x321c4d620 | 0x7000000 |
|
| 338 |
+
| 239 | blk.4.ffn_gate.weight | 0x328c4d620 | 0x7000000 |
|
| 339 |
+
| 240 | blk.4.ffn_up.weight | 0x32fc4d620 | 0x7000000 |
|
| 340 |
+
| 241 | blk.4.ffn_norm.weight | 0x336c4d620 | 0x4000 |
|
| 341 |
+
| 242 | blk.4.attn_k.weight | 0x336c51620 | 0x800000 |
|
| 342 |
+
| 243 | blk.4.attn_output.weight | 0x337451620 | 0x2000000 |
|
| 343 |
+
| 244 | blk.4.attn_q.weight | 0x339451620 | 0x2000000 |
|
| 344 |
+
| 245 | blk.4.attn_v.weight | 0x33b451620 | 0x800000 |
|
| 345 |
+
| 246 | blk.5.attn_norm.weight | 0x33bc51620 | 0x4000 |
|
| 346 |
+
| 247 | blk.5.ffn_down.weight | 0x33bc55620 | 0x7000000 |
|
| 347 |
+
| 248 | blk.5.ffn_gate.weight | 0x342c55620 | 0x7000000 |
|
| 348 |
+
| 249 | blk.5.ffn_up.weight | 0x349c55620 | 0x7000000 |
|
| 349 |
+
| 250 | blk.5.ffn_norm.weight | 0x350c55620 | 0x4000 |
|
| 350 |
+
| 251 | blk.5.attn_k.weight | 0x350c59620 | 0x800000 |
|
| 351 |
+
| 252 | blk.5.attn_output.weight | 0x351459620 | 0x2000000 |
|
| 352 |
+
| 253 | blk.5.attn_q.weight | 0x353459620 | 0x2000000 |
|
| 353 |
+
| 254 | blk.5.attn_v.weight | 0x355459620 | 0x800000 |
|
| 354 |
+
| 255 | blk.6.attn_norm.weight | 0x355c59620 | 0x4000 |
|
| 355 |
+
| 256 | blk.6.ffn_down.weight | 0x355c5d620 | 0x7000000 |
|
| 356 |
+
| 257 | blk.6.ffn_gate.weight | 0x35cc5d620 | 0x7000000 |
|
| 357 |
+
| 258 | blk.6.ffn_up.weight | 0x363c5d620 | 0x7000000 |
|
| 358 |
+
| 259 | blk.6.ffn_norm.weight | 0x36ac5d620 | 0x4000 |
|
| 359 |
+
| 260 | blk.6.attn_k.weight | 0x36ac61620 | 0x800000 |
|
| 360 |
+
| 261 | blk.6.attn_output.weight | 0x36b461620 | 0x2000000 |
|
| 361 |
+
| 262 | blk.6.attn_q.weight | 0x36d461620 | 0x2000000 |
|
| 362 |
+
| 263 | blk.6.attn_v.weight | 0x36f461620 | 0x800000 |
|
| 363 |
+
| 264 | blk.7.attn_norm.weight | 0x36fc61620 | 0x4000 |
|
| 364 |
+
| 265 | blk.7.ffn_down.weight | 0x36fc65620 | 0x7000000 |
|
| 365 |
+
| 266 | blk.7.ffn_gate.weight | 0x376c65620 | 0x7000000 |
|
| 366 |
+
| 267 | blk.7.ffn_up.weight | 0x37dc65620 | 0x7000000 |
|
| 367 |
+
| 268 | blk.7.ffn_norm.weight | 0x384c65620 | 0x4000 |
|
| 368 |
+
| 269 | blk.7.attn_k.weight | 0x384c69620 | 0x800000 |
|
| 369 |
+
| 270 | blk.7.attn_output.weight | 0x385469620 | 0x2000000 |
|
| 370 |
+
| 271 | blk.7.attn_q.weight | 0x387469620 | 0x2000000 |
|
| 371 |
+
| 272 | blk.7.attn_v.weight | 0x389469620 | 0x800000 |
|
| 372 |
+
| 273 | blk.8.attn_norm.weight | 0x389c69620 | 0x4000 |
|
| 373 |
+
| 274 | blk.8.ffn_down.weight | 0x389c6d620 | 0x7000000 |
|
| 374 |
+
| 275 | blk.8.ffn_gate.weight | 0x390c6d620 | 0x7000000 |
|
| 375 |
+
| 276 | blk.8.ffn_up.weight | 0x397c6d620 | 0x7000000 |
|
| 376 |
+
| 277 | blk.8.ffn_norm.weight | 0x39ec6d620 | 0x4000 |
|
| 377 |
+
| 278 | blk.8.attn_k.weight | 0x39ec71620 | 0x800000 |
|
| 378 |
+
| 279 | blk.8.attn_output.weight | 0x39f471620 | 0x2000000 |
|
| 379 |
+
| 280 | blk.8.attn_q.weight | 0x3a1471620 | 0x2000000 |
|
| 380 |
+
| 281 | blk.8.attn_v.weight | 0x3a3471620 | 0x800000 |
|
| 381 |
+
| 282 | blk.9.attn_norm.weight | 0x3a3c71620 | 0x4000 |
|
| 382 |
+
| 283 | blk.9.ffn_down.weight | 0x3a3c75620 | 0x7000000 |
|
| 383 |
+
| 284 | blk.9.ffn_gate.weight | 0x3aac75620 | 0x7000000 |
|
| 384 |
+
| 285 | blk.9.ffn_up.weight | 0x3b1c75620 | 0x7000000 |
|
| 385 |
+
| 286 | blk.9.ffn_norm.weight | 0x3b8c75620 | 0x4000 |
|
| 386 |
+
| 287 | blk.9.attn_k.weight | 0x3b8c79620 | 0x800000 |
|
| 387 |
+
| 288 | blk.9.attn_output.weight | 0x3b9479620 | 0x2000000 |
|
| 388 |
+
| 289 | blk.9.attn_q.weight | 0x3bb479620 | 0x2000000 |
|
| 389 |
+
| 290 | blk.9.attn_v.weight | 0x3bd479620 | 0x800000 |
|
| 390 |
+
| 291 | output_norm.weight | 0x3bdc79620 | 0x4000 |
|
| 391 |
+
|
| 392 |
+
### <a name="base">Base Tensor Group : ~1B Elements</a>
|
| 393 |
+
|
| 394 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 395 |
+
|-----:|:-------------------|:---------------------------------|:------------------|:----------------------|:-----|
|
| 396 |
+
| 0 | rope_freqs.weight | Rope_Freqs (W) | ( 64) 64 | 64 x 1 x 1 x 1 | F32 |
|
| 397 |
+
| 1 | output.weight | Output (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | F16 |
|
| 398 |
+
| 2 | token_embd.weight | Token Embedding (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | F16 |
|
| 399 |
+
| 291 | output_norm.weight | Output Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 400 |
+
|
| 401 |
+
- Total elements in base: ( ~1B) 1050677312
|
| 402 |
+
- Percentage of total elements: 13.08%
|
| 403 |
+
|
| 404 |
+
|
| 405 |
+
### <a name="blk_0">Block 0 Tensor Group : ~218M Elements</a>
|
| 406 |
+
|
| 407 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 408 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 409 |
+
| 3 | blk.0.attn_norm.weight | Block 0 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 410 |
+
| 4 | blk.0.ffn_down.weight | Block 0 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 411 |
+
| 5 | blk.0.ffn_gate.weight | Block 0 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 412 |
+
| 6 | blk.0.ffn_up.weight | Block 0 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 413 |
+
| 7 | blk.0.ffn_norm.weight | Block 0 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 414 |
+
| 8 | blk.0.attn_k.weight | Block 0 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 415 |
+
| 9 | blk.0.attn_output.weight | Block 0 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 416 |
+
| 10 | blk.0.attn_q.weight | Block 0 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 417 |
+
| 11 | blk.0.attn_v.weight | Block 0 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 418 |
+
|
| 419 |
+
- Total elements in blk.0: (~218M) 218112000
|
| 420 |
+
- Percentage of total elements: 2.72%
|
| 421 |
+
|
| 422 |
+
|
| 423 |
+
### <a name="blk_1">Block 1 Tensor Group : ~218M Elements</a>
|
| 424 |
+
|
| 425 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 426 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 427 |
+
| 12 | blk.1.attn_norm.weight | Block 1 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 428 |
+
| 13 | blk.1.ffn_down.weight | Block 1 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 429 |
+
| 14 | blk.1.ffn_gate.weight | Block 1 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 430 |
+
| 15 | blk.1.ffn_up.weight | Block 1 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 431 |
+
| 16 | blk.1.ffn_norm.weight | Block 1 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 432 |
+
| 17 | blk.1.attn_k.weight | Block 1 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 433 |
+
| 18 | blk.1.attn_output.weight | Block 1 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 434 |
+
| 19 | blk.1.attn_q.weight | Block 1 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 435 |
+
| 20 | blk.1.attn_v.weight | Block 1 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 436 |
+
|
| 437 |
+
- Total elements in blk.1: (~218M) 218112000
|
| 438 |
+
- Percentage of total elements: 2.72%
|
| 439 |
+
|
| 440 |
+
|
| 441 |
+
### <a name="blk_10">Block 10 Tensor Group : ~218M Elements</a>
|
| 442 |
+
|
| 443 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 444 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 445 |
+
| 21 | blk.10.attn_norm.weight | Block 10 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 446 |
+
| 22 | blk.10.ffn_down.weight | Block 10 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 447 |
+
| 23 | blk.10.ffn_gate.weight | Block 10 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 448 |
+
| 24 | blk.10.ffn_up.weight | Block 10 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 449 |
+
| 25 | blk.10.ffn_norm.weight | Block 10 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 450 |
+
| 26 | blk.10.attn_k.weight | Block 10 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 451 |
+
| 27 | blk.10.attn_output.weight | Block 10 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 452 |
+
| 28 | blk.10.attn_q.weight | Block 10 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 453 |
+
| 29 | blk.10.attn_v.weight | Block 10 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 454 |
+
|
| 455 |
+
- Total elements in blk.10: (~218M) 218112000
|
| 456 |
+
- Percentage of total elements: 2.72%
|
| 457 |
+
|
| 458 |
+
|
| 459 |
+
### <a name="blk_11">Block 11 Tensor Group : ~218M Elements</a>
|
| 460 |
+
|
| 461 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 462 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 463 |
+
| 30 | blk.11.attn_norm.weight | Block 11 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 464 |
+
| 31 | blk.11.ffn_down.weight | Block 11 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 465 |
+
| 32 | blk.11.ffn_gate.weight | Block 11 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 466 |
+
| 33 | blk.11.ffn_up.weight | Block 11 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 467 |
+
| 34 | blk.11.ffn_norm.weight | Block 11 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 468 |
+
| 35 | blk.11.attn_k.weight | Block 11 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 469 |
+
| 36 | blk.11.attn_output.weight | Block 11 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 470 |
+
| 37 | blk.11.attn_q.weight | Block 11 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 471 |
+
| 38 | blk.11.attn_v.weight | Block 11 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 472 |
+
|
| 473 |
+
- Total elements in blk.11: (~218M) 218112000
|
| 474 |
+
- Percentage of total elements: 2.72%
|
| 475 |
+
|
| 476 |
+
|
| 477 |
+
### <a name="blk_12">Block 12 Tensor Group : ~218M Elements</a>
|
| 478 |
+
|
| 479 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 480 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 481 |
+
| 39 | blk.12.attn_norm.weight | Block 12 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 482 |
+
| 40 | blk.12.ffn_down.weight | Block 12 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 483 |
+
| 41 | blk.12.ffn_gate.weight | Block 12 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 484 |
+
| 42 | blk.12.ffn_up.weight | Block 12 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 485 |
+
| 43 | blk.12.ffn_norm.weight | Block 12 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 486 |
+
| 44 | blk.12.attn_k.weight | Block 12 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 487 |
+
| 45 | blk.12.attn_output.weight | Block 12 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 488 |
+
| 46 | blk.12.attn_q.weight | Block 12 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 489 |
+
| 47 | blk.12.attn_v.weight | Block 12 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 490 |
+
|
| 491 |
+
- Total elements in blk.12: (~218M) 218112000
|
| 492 |
+
- Percentage of total elements: 2.72%
|
| 493 |
+
|
| 494 |
+
|
| 495 |
+
### <a name="blk_13">Block 13 Tensor Group : ~218M Elements</a>
|
| 496 |
+
|
| 497 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 498 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 499 |
+
| 48 | blk.13.attn_norm.weight | Block 13 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 500 |
+
| 49 | blk.13.ffn_down.weight | Block 13 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 501 |
+
| 50 | blk.13.ffn_gate.weight | Block 13 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 502 |
+
| 51 | blk.13.ffn_up.weight | Block 13 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 503 |
+
| 52 | blk.13.ffn_norm.weight | Block 13 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 504 |
+
| 53 | blk.13.attn_k.weight | Block 13 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 505 |
+
| 54 | blk.13.attn_output.weight | Block 13 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 506 |
+
| 55 | blk.13.attn_q.weight | Block 13 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 507 |
+
| 56 | blk.13.attn_v.weight | Block 13 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 508 |
+
|
| 509 |
+
- Total elements in blk.13: (~218M) 218112000
|
| 510 |
+
- Percentage of total elements: 2.72%
|
| 511 |
+
|
| 512 |
+
|
| 513 |
+
### <a name="blk_14">Block 14 Tensor Group : ~218M Elements</a>
|
| 514 |
+
|
| 515 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 516 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 517 |
+
| 57 | blk.14.attn_norm.weight | Block 14 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 518 |
+
| 58 | blk.14.ffn_down.weight | Block 14 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 519 |
+
| 59 | blk.14.ffn_gate.weight | Block 14 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 520 |
+
| 60 | blk.14.ffn_up.weight | Block 14 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 521 |
+
| 61 | blk.14.ffn_norm.weight | Block 14 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 522 |
+
| 62 | blk.14.attn_k.weight | Block 14 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 523 |
+
| 63 | blk.14.attn_output.weight | Block 14 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 524 |
+
| 64 | blk.14.attn_q.weight | Block 14 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 525 |
+
| 65 | blk.14.attn_v.weight | Block 14 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 526 |
+
|
| 527 |
+
- Total elements in blk.14: (~218M) 218112000
|
| 528 |
+
- Percentage of total elements: 2.72%
|
| 529 |
+
|
| 530 |
+
|
| 531 |
+
### <a name="blk_15">Block 15 Tensor Group : ~218M Elements</a>
|
| 532 |
+
|
| 533 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 534 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 535 |
+
| 66 | blk.15.attn_norm.weight | Block 15 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 536 |
+
| 67 | blk.15.ffn_down.weight | Block 15 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 537 |
+
| 68 | blk.15.ffn_gate.weight | Block 15 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 538 |
+
| 69 | blk.15.ffn_up.weight | Block 15 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 539 |
+
| 70 | blk.15.ffn_norm.weight | Block 15 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 540 |
+
| 71 | blk.15.attn_k.weight | Block 15 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 541 |
+
| 72 | blk.15.attn_output.weight | Block 15 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 542 |
+
| 73 | blk.15.attn_q.weight | Block 15 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 543 |
+
| 74 | blk.15.attn_v.weight | Block 15 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 544 |
+
|
| 545 |
+
- Total elements in blk.15: (~218M) 218112000
|
| 546 |
+
- Percentage of total elements: 2.72%
|
| 547 |
+
|
| 548 |
+
|
| 549 |
+
### <a name="blk_16">Block 16 Tensor Group : ~218M Elements</a>
|
| 550 |
+
|
| 551 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 552 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 553 |
+
| 75 | blk.16.attn_norm.weight | Block 16 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 554 |
+
| 76 | blk.16.ffn_down.weight | Block 16 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 555 |
+
| 77 | blk.16.ffn_gate.weight | Block 16 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 556 |
+
| 78 | blk.16.ffn_up.weight | Block 16 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 557 |
+
| 79 | blk.16.ffn_norm.weight | Block 16 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 558 |
+
| 80 | blk.16.attn_k.weight | Block 16 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 559 |
+
| 81 | blk.16.attn_output.weight | Block 16 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 560 |
+
| 82 | blk.16.attn_q.weight | Block 16 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 561 |
+
| 83 | blk.16.attn_v.weight | Block 16 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 562 |
+
|
| 563 |
+
- Total elements in blk.16: (~218M) 218112000
|
| 564 |
+
- Percentage of total elements: 2.72%
|
| 565 |
+
|
| 566 |
+
|
| 567 |
+
### <a name="blk_17">Block 17 Tensor Group : ~218M Elements</a>
|
| 568 |
+
|
| 569 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 570 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 571 |
+
| 84 | blk.17.attn_norm.weight | Block 17 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 572 |
+
| 85 | blk.17.ffn_down.weight | Block 17 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 573 |
+
| 86 | blk.17.ffn_gate.weight | Block 17 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 574 |
+
| 87 | blk.17.ffn_up.weight | Block 17 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 575 |
+
| 88 | blk.17.ffn_norm.weight | Block 17 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 576 |
+
| 89 | blk.17.attn_k.weight | Block 17 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 577 |
+
| 90 | blk.17.attn_output.weight | Block 17 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 578 |
+
| 91 | blk.17.attn_q.weight | Block 17 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 579 |
+
| 92 | blk.17.attn_v.weight | Block 17 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 580 |
+
|
| 581 |
+
- Total elements in blk.17: (~218M) 218112000
|
| 582 |
+
- Percentage of total elements: 2.72%
|
| 583 |
+
|
| 584 |
+
|
| 585 |
+
### <a name="blk_18">Block 18 Tensor Group : ~218M Elements</a>
|
| 586 |
+
|
| 587 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 588 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 589 |
+
| 93 | blk.18.attn_norm.weight | Block 18 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 590 |
+
| 94 | blk.18.ffn_down.weight | Block 18 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 591 |
+
| 95 | blk.18.ffn_gate.weight | Block 18 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 592 |
+
| 96 | blk.18.ffn_up.weight | Block 18 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 593 |
+
| 97 | blk.18.ffn_norm.weight | Block 18 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 594 |
+
| 98 | blk.18.attn_k.weight | Block 18 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 595 |
+
| 99 | blk.18.attn_output.weight | Block 18 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 596 |
+
| 100 | blk.18.attn_q.weight | Block 18 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 597 |
+
| 101 | blk.18.attn_v.weight | Block 18 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 598 |
+
|
| 599 |
+
- Total elements in blk.18: (~218M) 218112000
|
| 600 |
+
- Percentage of total elements: 2.72%
|
| 601 |
+
|
| 602 |
+
|
| 603 |
+
### <a name="blk_19">Block 19 Tensor Group : ~218M Elements</a>
|
| 604 |
+
|
| 605 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 606 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 607 |
+
| 102 | blk.19.attn_norm.weight | Block 19 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 608 |
+
| 103 | blk.19.ffn_down.weight | Block 19 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 609 |
+
| 104 | blk.19.ffn_gate.weight | Block 19 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 610 |
+
| 105 | blk.19.ffn_up.weight | Block 19 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 611 |
+
| 106 | blk.19.ffn_norm.weight | Block 19 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 612 |
+
| 107 | blk.19.attn_k.weight | Block 19 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 613 |
+
| 108 | blk.19.attn_output.weight | Block 19 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 614 |
+
| 109 | blk.19.attn_q.weight | Block 19 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 615 |
+
| 110 | blk.19.attn_v.weight | Block 19 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 616 |
+
|
| 617 |
+
- Total elements in blk.19: (~218M) 218112000
|
| 618 |
+
- Percentage of total elements: 2.72%
|
| 619 |
+
|
| 620 |
+
|
| 621 |
+
### <a name="blk_2">Block 2 Tensor Group : ~218M Elements</a>
|
| 622 |
+
|
| 623 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 624 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 625 |
+
| 111 | blk.2.attn_norm.weight | Block 2 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 626 |
+
| 112 | blk.2.ffn_down.weight | Block 2 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 627 |
+
| 113 | blk.2.ffn_gate.weight | Block 2 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 628 |
+
| 114 | blk.2.ffn_up.weight | Block 2 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 629 |
+
| 115 | blk.2.ffn_norm.weight | Block 2 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 630 |
+
| 116 | blk.2.attn_k.weight | Block 2 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 631 |
+
| 117 | blk.2.attn_output.weight | Block 2 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 632 |
+
| 118 | blk.2.attn_q.weight | Block 2 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 633 |
+
| 119 | blk.2.attn_v.weight | Block 2 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 634 |
+
|
| 635 |
+
- Total elements in blk.2: (~218M) 218112000
|
| 636 |
+
- Percentage of total elements: 2.72%
|
| 637 |
+
|
| 638 |
+
|
| 639 |
+
### <a name="blk_20">Block 20 Tensor Group : ~218M Elements</a>
|
| 640 |
+
|
| 641 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 642 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 643 |
+
| 120 | blk.20.attn_norm.weight | Block 20 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 644 |
+
| 121 | blk.20.ffn_down.weight | Block 20 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 645 |
+
| 122 | blk.20.ffn_gate.weight | Block 20 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 646 |
+
| 123 | blk.20.ffn_up.weight | Block 20 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 647 |
+
| 124 | blk.20.ffn_norm.weight | Block 20 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 648 |
+
| 125 | blk.20.attn_k.weight | Block 20 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 649 |
+
| 126 | blk.20.attn_output.weight | Block 20 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 650 |
+
| 127 | blk.20.attn_q.weight | Block 20 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 651 |
+
| 128 | blk.20.attn_v.weight | Block 20 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 652 |
+
|
| 653 |
+
- Total elements in blk.20: (~218M) 218112000
|
| 654 |
+
- Percentage of total elements: 2.72%
|
| 655 |
+
|
| 656 |
+
|
| 657 |
+
### <a name="blk_21">Block 21 Tensor Group : ~218M Elements</a>
|
| 658 |
+
|
| 659 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 660 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 661 |
+
| 129 | blk.21.attn_norm.weight | Block 21 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 662 |
+
| 130 | blk.21.ffn_down.weight | Block 21 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 663 |
+
| 131 | blk.21.ffn_gate.weight | Block 21 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 664 |
+
| 132 | blk.21.ffn_up.weight | Block 21 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 665 |
+
| 133 | blk.21.ffn_norm.weight | Block 21 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 666 |
+
| 134 | blk.21.attn_k.weight | Block 21 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 667 |
+
| 135 | blk.21.attn_output.weight | Block 21 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 668 |
+
| 136 | blk.21.attn_q.weight | Block 21 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 669 |
+
| 137 | blk.21.attn_v.weight | Block 21 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 670 |
+
|
| 671 |
+
- Total elements in blk.21: (~218M) 218112000
|
| 672 |
+
- Percentage of total elements: 2.72%
|
| 673 |
+
|
| 674 |
+
|
| 675 |
+
### <a name="blk_22">Block 22 Tensor Group : ~218M Elements</a>
|
| 676 |
+
|
| 677 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 678 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 679 |
+
| 138 | blk.22.attn_norm.weight | Block 22 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 680 |
+
| 139 | blk.22.ffn_down.weight | Block 22 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 681 |
+
| 140 | blk.22.ffn_gate.weight | Block 22 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 682 |
+
| 141 | blk.22.ffn_up.weight | Block 22 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 683 |
+
| 142 | blk.22.ffn_norm.weight | Block 22 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 684 |
+
| 143 | blk.22.attn_k.weight | Block 22 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 685 |
+
| 144 | blk.22.attn_output.weight | Block 22 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 686 |
+
| 145 | blk.22.attn_q.weight | Block 22 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 687 |
+
| 146 | blk.22.attn_v.weight | Block 22 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 688 |
+
|
| 689 |
+
- Total elements in blk.22: (~218M) 218112000
|
| 690 |
+
- Percentage of total elements: 2.72%
|
| 691 |
+
|
| 692 |
+
|
| 693 |
+
### <a name="blk_23">Block 23 Tensor Group : ~218M Elements</a>
|
| 694 |
+
|
| 695 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 696 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 697 |
+
| 147 | blk.23.attn_norm.weight | Block 23 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 698 |
+
| 148 | blk.23.ffn_down.weight | Block 23 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 699 |
+
| 149 | blk.23.ffn_gate.weight | Block 23 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 700 |
+
| 150 | blk.23.ffn_up.weight | Block 23 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 701 |
+
| 151 | blk.23.ffn_norm.weight | Block 23 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 702 |
+
| 152 | blk.23.attn_k.weight | Block 23 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 703 |
+
| 153 | blk.23.attn_output.weight | Block 23 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 704 |
+
| 154 | blk.23.attn_q.weight | Block 23 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 705 |
+
| 155 | blk.23.attn_v.weight | Block 23 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 706 |
+
|
| 707 |
+
- Total elements in blk.23: (~218M) 218112000
|
| 708 |
+
- Percentage of total elements: 2.72%
|
| 709 |
+
|
| 710 |
+
|
| 711 |
+
### <a name="blk_24">Block 24 Tensor Group : ~218M Elements</a>
|
| 712 |
+
|
| 713 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 714 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 715 |
+
| 156 | blk.24.attn_norm.weight | Block 24 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 716 |
+
| 157 | blk.24.ffn_down.weight | Block 24 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 717 |
+
| 158 | blk.24.ffn_gate.weight | Block 24 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 718 |
+
| 159 | blk.24.ffn_up.weight | Block 24 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 719 |
+
| 160 | blk.24.ffn_norm.weight | Block 24 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 720 |
+
| 161 | blk.24.attn_k.weight | Block 24 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 721 |
+
| 162 | blk.24.attn_output.weight | Block 24 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 722 |
+
| 163 | blk.24.attn_q.weight | Block 24 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 723 |
+
| 164 | blk.24.attn_v.weight | Block 24 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 724 |
+
|
| 725 |
+
- Total elements in blk.24: (~218M) 218112000
|
| 726 |
+
- Percentage of total elements: 2.72%
|
| 727 |
+
|
| 728 |
+
|
| 729 |
+
### <a name="blk_25">Block 25 Tensor Group : ~218M Elements</a>
|
| 730 |
+
|
| 731 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 732 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 733 |
+
| 165 | blk.25.attn_norm.weight | Block 25 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 734 |
+
| 166 | blk.25.ffn_down.weight | Block 25 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 735 |
+
| 167 | blk.25.ffn_gate.weight | Block 25 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 736 |
+
| 168 | blk.25.ffn_up.weight | Block 25 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 737 |
+
| 169 | blk.25.ffn_norm.weight | Block 25 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 738 |
+
| 170 | blk.25.attn_k.weight | Block 25 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 739 |
+
| 171 | blk.25.attn_output.weight | Block 25 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 740 |
+
| 172 | blk.25.attn_q.weight | Block 25 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 741 |
+
| 173 | blk.25.attn_v.weight | Block 25 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 742 |
+
|
| 743 |
+
- Total elements in blk.25: (~218M) 218112000
|
| 744 |
+
- Percentage of total elements: 2.72%
|
| 745 |
+
|
| 746 |
+
|
| 747 |
+
### <a name="blk_26">Block 26 Tensor Group : ~218M Elements</a>
|
| 748 |
+
|
| 749 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 750 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 751 |
+
| 174 | blk.26.attn_norm.weight | Block 26 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 752 |
+
| 175 | blk.26.ffn_down.weight | Block 26 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 753 |
+
| 176 | blk.26.ffn_gate.weight | Block 26 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 754 |
+
| 177 | blk.26.ffn_up.weight | Block 26 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 755 |
+
| 178 | blk.26.ffn_norm.weight | Block 26 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 756 |
+
| 179 | blk.26.attn_k.weight | Block 26 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 757 |
+
| 180 | blk.26.attn_output.weight | Block 26 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 758 |
+
| 181 | blk.26.attn_q.weight | Block 26 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 759 |
+
| 182 | blk.26.attn_v.weight | Block 26 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 760 |
+
|
| 761 |
+
- Total elements in blk.26: (~218M) 218112000
|
| 762 |
+
- Percentage of total elements: 2.72%
|
| 763 |
+
|
| 764 |
+
|
| 765 |
+
### <a name="blk_27">Block 27 Tensor Group : ~218M Elements</a>
|
| 766 |
+
|
| 767 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 768 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 769 |
+
| 183 | blk.27.attn_norm.weight | Block 27 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 770 |
+
| 184 | blk.27.ffn_down.weight | Block 27 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 771 |
+
| 185 | blk.27.ffn_gate.weight | Block 27 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 772 |
+
| 186 | blk.27.ffn_up.weight | Block 27 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 773 |
+
| 187 | blk.27.ffn_norm.weight | Block 27 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 774 |
+
| 188 | blk.27.attn_k.weight | Block 27 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 775 |
+
| 189 | blk.27.attn_output.weight | Block 27 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 776 |
+
| 190 | blk.27.attn_q.weight | Block 27 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 777 |
+
| 191 | blk.27.attn_v.weight | Block 27 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 778 |
+
|
| 779 |
+
- Total elements in blk.27: (~218M) 218112000
|
| 780 |
+
- Percentage of total elements: 2.72%
|
| 781 |
+
|
| 782 |
+
|
| 783 |
+
### <a name="blk_28">Block 28 Tensor Group : ~218M Elements</a>
|
| 784 |
+
|
| 785 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 786 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 787 |
+
| 192 | blk.28.attn_norm.weight | Block 28 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 788 |
+
| 193 | blk.28.ffn_down.weight | Block 28 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 789 |
+
| 194 | blk.28.ffn_gate.weight | Block 28 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 790 |
+
| 195 | blk.28.ffn_up.weight | Block 28 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 791 |
+
| 196 | blk.28.ffn_norm.weight | Block 28 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 792 |
+
| 197 | blk.28.attn_k.weight | Block 28 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 793 |
+
| 198 | blk.28.attn_output.weight | Block 28 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 794 |
+
| 199 | blk.28.attn_q.weight | Block 28 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 795 |
+
| 200 | blk.28.attn_v.weight | Block 28 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 796 |
+
|
| 797 |
+
- Total elements in blk.28: (~218M) 218112000
|
| 798 |
+
- Percentage of total elements: 2.72%
|
| 799 |
+
|
| 800 |
+
|
| 801 |
+
### <a name="blk_29">Block 29 Tensor Group : ~218M Elements</a>
|
| 802 |
+
|
| 803 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 804 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 805 |
+
| 201 | blk.29.attn_norm.weight | Block 29 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 806 |
+
| 202 | blk.29.ffn_down.weight | Block 29 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 807 |
+
| 203 | blk.29.ffn_gate.weight | Block 29 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 808 |
+
| 204 | blk.29.ffn_up.weight | Block 29 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 809 |
+
| 205 | blk.29.ffn_norm.weight | Block 29 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 810 |
+
| 206 | blk.29.attn_k.weight | Block 29 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 811 |
+
| 207 | blk.29.attn_output.weight | Block 29 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 812 |
+
| 208 | blk.29.attn_q.weight | Block 29 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 813 |
+
| 209 | blk.29.attn_v.weight | Block 29 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 814 |
+
|
| 815 |
+
- Total elements in blk.29: (~218M) 218112000
|
| 816 |
+
- Percentage of total elements: 2.72%
|
| 817 |
+
|
| 818 |
+
|
| 819 |
+
### <a name="blk_3">Block 3 Tensor Group : ~218M Elements</a>
|
| 820 |
+
|
| 821 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 822 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 823 |
+
| 210 | blk.3.attn_norm.weight | Block 3 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 824 |
+
| 211 | blk.3.ffn_down.weight | Block 3 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 825 |
+
| 212 | blk.3.ffn_gate.weight | Block 3 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 826 |
+
| 213 | blk.3.ffn_up.weight | Block 3 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 827 |
+
| 214 | blk.3.ffn_norm.weight | Block 3 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 828 |
+
| 215 | blk.3.attn_k.weight | Block 3 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 829 |
+
| 216 | blk.3.attn_output.weight | Block 3 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 830 |
+
| 217 | blk.3.attn_q.weight | Block 3 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 831 |
+
| 218 | blk.3.attn_v.weight | Block 3 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 832 |
+
|
| 833 |
+
- Total elements in blk.3: (~218M) 218112000
|
| 834 |
+
- Percentage of total elements: 2.72%
|
| 835 |
+
|
| 836 |
+
|
| 837 |
+
### <a name="blk_30">Block 30 Tensor Group : ~218M Elements</a>
|
| 838 |
+
|
| 839 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 840 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 841 |
+
| 219 | blk.30.attn_norm.weight | Block 30 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 842 |
+
| 220 | blk.30.ffn_down.weight | Block 30 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 843 |
+
| 221 | blk.30.ffn_gate.weight | Block 30 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 844 |
+
| 222 | blk.30.ffn_up.weight | Block 30 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 845 |
+
| 223 | blk.30.ffn_norm.weight | Block 30 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 846 |
+
| 224 | blk.30.attn_k.weight | Block 30 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 847 |
+
| 225 | blk.30.attn_output.weight | Block 30 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 848 |
+
| 226 | blk.30.attn_q.weight | Block 30 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 849 |
+
| 227 | blk.30.attn_v.weight | Block 30 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 850 |
+
|
| 851 |
+
- Total elements in blk.30: (~218M) 218112000
|
| 852 |
+
- Percentage of total elements: 2.72%
|
| 853 |
+
|
| 854 |
+
|
| 855 |
+
### <a name="blk_31">Block 31 Tensor Group : ~218M Elements</a>
|
| 856 |
+
|
| 857 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 858 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 859 |
+
| 228 | blk.31.attn_norm.weight | Block 31 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 860 |
+
| 229 | blk.31.ffn_down.weight | Block 31 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 861 |
+
| 230 | blk.31.ffn_gate.weight | Block 31 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 862 |
+
| 231 | blk.31.ffn_up.weight | Block 31 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 863 |
+
| 232 | blk.31.ffn_norm.weight | Block 31 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 864 |
+
| 233 | blk.31.attn_k.weight | Block 31 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 865 |
+
| 234 | blk.31.attn_output.weight | Block 31 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 866 |
+
| 235 | blk.31.attn_q.weight | Block 31 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 867 |
+
| 236 | blk.31.attn_v.weight | Block 31 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 868 |
+
|
| 869 |
+
- Total elements in blk.31: (~218M) 218112000
|
| 870 |
+
- Percentage of total elements: 2.72%
|
| 871 |
+
|
| 872 |
+
|
| 873 |
+
### <a name="blk_4">Block 4 Tensor Group : ~218M Elements</a>
|
| 874 |
+
|
| 875 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 876 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 877 |
+
| 237 | blk.4.attn_norm.weight | Block 4 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 878 |
+
| 238 | blk.4.ffn_down.weight | Block 4 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 879 |
+
| 239 | blk.4.ffn_gate.weight | Block 4 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 880 |
+
| 240 | blk.4.ffn_up.weight | Block 4 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 881 |
+
| 241 | blk.4.ffn_norm.weight | Block 4 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 882 |
+
| 242 | blk.4.attn_k.weight | Block 4 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 883 |
+
| 243 | blk.4.attn_output.weight | Block 4 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 884 |
+
| 244 | blk.4.attn_q.weight | Block 4 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 885 |
+
| 245 | blk.4.attn_v.weight | Block 4 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 886 |
+
|
| 887 |
+
- Total elements in blk.4: (~218M) 218112000
|
| 888 |
+
- Percentage of total elements: 2.72%
|
| 889 |
+
|
| 890 |
+
|
| 891 |
+
### <a name="blk_5">Block 5 Tensor Group : ~218M Elements</a>
|
| 892 |
+
|
| 893 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 894 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 895 |
+
| 246 | blk.5.attn_norm.weight | Block 5 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 896 |
+
| 247 | blk.5.ffn_down.weight | Block 5 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 897 |
+
| 248 | blk.5.ffn_gate.weight | Block 5 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 898 |
+
| 249 | blk.5.ffn_up.weight | Block 5 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 899 |
+
| 250 | blk.5.ffn_norm.weight | Block 5 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 900 |
+
| 251 | blk.5.attn_k.weight | Block 5 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 901 |
+
| 252 | blk.5.attn_output.weight | Block 5 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 902 |
+
| 253 | blk.5.attn_q.weight | Block 5 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 903 |
+
| 254 | blk.5.attn_v.weight | Block 5 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 904 |
+
|
| 905 |
+
- Total elements in blk.5: (~218M) 218112000
|
| 906 |
+
- Percentage of total elements: 2.72%
|
| 907 |
+
|
| 908 |
+
|
| 909 |
+
### <a name="blk_6">Block 6 Tensor Group : ~218M Elements</a>
|
| 910 |
+
|
| 911 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 912 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 913 |
+
| 255 | blk.6.attn_norm.weight | Block 6 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 914 |
+
| 256 | blk.6.ffn_down.weight | Block 6 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 915 |
+
| 257 | blk.6.ffn_gate.weight | Block 6 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 916 |
+
| 258 | blk.6.ffn_up.weight | Block 6 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 917 |
+
| 259 | blk.6.ffn_norm.weight | Block 6 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 918 |
+
| 260 | blk.6.attn_k.weight | Block 6 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 919 |
+
| 261 | blk.6.attn_output.weight | Block 6 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 920 |
+
| 262 | blk.6.attn_q.weight | Block 6 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 921 |
+
| 263 | blk.6.attn_v.weight | Block 6 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 922 |
+
|
| 923 |
+
- Total elements in blk.6: (~218M) 218112000
|
| 924 |
+
- Percentage of total elements: 2.72%
|
| 925 |
+
|
| 926 |
+
|
| 927 |
+
### <a name="blk_7">Block 7 Tensor Group : ~218M Elements</a>
|
| 928 |
+
|
| 929 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 930 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 931 |
+
| 264 | blk.7.attn_norm.weight | Block 7 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 932 |
+
| 265 | blk.7.ffn_down.weight | Block 7 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 933 |
+
| 266 | blk.7.ffn_gate.weight | Block 7 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 934 |
+
| 267 | blk.7.ffn_up.weight | Block 7 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 935 |
+
| 268 | blk.7.ffn_norm.weight | Block 7 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 936 |
+
| 269 | blk.7.attn_k.weight | Block 7 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 937 |
+
| 270 | blk.7.attn_output.weight | Block 7 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 938 |
+
| 271 | blk.7.attn_q.weight | Block 7 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 939 |
+
| 272 | blk.7.attn_v.weight | Block 7 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 940 |
+
|
| 941 |
+
- Total elements in blk.7: (~218M) 218112000
|
| 942 |
+
- Percentage of total elements: 2.72%
|
| 943 |
+
|
| 944 |
+
|
| 945 |
+
### <a name="blk_8">Block 8 Tensor Group : ~218M Elements</a>
|
| 946 |
+
|
| 947 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 948 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 949 |
+
| 273 | blk.8.attn_norm.weight | Block 8 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 950 |
+
| 274 | blk.8.ffn_down.weight | Block 8 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 951 |
+
| 275 | blk.8.ffn_gate.weight | Block 8 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 952 |
+
| 276 | blk.8.ffn_up.weight | Block 8 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 953 |
+
| 277 | blk.8.ffn_norm.weight | Block 8 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 954 |
+
| 278 | blk.8.attn_k.weight | Block 8 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 955 |
+
| 279 | blk.8.attn_output.weight | Block 8 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 956 |
+
| 280 | blk.8.attn_q.weight | Block 8 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 957 |
+
| 281 | blk.8.attn_v.weight | Block 8 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 958 |
+
|
| 959 |
+
- Total elements in blk.8: (~218M) 218112000
|
| 960 |
+
- Percentage of total elements: 2.72%
|
| 961 |
+
|
| 962 |
+
|
| 963 |
+
### <a name="blk_9">Block 9 Tensor Group : ~218M Elements</a>
|
| 964 |
+
|
| 965 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 966 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 967 |
+
| 282 | blk.9.attn_norm.weight | Block 9 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 968 |
+
| 283 | blk.9.ffn_down.weight | Block 9 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | F16 |
|
| 969 |
+
| 284 | blk.9.ffn_gate.weight | Block 9 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 970 |
+
| 285 | blk.9.ffn_up.weight | Block 9 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | F16 |
|
| 971 |
+
| 286 | blk.9.ffn_norm.weight | Block 9 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 972 |
+
| 287 | blk.9.attn_k.weight | Block 9 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 973 |
+
| 288 | blk.9.attn_output.weight | Block 9 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 974 |
+
| 289 | blk.9.attn_q.weight | Block 9 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | F16 |
|
| 975 |
+
| 290 | blk.9.attn_v.weight | Block 9 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 976 |
+
|
| 977 |
+
- Total elements in blk.9: (~218M) 218112000
|
| 978 |
+
- Percentage of total elements: 2.72%
|
scores/Watt-Tool-8B-IQ3_M.md
ADDED
|
@@ -0,0 +1,982 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Watt-Tool-8B-IQ3_M.gguf - GGUF Internal File Dump
|
| 2 |
+
|
| 3 |
+
- Endian: LITTLE endian
|
| 4 |
+
|
| 5 |
+
## Key Value Metadata Store
|
| 6 |
+
|
| 7 |
+
There are 43 key-value pairs in this file
|
| 8 |
+
|
| 9 |
+
| POS | TYPE | Count | Key | Value |
|
| 10 |
+
|----:|:---------|-------:|:---------------------------------------|:--------------------------------------------------------------------|
|
| 11 |
+
| 1 | UINT32 | 1 | GGUF.version | 3 |
|
| 12 |
+
| 2 | UINT64 | 1 | GGUF.tensor_count | 292 |
|
| 13 |
+
| 3 | UINT64 | 1 | GGUF.kv_count | 40 |
|
| 14 |
+
| 4 | STRING | 1 | general.architecture | `llama` |
|
| 15 |
+
| 5 | STRING | 1 | general.type | `model` |
|
| 16 |
+
| 6 | STRING | 1 | general.name | `Watt Tool 8B GGUF` |
|
| 17 |
+
| 7 | STRING | 1 | general.finetune | `GGUF` |
|
| 18 |
+
| 8 | STRING | 1 | general.basename | `Watt-Tool` |
|
| 19 |
+
| 9 | STRING | 1 | general.size_label | `8B` |
|
| 20 |
+
| 10 | STRING | 1 | general.license | `apache-2.0` |
|
| 21 |
+
| 11 | UINT32 | 1 | general.base_model.count | 1 |
|
| 22 |
+
| 12 | STRING | 1 | general.base_model.0.name | `Llama 3.1 8B Instruct` |
|
| 23 |
+
| 13 | STRING | 1 | general.base_model.0.organization | `Meta Llama` |
|
| 24 |
+
| 14 | STRING | 1 | general.base_model.0.repo_url | `https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct` |
|
| 25 |
+
| 15 | [STRING] | 4 | general.tags | [ `function-calling`, `tool-use`, `llama`, `bfcl` ] |
|
| 26 |
+
| 16 | [STRING] | 1 | general.languages | [ `en` ] |
|
| 27 |
+
| 17 | UINT32 | 1 | llama.block_count | 32 |
|
| 28 |
+
| 18 | UINT32 | 1 | llama.context_length | 131072 |
|
| 29 |
+
| 19 | UINT32 | 1 | llama.embedding_length | 4096 |
|
| 30 |
+
| 20 | UINT32 | 1 | llama.feed_forward_length | 14336 |
|
| 31 |
+
| 21 | UINT32 | 1 | llama.attention.head_count | 32 |
|
| 32 |
+
| 22 | UINT32 | 1 | llama.attention.head_count_kv | 8 |
|
| 33 |
+
| 23 | FLOAT32 | 1 | llama.rope.freq_base | 500000.0 |
|
| 34 |
+
| 24 | FLOAT32 | 1 | llama.attention.layer_norm_rms_epsilon | 1e-05 |
|
| 35 |
+
| 25 | UINT32 | 1 | llama.attention.key_length | 128 |
|
| 36 |
+
| 26 | UINT32 | 1 | llama.attention.value_length | 128 |
|
| 37 |
+
| 27 | UINT32 | 1 | llama.vocab_size | 128256 |
|
| 38 |
+
| 28 | UINT32 | 1 | llama.rope.dimension_count | 128 |
|
| 39 |
+
| 29 | STRING | 1 | tokenizer.ggml.model | `gpt2` |
|
| 40 |
+
| 30 | STRING | 1 | tokenizer.ggml.pre | `llama-bpe` |
|
| 41 |
+
| 31 | [STRING] | 128256 | tokenizer.ggml.tokens | [ `!`, `"`, `#`, `$`, `%`, ... ] |
|
| 42 |
+
| 32 | [INT32] | 128256 | tokenizer.ggml.token_type | [ 1, 1, 1, 1, 1, 1, 1, ... ] |
|
| 43 |
+
| 33 | [STRING] | 280147 | tokenizer.ggml.merges | [ `Ġ Ġ`, `Ġ ĠĠĠ`, `ĠĠ ĠĠ`, `ĠĠĠ Ġ`, `i n`, ... ] |
|
| 44 |
+
| 34 | UINT32 | 1 | tokenizer.ggml.bos_token_id | 128000 |
|
| 45 |
+
| 35 | UINT32 | 1 | tokenizer.ggml.eos_token_id | 128009 |
|
| 46 |
+
| 36 | UINT32 | 1 | tokenizer.ggml.padding_token_id | 128009 |
|
| 47 |
+
| 37 | STRING | 1 | tokenizer.chat_template | `{{ '<|begin_of_text|>' }}{% if`...`d|>' }}{% endif %}{% endfor %}` |
|
| 48 |
+
| 38 | UINT32 | 1 | general.quantization_version | 2 |
|
| 49 |
+
| 39 | UINT32 | 1 | general.file_type | 27 |
|
| 50 |
+
| 40 | STRING | 1 | quantize.imatrix.file | `./imatrix/imatrix-Watt-Tool-8B-small.dat` |
|
| 51 |
+
| 41 | STRING | 1 | quantize.imatrix.dataset | `../../datasets/imatrix/calibration_eur_small.txt` |
|
| 52 |
+
| 42 | INT32 | 1 | quantize.imatrix.entries_count | 225 |
|
| 53 |
+
| 43 | INT32 | 1 | quantize.imatrix.chunks_count | 962 |
|
| 54 |
+
|
| 55 |
+
## Tensors Overview ~8B Elements
|
| 56 |
+
|
| 57 |
+
Total number of elements in all tensors: 8030261312 Elements
|
| 58 |
+
|
| 59 |
+
- [Watt-Tool-8B-IQ3\_M.gguf - GGUF Internal File Dump](#watt-tool-8b-iq3_mgguf---gguf-internal-file-dump)
|
| 60 |
+
- [Key Value Metadata Store](#key-value-metadata-store)
|
| 61 |
+
- [Tensors Overview ~8B Elements](#tensors-overview-8b-elements)
|
| 62 |
+
- [Tensor Data Offset](#tensor-data-offset)
|
| 63 |
+
- [Base Tensor Group : ~1B Elements](#base-tensor-group--1b-elements)
|
| 64 |
+
- [Block 0 Tensor Group : ~218M Elements](#block-0-tensor-group--218m-elements)
|
| 65 |
+
- [Block 1 Tensor Group : ~218M Elements](#block-1-tensor-group--218m-elements)
|
| 66 |
+
- [Block 2 Tensor Group : ~218M Elements](#block-2-tensor-group--218m-elements)
|
| 67 |
+
- [Block 3 Tensor Group : ~218M Elements](#block-3-tensor-group--218m-elements)
|
| 68 |
+
- [Block 4 Tensor Group : ~218M Elements](#block-4-tensor-group--218m-elements)
|
| 69 |
+
- [Block 5 Tensor Group : ~218M Elements](#block-5-tensor-group--218m-elements)
|
| 70 |
+
- [Block 6 Tensor Group : ~218M Elements](#block-6-tensor-group--218m-elements)
|
| 71 |
+
- [Block 7 Tensor Group : ~218M Elements](#block-7-tensor-group--218m-elements)
|
| 72 |
+
- [Block 8 Tensor Group : ~218M Elements](#block-8-tensor-group--218m-elements)
|
| 73 |
+
- [Block 9 Tensor Group : ~218M Elements](#block-9-tensor-group--218m-elements)
|
| 74 |
+
- [Block 10 Tensor Group : ~218M Elements](#block-10-tensor-group--218m-elements)
|
| 75 |
+
- [Block 11 Tensor Group : ~218M Elements](#block-11-tensor-group--218m-elements)
|
| 76 |
+
- [Block 12 Tensor Group : ~218M Elements](#block-12-tensor-group--218m-elements)
|
| 77 |
+
- [Block 13 Tensor Group : ~218M Elements](#block-13-tensor-group--218m-elements)
|
| 78 |
+
- [Block 14 Tensor Group : ~218M Elements](#block-14-tensor-group--218m-elements)
|
| 79 |
+
- [Block 15 Tensor Group : ~218M Elements](#block-15-tensor-group--218m-elements)
|
| 80 |
+
- [Block 16 Tensor Group : ~218M Elements](#block-16-tensor-group--218m-elements)
|
| 81 |
+
- [Block 17 Tensor Group : ~218M Elements](#block-17-tensor-group--218m-elements)
|
| 82 |
+
- [Block 18 Tensor Group : ~218M Elements](#block-18-tensor-group--218m-elements)
|
| 83 |
+
- [Block 19 Tensor Group : ~218M Elements](#block-19-tensor-group--218m-elements)
|
| 84 |
+
- [Block 20 Tensor Group : ~218M Elements](#block-20-tensor-group--218m-elements)
|
| 85 |
+
- [Block 21 Tensor Group : ~218M Elements](#block-21-tensor-group--218m-elements)
|
| 86 |
+
- [Block 22 Tensor Group : ~218M Elements](#block-22-tensor-group--218m-elements)
|
| 87 |
+
- [Block 23 Tensor Group : ~218M Elements](#block-23-tensor-group--218m-elements)
|
| 88 |
+
- [Block 24 Tensor Group : ~218M Elements](#block-24-tensor-group--218m-elements)
|
| 89 |
+
- [Block 25 Tensor Group : ~218M Elements](#block-25-tensor-group--218m-elements)
|
| 90 |
+
- [Block 26 Tensor Group : ~218M Elements](#block-26-tensor-group--218m-elements)
|
| 91 |
+
- [Block 27 Tensor Group : ~218M Elements](#block-27-tensor-group--218m-elements)
|
| 92 |
+
- [Block 28 Tensor Group : ~218M Elements](#block-28-tensor-group--218m-elements)
|
| 93 |
+
- [Block 29 Tensor Group : ~218M Elements](#block-29-tensor-group--218m-elements)
|
| 94 |
+
- [Block 30 Tensor Group : ~218M Elements](#block-30-tensor-group--218m-elements)
|
| 95 |
+
- [Block 31 Tensor Group : ~218M Elements](#block-31-tensor-group--218m-elements)
|
| 96 |
+
|
| 97 |
+
### Tensor Data Offset
|
| 98 |
+
|
| 99 |
+
This table contains the offset and data segment relative to start of file
|
| 100 |
+
|
| 101 |
+
| T_ID | Tensor Layer Name | Data Offset (B) | Data Size (B) |
|
| 102 |
+
|-----:|:--------------------------|-----------------:|-----------------:|
|
| 103 |
+
| 0 | output.weight | 0x779620 | 0xd746000 |
|
| 104 |
+
| 1 | output_norm.weight | 0xdebf620 | 0x4000 |
|
| 105 |
+
| 2 | rope_freqs.weight | 0xdec3620 | 0x100 |
|
| 106 |
+
| 3 | token_embd.weight | 0xdec3720 | 0xd746000 |
|
| 107 |
+
| 4 | blk.0.attn_k.weight | 0x1b609720 | 0x188000 |
|
| 108 |
+
| 5 | blk.0.attn_norm.weight | 0x1b791720 | 0x4000 |
|
| 109 |
+
| 6 | blk.0.attn_output.weight | 0x1b795720 | 0x900000 |
|
| 110 |
+
| 7 | blk.0.attn_q.weight | 0x1c095720 | 0x620000 |
|
| 111 |
+
| 8 | blk.0.attn_v.weight | 0x1c6b5720 | 0x1b8000 |
|
| 112 |
+
| 9 | blk.0.ffn_down.weight | 0x1c86d720 | 0x1f80000 |
|
| 113 |
+
| 10 | blk.0.ffn_gate.weight | 0x1e7ed720 | 0x1570000 |
|
| 114 |
+
| 11 | blk.0.ffn_norm.weight | 0x1fd5d720 | 0x4000 |
|
| 115 |
+
| 12 | blk.0.ffn_up.weight | 0x1fd61720 | 0x1570000 |
|
| 116 |
+
| 13 | blk.1.attn_k.weight | 0x212d1720 | 0x188000 |
|
| 117 |
+
| 14 | blk.1.attn_norm.weight | 0x21459720 | 0x4000 |
|
| 118 |
+
| 15 | blk.1.attn_output.weight | 0x2145d720 | 0x900000 |
|
| 119 |
+
| 16 | blk.1.attn_q.weight | 0x21d5d720 | 0x620000 |
|
| 120 |
+
| 17 | blk.1.attn_v.weight | 0x2237d720 | 0x1b8000 |
|
| 121 |
+
| 18 | blk.1.ffn_down.weight | 0x22535720 | 0x1f80000 |
|
| 122 |
+
| 19 | blk.1.ffn_gate.weight | 0x244b5720 | 0x1570000 |
|
| 123 |
+
| 20 | blk.1.ffn_norm.weight | 0x25a25720 | 0x4000 |
|
| 124 |
+
| 21 | blk.1.ffn_up.weight | 0x25a29720 | 0x1570000 |
|
| 125 |
+
| 22 | blk.2.attn_k.weight | 0x26f99720 | 0x188000 |
|
| 126 |
+
| 23 | blk.2.attn_norm.weight | 0x27121720 | 0x4000 |
|
| 127 |
+
| 24 | blk.2.attn_output.weight | 0x27125720 | 0x900000 |
|
| 128 |
+
| 25 | blk.2.attn_q.weight | 0x27a25720 | 0x620000 |
|
| 129 |
+
| 26 | blk.2.attn_v.weight | 0x28045720 | 0x1b8000 |
|
| 130 |
+
| 27 | blk.2.ffn_down.weight | 0x281fd720 | 0x1f80000 |
|
| 131 |
+
| 28 | blk.2.ffn_gate.weight | 0x2a17d720 | 0x1570000 |
|
| 132 |
+
| 29 | blk.2.ffn_norm.weight | 0x2b6ed720 | 0x4000 |
|
| 133 |
+
| 30 | blk.2.ffn_up.weight | 0x2b6f1720 | 0x1570000 |
|
| 134 |
+
| 31 | blk.3.attn_k.weight | 0x2cc61720 | 0x188000 |
|
| 135 |
+
| 32 | blk.3.attn_norm.weight | 0x2cde9720 | 0x4000 |
|
| 136 |
+
| 33 | blk.3.attn_output.weight | 0x2cded720 | 0x900000 |
|
| 137 |
+
| 34 | blk.3.attn_q.weight | 0x2d6ed720 | 0x620000 |
|
| 138 |
+
| 35 | blk.3.attn_v.weight | 0x2dd0d720 | 0x1b8000 |
|
| 139 |
+
| 36 | blk.3.ffn_down.weight | 0x2dec5720 | 0x1f80000 |
|
| 140 |
+
| 37 | blk.3.ffn_gate.weight | 0x2fe45720 | 0x1570000 |
|
| 141 |
+
| 38 | blk.3.ffn_norm.weight | 0x313b5720 | 0x4000 |
|
| 142 |
+
| 39 | blk.3.ffn_up.weight | 0x313b9720 | 0x1570000 |
|
| 143 |
+
| 40 | blk.4.attn_k.weight | 0x32929720 | 0x188000 |
|
| 144 |
+
| 41 | blk.4.attn_norm.weight | 0x32ab1720 | 0x4000 |
|
| 145 |
+
| 42 | blk.4.attn_output.weight | 0x32ab5720 | 0x900000 |
|
| 146 |
+
| 43 | blk.4.attn_q.weight | 0x333b5720 | 0x620000 |
|
| 147 |
+
| 44 | blk.4.attn_v.weight | 0x339d5720 | 0x1b8000 |
|
| 148 |
+
| 45 | blk.4.ffn_down.weight | 0x33b8d720 | 0x1f80000 |
|
| 149 |
+
| 46 | blk.4.ffn_gate.weight | 0x35b0d720 | 0x1570000 |
|
| 150 |
+
| 47 | blk.4.ffn_norm.weight | 0x3707d720 | 0x4000 |
|
| 151 |
+
| 48 | blk.4.ffn_up.weight | 0x37081720 | 0x1570000 |
|
| 152 |
+
| 49 | blk.5.attn_k.weight | 0x385f1720 | 0x188000 |
|
| 153 |
+
| 50 | blk.5.attn_norm.weight | 0x38779720 | 0x4000 |
|
| 154 |
+
| 51 | blk.5.attn_output.weight | 0x3877d720 | 0x900000 |
|
| 155 |
+
| 52 | blk.5.attn_q.weight | 0x3907d720 | 0x620000 |
|
| 156 |
+
| 53 | blk.5.attn_v.weight | 0x3969d720 | 0x1b8000 |
|
| 157 |
+
| 54 | blk.5.ffn_down.weight | 0x39855720 | 0x1f80000 |
|
| 158 |
+
| 55 | blk.5.ffn_gate.weight | 0x3b7d5720 | 0x1570000 |
|
| 159 |
+
| 56 | blk.5.ffn_norm.weight | 0x3cd45720 | 0x4000 |
|
| 160 |
+
| 57 | blk.5.ffn_up.weight | 0x3cd49720 | 0x1570000 |
|
| 161 |
+
| 58 | blk.6.attn_k.weight | 0x3e2b9720 | 0x188000 |
|
| 162 |
+
| 59 | blk.6.attn_norm.weight | 0x3e441720 | 0x4000 |
|
| 163 |
+
| 60 | blk.6.attn_output.weight | 0x3e445720 | 0x900000 |
|
| 164 |
+
| 61 | blk.6.attn_q.weight | 0x3ed45720 | 0x620000 |
|
| 165 |
+
| 62 | blk.6.attn_v.weight | 0x3f365720 | 0x1b8000 |
|
| 166 |
+
| 63 | blk.6.ffn_down.weight | 0x3f51d720 | 0x1f80000 |
|
| 167 |
+
| 64 | blk.6.ffn_gate.weight | 0x4149d720 | 0x1570000 |
|
| 168 |
+
| 65 | blk.6.ffn_norm.weight | 0x42a0d720 | 0x4000 |
|
| 169 |
+
| 66 | blk.6.ffn_up.weight | 0x42a11720 | 0x1570000 |
|
| 170 |
+
| 67 | blk.7.attn_k.weight | 0x43f81720 | 0x188000 |
|
| 171 |
+
| 68 | blk.7.attn_norm.weight | 0x44109720 | 0x4000 |
|
| 172 |
+
| 69 | blk.7.attn_output.weight | 0x4410d720 | 0x900000 |
|
| 173 |
+
| 70 | blk.7.attn_q.weight | 0x44a0d720 | 0x620000 |
|
| 174 |
+
| 71 | blk.7.attn_v.weight | 0x4502d720 | 0x1b8000 |
|
| 175 |
+
| 72 | blk.7.ffn_down.weight | 0x451e5720 | 0x1f80000 |
|
| 176 |
+
| 73 | blk.7.ffn_gate.weight | 0x47165720 | 0x1570000 |
|
| 177 |
+
| 74 | blk.7.ffn_norm.weight | 0x486d5720 | 0x4000 |
|
| 178 |
+
| 75 | blk.7.ffn_up.weight | 0x486d9720 | 0x1570000 |
|
| 179 |
+
| 76 | blk.8.attn_k.weight | 0x49c49720 | 0x188000 |
|
| 180 |
+
| 77 | blk.8.attn_norm.weight | 0x49dd1720 | 0x4000 |
|
| 181 |
+
| 78 | blk.8.attn_output.weight | 0x49dd5720 | 0x900000 |
|
| 182 |
+
| 79 | blk.8.attn_q.weight | 0x4a6d5720 | 0x620000 |
|
| 183 |
+
| 80 | blk.8.attn_v.weight | 0x4acf5720 | 0x1b8000 |
|
| 184 |
+
| 81 | blk.8.ffn_down.weight | 0x4aead720 | 0x1f80000 |
|
| 185 |
+
| 82 | blk.8.ffn_gate.weight | 0x4ce2d720 | 0x1570000 |
|
| 186 |
+
| 83 | blk.8.ffn_norm.weight | 0x4e39d720 | 0x4000 |
|
| 187 |
+
| 84 | blk.8.ffn_up.weight | 0x4e3a1720 | 0x1570000 |
|
| 188 |
+
| 85 | blk.9.attn_k.weight | 0x4f911720 | 0x188000 |
|
| 189 |
+
| 86 | blk.9.attn_norm.weight | 0x4fa99720 | 0x4000 |
|
| 190 |
+
| 87 | blk.9.attn_output.weight | 0x4fa9d720 | 0x900000 |
|
| 191 |
+
| 88 | blk.9.attn_q.weight | 0x5039d720 | 0x620000 |
|
| 192 |
+
| 89 | blk.9.attn_v.weight | 0x509bd720 | 0x1b8000 |
|
| 193 |
+
| 90 | blk.9.ffn_down.weight | 0x50b75720 | 0x1f80000 |
|
| 194 |
+
| 91 | blk.9.ffn_gate.weight | 0x52af5720 | 0x1570000 |
|
| 195 |
+
| 92 | blk.9.ffn_norm.weight | 0x54065720 | 0x4000 |
|
| 196 |
+
| 93 | blk.9.ffn_up.weight | 0x54069720 | 0x1570000 |
|
| 197 |
+
| 94 | blk.10.attn_k.weight | 0x555d9720 | 0x188000 |
|
| 198 |
+
| 95 | blk.10.attn_norm.weight | 0x55761720 | 0x4000 |
|
| 199 |
+
| 96 | blk.10.attn_output.weight | 0x55765720 | 0x900000 |
|
| 200 |
+
| 97 | blk.10.attn_q.weight | 0x56065720 | 0x620000 |
|
| 201 |
+
| 98 | blk.10.attn_v.weight | 0x56685720 | 0x1b8000 |
|
| 202 |
+
| 99 | blk.10.ffn_down.weight | 0x5683d720 | 0x1f80000 |
|
| 203 |
+
| 100 | blk.10.ffn_gate.weight | 0x587bd720 | 0x1570000 |
|
| 204 |
+
| 101 | blk.10.ffn_norm.weight | 0x59d2d720 | 0x4000 |
|
| 205 |
+
| 102 | blk.10.ffn_up.weight | 0x59d31720 | 0x1570000 |
|
| 206 |
+
| 103 | blk.11.attn_k.weight | 0x5b2a1720 | 0x188000 |
|
| 207 |
+
| 104 | blk.11.attn_norm.weight | 0x5b429720 | 0x4000 |
|
| 208 |
+
| 105 | blk.11.attn_output.weight | 0x5b42d720 | 0x900000 |
|
| 209 |
+
| 106 | blk.11.attn_q.weight | 0x5bd2d720 | 0x620000 |
|
| 210 |
+
| 107 | blk.11.attn_v.weight | 0x5c34d720 | 0x1b8000 |
|
| 211 |
+
| 108 | blk.11.ffn_down.weight | 0x5c505720 | 0x1f80000 |
|
| 212 |
+
| 109 | blk.11.ffn_gate.weight | 0x5e485720 | 0x1570000 |
|
| 213 |
+
| 110 | blk.11.ffn_norm.weight | 0x5f9f5720 | 0x4000 |
|
| 214 |
+
| 111 | blk.11.ffn_up.weight | 0x5f9f9720 | 0x1570000 |
|
| 215 |
+
| 112 | blk.12.attn_k.weight | 0x60f69720 | 0x188000 |
|
| 216 |
+
| 113 | blk.12.attn_norm.weight | 0x610f1720 | 0x4000 |
|
| 217 |
+
| 114 | blk.12.attn_output.weight | 0x610f5720 | 0x900000 |
|
| 218 |
+
| 115 | blk.12.attn_q.weight | 0x619f5720 | 0x620000 |
|
| 219 |
+
| 116 | blk.12.attn_v.weight | 0x62015720 | 0x1b8000 |
|
| 220 |
+
| 117 | blk.12.ffn_down.weight | 0x621cd720 | 0x1f80000 |
|
| 221 |
+
| 118 | blk.12.ffn_gate.weight | 0x6414d720 | 0x1570000 |
|
| 222 |
+
| 119 | blk.12.ffn_norm.weight | 0x656bd720 | 0x4000 |
|
| 223 |
+
| 120 | blk.12.ffn_up.weight | 0x656c1720 | 0x1570000 |
|
| 224 |
+
| 121 | blk.13.attn_k.weight | 0x66c31720 | 0x1b8000 |
|
| 225 |
+
| 122 | blk.13.attn_norm.weight | 0x66de9720 | 0x4000 |
|
| 226 |
+
| 123 | blk.13.attn_output.weight | 0x66ded720 | 0x900000 |
|
| 227 |
+
| 124 | blk.13.attn_q.weight | 0x676ed720 | 0x6e0000 |
|
| 228 |
+
| 125 | blk.13.attn_v.weight | 0x67dcd720 | 0x240000 |
|
| 229 |
+
| 126 | blk.13.ffn_down.weight | 0x6800d720 | 0x1f80000 |
|
| 230 |
+
| 127 | blk.13.ffn_gate.weight | 0x69f8d720 | 0x1570000 |
|
| 231 |
+
| 128 | blk.13.ffn_norm.weight | 0x6b4fd720 | 0x4000 |
|
| 232 |
+
| 129 | blk.13.ffn_up.weight | 0x6b501720 | 0x1570000 |
|
| 233 |
+
| 130 | blk.14.attn_k.weight | 0x6ca71720 | 0x1b8000 |
|
| 234 |
+
| 131 | blk.14.attn_norm.weight | 0x6cc29720 | 0x4000 |
|
| 235 |
+
| 132 | blk.14.attn_output.weight | 0x6cc2d720 | 0x900000 |
|
| 236 |
+
| 133 | blk.14.attn_q.weight | 0x6d52d720 | 0x6e0000 |
|
| 237 |
+
| 134 | blk.14.attn_v.weight | 0x6dc0d720 | 0x240000 |
|
| 238 |
+
| 135 | blk.14.ffn_down.weight | 0x6de4d720 | 0x1f80000 |
|
| 239 |
+
| 136 | blk.14.ffn_gate.weight | 0x6fdcd720 | 0x1570000 |
|
| 240 |
+
| 137 | blk.14.ffn_norm.weight | 0x7133d720 | 0x4000 |
|
| 241 |
+
| 138 | blk.14.ffn_up.weight | 0x71341720 | 0x1570000 |
|
| 242 |
+
| 139 | blk.15.attn_k.weight | 0x728b1720 | 0x188000 |
|
| 243 |
+
| 140 | blk.15.attn_norm.weight | 0x72a39720 | 0x4000 |
|
| 244 |
+
| 141 | blk.15.attn_output.weight | 0x72a3d720 | 0x900000 |
|
| 245 |
+
| 142 | blk.15.attn_q.weight | 0x7333d720 | 0x620000 |
|
| 246 |
+
| 143 | blk.15.attn_v.weight | 0x7395d720 | 0x1b8000 |
|
| 247 |
+
| 144 | blk.15.ffn_down.weight | 0x73b15720 | 0x1f80000 |
|
| 248 |
+
| 145 | blk.15.ffn_gate.weight | 0x75a95720 | 0x1570000 |
|
| 249 |
+
| 146 | blk.15.ffn_norm.weight | 0x77005720 | 0x4000 |
|
| 250 |
+
| 147 | blk.15.ffn_up.weight | 0x77009720 | 0x1570000 |
|
| 251 |
+
| 148 | blk.16.attn_k.weight | 0x78579720 | 0x1b8000 |
|
| 252 |
+
| 149 | blk.16.attn_norm.weight | 0x78731720 | 0x4000 |
|
| 253 |
+
| 150 | blk.16.attn_output.weight | 0x78735720 | 0x900000 |
|
| 254 |
+
| 151 | blk.16.attn_q.weight | 0x79035720 | 0x6e0000 |
|
| 255 |
+
| 152 | blk.16.attn_v.weight | 0x79715720 | 0x240000 |
|
| 256 |
+
| 153 | blk.16.ffn_down.weight | 0x79955720 | 0x1f80000 |
|
| 257 |
+
| 154 | blk.16.ffn_gate.weight | 0x7b8d5720 | 0x1810000 |
|
| 258 |
+
| 155 | blk.16.ffn_norm.weight | 0x7d0e5720 | 0x4000 |
|
| 259 |
+
| 156 | blk.16.ffn_up.weight | 0x7d0e9720 | 0x1810000 |
|
| 260 |
+
| 157 | blk.17.attn_k.weight | 0x7e8f9720 | 0x188000 |
|
| 261 |
+
| 158 | blk.17.attn_norm.weight | 0x7ea81720 | 0x4000 |
|
| 262 |
+
| 159 | blk.17.attn_output.weight | 0x7ea85720 | 0x900000 |
|
| 263 |
+
| 160 | blk.17.attn_q.weight | 0x7f385720 | 0x620000 |
|
| 264 |
+
| 161 | blk.17.attn_v.weight | 0x7f9a5720 | 0x1b8000 |
|
| 265 |
+
| 162 | blk.17.ffn_down.weight | 0x7fb5d720 | 0x1f80000 |
|
| 266 |
+
| 163 | blk.17.ffn_gate.weight | 0x81add720 | 0x1810000 |
|
| 267 |
+
| 164 | blk.17.ffn_norm.weight | 0x832ed720 | 0x4000 |
|
| 268 |
+
| 165 | blk.17.ffn_up.weight | 0x832f1720 | 0x1810000 |
|
| 269 |
+
| 166 | blk.18.attn_k.weight | 0x84b01720 | 0x1b8000 |
|
| 270 |
+
| 167 | blk.18.attn_norm.weight | 0x84cb9720 | 0x4000 |
|
| 271 |
+
| 168 | blk.18.attn_output.weight | 0x84cbd720 | 0x900000 |
|
| 272 |
+
| 169 | blk.18.attn_q.weight | 0x855bd720 | 0x6e0000 |
|
| 273 |
+
| 170 | blk.18.attn_v.weight | 0x85c9d720 | 0x240000 |
|
| 274 |
+
| 171 | blk.18.ffn_down.weight | 0x85edd720 | 0x1f80000 |
|
| 275 |
+
| 172 | blk.18.ffn_gate.weight | 0x87e5d720 | 0x1810000 |
|
| 276 |
+
| 173 | blk.18.ffn_norm.weight | 0x8966d720 | 0x4000 |
|
| 277 |
+
| 174 | blk.18.ffn_up.weight | 0x89671720 | 0x1810000 |
|
| 278 |
+
| 175 | blk.19.attn_k.weight | 0x8ae81720 | 0x1b8000 |
|
| 279 |
+
| 176 | blk.19.attn_norm.weight | 0x8b039720 | 0x4000 |
|
| 280 |
+
| 177 | blk.19.attn_output.weight | 0x8b03d720 | 0x900000 |
|
| 281 |
+
| 178 | blk.19.attn_q.weight | 0x8b93d720 | 0x6e0000 |
|
| 282 |
+
| 179 | blk.19.attn_v.weight | 0x8c01d720 | 0x240000 |
|
| 283 |
+
| 180 | blk.19.ffn_down.weight | 0x8c25d720 | 0x1f80000 |
|
| 284 |
+
| 181 | blk.19.ffn_gate.weight | 0x8e1dd720 | 0x1810000 |
|
| 285 |
+
| 182 | blk.19.ffn_norm.weight | 0x8f9ed720 | 0x4000 |
|
| 286 |
+
| 183 | blk.19.ffn_up.weight | 0x8f9f1720 | 0x1810000 |
|
| 287 |
+
| 184 | blk.20.attn_k.weight | 0x91201720 | 0x1b8000 |
|
| 288 |
+
| 185 | blk.20.attn_norm.weight | 0x913b9720 | 0x4000 |
|
| 289 |
+
| 186 | blk.20.attn_output.weight | 0x913bd720 | 0x900000 |
|
| 290 |
+
| 187 | blk.20.attn_q.weight | 0x91cbd720 | 0x6e0000 |
|
| 291 |
+
| 188 | blk.20.attn_v.weight | 0x9239d720 | 0x240000 |
|
| 292 |
+
| 189 | blk.20.ffn_down.weight | 0x925dd720 | 0x1f80000 |
|
| 293 |
+
| 190 | blk.20.ffn_gate.weight | 0x9455d720 | 0x1810000 |
|
| 294 |
+
| 191 | blk.20.ffn_norm.weight | 0x95d6d720 | 0x4000 |
|
| 295 |
+
| 192 | blk.20.ffn_up.weight | 0x95d71720 | 0x1810000 |
|
| 296 |
+
| 193 | blk.21.attn_k.weight | 0x97581720 | 0x1b8000 |
|
| 297 |
+
| 194 | blk.21.attn_norm.weight | 0x97739720 | 0x4000 |
|
| 298 |
+
| 195 | blk.21.attn_output.weight | 0x9773d720 | 0x900000 |
|
| 299 |
+
| 196 | blk.21.attn_q.weight | 0x9803d720 | 0x6e0000 |
|
| 300 |
+
| 197 | blk.21.attn_v.weight | 0x9871d720 | 0x240000 |
|
| 301 |
+
| 198 | blk.21.ffn_down.weight | 0x9895d720 | 0x1f80000 |
|
| 302 |
+
| 199 | blk.21.ffn_gate.weight | 0x9a8dd720 | 0x1810000 |
|
| 303 |
+
| 200 | blk.21.ffn_norm.weight | 0x9c0ed720 | 0x4000 |
|
| 304 |
+
| 201 | blk.21.ffn_up.weight | 0x9c0f1720 | 0x1810000 |
|
| 305 |
+
| 202 | blk.22.attn_k.weight | 0x9d901720 | 0x1b8000 |
|
| 306 |
+
| 203 | blk.22.attn_norm.weight | 0x9dab9720 | 0x4000 |
|
| 307 |
+
| 204 | blk.22.attn_output.weight | 0x9dabd720 | 0x900000 |
|
| 308 |
+
| 205 | blk.22.attn_q.weight | 0x9e3bd720 | 0x6e0000 |
|
| 309 |
+
| 206 | blk.22.attn_v.weight | 0x9ea9d720 | 0x240000 |
|
| 310 |
+
| 207 | blk.22.ffn_down.weight | 0x9ecdd720 | 0x1f80000 |
|
| 311 |
+
| 208 | blk.22.ffn_gate.weight | 0xa0c5d720 | 0x1810000 |
|
| 312 |
+
| 209 | blk.22.ffn_norm.weight | 0xa246d720 | 0x4000 |
|
| 313 |
+
| 210 | blk.22.ffn_up.weight | 0xa2471720 | 0x1810000 |
|
| 314 |
+
| 211 | blk.23.attn_k.weight | 0xa3c81720 | 0x1b8000 |
|
| 315 |
+
| 212 | blk.23.attn_norm.weight | 0xa3e39720 | 0x4000 |
|
| 316 |
+
| 213 | blk.23.attn_output.weight | 0xa3e3d720 | 0x900000 |
|
| 317 |
+
| 214 | blk.23.attn_q.weight | 0xa473d720 | 0x6e0000 |
|
| 318 |
+
| 215 | blk.23.attn_v.weight | 0xa4e1d720 | 0x240000 |
|
| 319 |
+
| 216 | blk.23.ffn_down.weight | 0xa505d720 | 0x1f80000 |
|
| 320 |
+
| 217 | blk.23.ffn_gate.weight | 0xa6fdd720 | 0x1810000 |
|
| 321 |
+
| 218 | blk.23.ffn_norm.weight | 0xa87ed720 | 0x4000 |
|
| 322 |
+
| 219 | blk.23.ffn_up.weight | 0xa87f1720 | 0x1810000 |
|
| 323 |
+
| 220 | blk.24.attn_k.weight | 0xaa001720 | 0x1b8000 |
|
| 324 |
+
| 221 | blk.24.attn_norm.weight | 0xaa1b9720 | 0x4000 |
|
| 325 |
+
| 222 | blk.24.attn_output.weight | 0xaa1bd720 | 0x900000 |
|
| 326 |
+
| 223 | blk.24.attn_q.weight | 0xaaabd720 | 0x6e0000 |
|
| 327 |
+
| 224 | blk.24.attn_v.weight | 0xab19d720 | 0x240000 |
|
| 328 |
+
| 225 | blk.24.ffn_down.weight | 0xab3dd720 | 0x1f80000 |
|
| 329 |
+
| 226 | blk.24.ffn_gate.weight | 0xad35d720 | 0x1810000 |
|
| 330 |
+
| 227 | blk.24.ffn_norm.weight | 0xaeb6d720 | 0x4000 |
|
| 331 |
+
| 228 | blk.24.ffn_up.weight | 0xaeb71720 | 0x1810000 |
|
| 332 |
+
| 229 | blk.25.attn_k.weight | 0xb0381720 | 0x1b8000 |
|
| 333 |
+
| 230 | blk.25.attn_norm.weight | 0xb0539720 | 0x4000 |
|
| 334 |
+
| 231 | blk.25.attn_output.weight | 0xb053d720 | 0x900000 |
|
| 335 |
+
| 232 | blk.25.attn_q.weight | 0xb0e3d720 | 0x6e0000 |
|
| 336 |
+
| 233 | blk.25.attn_v.weight | 0xb151d720 | 0x240000 |
|
| 337 |
+
| 234 | blk.25.ffn_down.weight | 0xb175d720 | 0x1f80000 |
|
| 338 |
+
| 235 | blk.25.ffn_gate.weight | 0xb36dd720 | 0x1810000 |
|
| 339 |
+
| 236 | blk.25.ffn_norm.weight | 0xb4eed720 | 0x4000 |
|
| 340 |
+
| 237 | blk.25.ffn_up.weight | 0xb4ef1720 | 0x1810000 |
|
| 341 |
+
| 238 | blk.26.attn_k.weight | 0xb6701720 | 0x1b8000 |
|
| 342 |
+
| 239 | blk.26.attn_norm.weight | 0xb68b9720 | 0x4000 |
|
| 343 |
+
| 240 | blk.26.attn_output.weight | 0xb68bd720 | 0x900000 |
|
| 344 |
+
| 241 | blk.26.attn_q.weight | 0xb71bd720 | 0x6e0000 |
|
| 345 |
+
| 242 | blk.26.attn_v.weight | 0xb789d720 | 0x240000 |
|
| 346 |
+
| 243 | blk.26.ffn_down.weight | 0xb7add720 | 0x1f80000 |
|
| 347 |
+
| 244 | blk.26.ffn_gate.weight | 0xb9a5d720 | 0x1810000 |
|
| 348 |
+
| 245 | blk.26.ffn_norm.weight | 0xbb26d720 | 0x4000 |
|
| 349 |
+
| 246 | blk.26.ffn_up.weight | 0xbb271720 | 0x1810000 |
|
| 350 |
+
| 247 | blk.27.attn_k.weight | 0xbca81720 | 0x1b8000 |
|
| 351 |
+
| 248 | blk.27.attn_norm.weight | 0xbcc39720 | 0x4000 |
|
| 352 |
+
| 249 | blk.27.attn_output.weight | 0xbcc3d720 | 0x900000 |
|
| 353 |
+
| 250 | blk.27.attn_q.weight | 0xbd53d720 | 0x6e0000 |
|
| 354 |
+
| 251 | blk.27.attn_v.weight | 0xbdc1d720 | 0x240000 |
|
| 355 |
+
| 252 | blk.27.ffn_down.weight | 0xbde5d720 | 0x1f80000 |
|
| 356 |
+
| 253 | blk.27.ffn_gate.weight | 0xbfddd720 | 0x1810000 |
|
| 357 |
+
| 254 | blk.27.ffn_norm.weight | 0xc15ed720 | 0x4000 |
|
| 358 |
+
| 255 | blk.27.ffn_up.weight | 0xc15f1720 | 0x1810000 |
|
| 359 |
+
| 256 | blk.28.attn_k.weight | 0xc2e01720 | 0x1b8000 |
|
| 360 |
+
| 257 | blk.28.attn_norm.weight | 0xc2fb9720 | 0x4000 |
|
| 361 |
+
| 258 | blk.28.attn_output.weight | 0xc2fbd720 | 0x900000 |
|
| 362 |
+
| 259 | blk.28.attn_q.weight | 0xc38bd720 | 0x6e0000 |
|
| 363 |
+
| 260 | blk.28.attn_v.weight | 0xc3f9d720 | 0x240000 |
|
| 364 |
+
| 261 | blk.28.ffn_down.weight | 0xc41dd720 | 0x1f80000 |
|
| 365 |
+
| 262 | blk.28.ffn_gate.weight | 0xc615d720 | 0x1810000 |
|
| 366 |
+
| 263 | blk.28.ffn_norm.weight | 0xc796d720 | 0x4000 |
|
| 367 |
+
| 264 | blk.28.ffn_up.weight | 0xc7971720 | 0x1810000 |
|
| 368 |
+
| 265 | blk.29.attn_k.weight | 0xc9181720 | 0x1b8000 |
|
| 369 |
+
| 266 | blk.29.attn_norm.weight | 0xc9339720 | 0x4000 |
|
| 370 |
+
| 267 | blk.29.attn_output.weight | 0xc933d720 | 0x900000 |
|
| 371 |
+
| 268 | blk.29.attn_q.weight | 0xc9c3d720 | 0x6e0000 |
|
| 372 |
+
| 269 | blk.29.attn_v.weight | 0xca31d720 | 0x240000 |
|
| 373 |
+
| 270 | blk.29.ffn_down.weight | 0xca55d720 | 0x1f80000 |
|
| 374 |
+
| 271 | blk.29.ffn_gate.weight | 0xcc4dd720 | 0x1810000 |
|
| 375 |
+
| 272 | blk.29.ffn_norm.weight | 0xcdced720 | 0x4000 |
|
| 376 |
+
| 273 | blk.29.ffn_up.weight | 0xcdcf1720 | 0x1810000 |
|
| 377 |
+
| 274 | blk.30.attn_k.weight | 0xcf501720 | 0x1b8000 |
|
| 378 |
+
| 275 | blk.30.attn_norm.weight | 0xcf6b9720 | 0x4000 |
|
| 379 |
+
| 276 | blk.30.attn_output.weight | 0xcf6bd720 | 0x900000 |
|
| 380 |
+
| 277 | blk.30.attn_q.weight | 0xcffbd720 | 0x6e0000 |
|
| 381 |
+
| 278 | blk.30.attn_v.weight | 0xd069d720 | 0x240000 |
|
| 382 |
+
| 279 | blk.30.ffn_down.weight | 0xd08dd720 | 0x1f80000 |
|
| 383 |
+
| 280 | blk.30.ffn_gate.weight | 0xd285d720 | 0x1810000 |
|
| 384 |
+
| 281 | blk.30.ffn_norm.weight | 0xd406d720 | 0x4000 |
|
| 385 |
+
| 282 | blk.30.ffn_up.weight | 0xd4071720 | 0x1810000 |
|
| 386 |
+
| 283 | blk.31.attn_k.weight | 0xd5881720 | 0x188000 |
|
| 387 |
+
| 284 | blk.31.attn_norm.weight | 0xd5a09720 | 0x4000 |
|
| 388 |
+
| 285 | blk.31.attn_output.weight | 0xd5a0d720 | 0x900000 |
|
| 389 |
+
| 286 | blk.31.attn_q.weight | 0xd630d720 | 0x620000 |
|
| 390 |
+
| 287 | blk.31.attn_v.weight | 0xd692d720 | 0x1b8000 |
|
| 391 |
+
| 288 | blk.31.ffn_down.weight | 0xd6ae5720 | 0x1f80000 |
|
| 392 |
+
| 289 | blk.31.ffn_gate.weight | 0xd8a65720 | 0x1810000 |
|
| 393 |
+
| 290 | blk.31.ffn_norm.weight | 0xda275720 | 0x4000 |
|
| 394 |
+
| 291 | blk.31.ffn_up.weight | 0xda279720 | 0x1810000 |
|
| 395 |
+
|
| 396 |
+
### <a name="base">Base Tensor Group : ~1B Elements</a>
|
| 397 |
+
|
| 398 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 399 |
+
|-----:|:-------------------|:---------------------------------|:------------------|:----------------------|:------|
|
| 400 |
+
| 0 | output.weight | Output (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | IQ3_S |
|
| 401 |
+
| 1 | output_norm.weight | Output Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 402 |
+
| 2 | rope_freqs.weight | Rope_Freqs (W) | ( 64) 64 | 64 x 1 x 1 x 1 | F32 |
|
| 403 |
+
| 3 | token_embd.weight | Token Embedding (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | IQ3_S |
|
| 404 |
+
|
| 405 |
+
- Total elements in base: ( ~1B) 1050677312
|
| 406 |
+
- Percentage of total elements: 13.08%
|
| 407 |
+
|
| 408 |
+
|
| 409 |
+
### <a name="blk_0">Block 0 Tensor Group : ~218M Elements</a>
|
| 410 |
+
|
| 411 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 412 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:--------|
|
| 413 |
+
| 4 | blk.0.attn_k.weight | Block 0 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 414 |
+
| 5 | blk.0.attn_norm.weight | Block 0 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 415 |
+
| 6 | blk.0.attn_output.weight | Block 0 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 416 |
+
| 7 | blk.0.attn_q.weight | Block 0 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 417 |
+
| 8 | blk.0.attn_v.weight | Block 0 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 418 |
+
| 9 | blk.0.ffn_down.weight | Block 0 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 419 |
+
| 10 | blk.0.ffn_gate.weight | Block 0 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 420 |
+
| 11 | blk.0.ffn_norm.weight | Block 0 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 421 |
+
| 12 | blk.0.ffn_up.weight | Block 0 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 422 |
+
|
| 423 |
+
- Total elements in blk.0: (~218M) 218112000
|
| 424 |
+
- Percentage of total elements: 2.72%
|
| 425 |
+
|
| 426 |
+
|
| 427 |
+
### <a name="blk_1">Block 1 Tensor Group : ~218M Elements</a>
|
| 428 |
+
|
| 429 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 430 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:--------|
|
| 431 |
+
| 13 | blk.1.attn_k.weight | Block 1 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 432 |
+
| 14 | blk.1.attn_norm.weight | Block 1 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 433 |
+
| 15 | blk.1.attn_output.weight | Block 1 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 434 |
+
| 16 | blk.1.attn_q.weight | Block 1 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 435 |
+
| 17 | blk.1.attn_v.weight | Block 1 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 436 |
+
| 18 | blk.1.ffn_down.weight | Block 1 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 437 |
+
| 19 | blk.1.ffn_gate.weight | Block 1 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 438 |
+
| 20 | blk.1.ffn_norm.weight | Block 1 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 439 |
+
| 21 | blk.1.ffn_up.weight | Block 1 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 440 |
+
|
| 441 |
+
- Total elements in blk.1: (~218M) 218112000
|
| 442 |
+
- Percentage of total elements: 2.72%
|
| 443 |
+
|
| 444 |
+
|
| 445 |
+
### <a name="blk_2">Block 2 Tensor Group : ~218M Elements</a>
|
| 446 |
+
|
| 447 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 448 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:--------|
|
| 449 |
+
| 22 | blk.2.attn_k.weight | Block 2 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 450 |
+
| 23 | blk.2.attn_norm.weight | Block 2 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 451 |
+
| 24 | blk.2.attn_output.weight | Block 2 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 452 |
+
| 25 | blk.2.attn_q.weight | Block 2 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 453 |
+
| 26 | blk.2.attn_v.weight | Block 2 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 454 |
+
| 27 | blk.2.ffn_down.weight | Block 2 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 455 |
+
| 28 | blk.2.ffn_gate.weight | Block 2 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 456 |
+
| 29 | blk.2.ffn_norm.weight | Block 2 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 457 |
+
| 30 | blk.2.ffn_up.weight | Block 2 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 458 |
+
|
| 459 |
+
- Total elements in blk.2: (~218M) 218112000
|
| 460 |
+
- Percentage of total elements: 2.72%
|
| 461 |
+
|
| 462 |
+
|
| 463 |
+
### <a name="blk_3">Block 3 Tensor Group : ~218M Elements</a>
|
| 464 |
+
|
| 465 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 466 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:--------|
|
| 467 |
+
| 31 | blk.3.attn_k.weight | Block 3 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 468 |
+
| 32 | blk.3.attn_norm.weight | Block 3 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 469 |
+
| 33 | blk.3.attn_output.weight | Block 3 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 470 |
+
| 34 | blk.3.attn_q.weight | Block 3 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 471 |
+
| 35 | blk.3.attn_v.weight | Block 3 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 472 |
+
| 36 | blk.3.ffn_down.weight | Block 3 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 473 |
+
| 37 | blk.3.ffn_gate.weight | Block 3 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 474 |
+
| 38 | blk.3.ffn_norm.weight | Block 3 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 475 |
+
| 39 | blk.3.ffn_up.weight | Block 3 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 476 |
+
|
| 477 |
+
- Total elements in blk.3: (~218M) 218112000
|
| 478 |
+
- Percentage of total elements: 2.72%
|
| 479 |
+
|
| 480 |
+
|
| 481 |
+
### <a name="blk_4">Block 4 Tensor Group : ~218M Elements</a>
|
| 482 |
+
|
| 483 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 484 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:--------|
|
| 485 |
+
| 40 | blk.4.attn_k.weight | Block 4 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 486 |
+
| 41 | blk.4.attn_norm.weight | Block 4 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 487 |
+
| 42 | blk.4.attn_output.weight | Block 4 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 488 |
+
| 43 | blk.4.attn_q.weight | Block 4 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 489 |
+
| 44 | blk.4.attn_v.weight | Block 4 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 490 |
+
| 45 | blk.4.ffn_down.weight | Block 4 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 491 |
+
| 46 | blk.4.ffn_gate.weight | Block 4 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 492 |
+
| 47 | blk.4.ffn_norm.weight | Block 4 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 493 |
+
| 48 | blk.4.ffn_up.weight | Block 4 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 494 |
+
|
| 495 |
+
- Total elements in blk.4: (~218M) 218112000
|
| 496 |
+
- Percentage of total elements: 2.72%
|
| 497 |
+
|
| 498 |
+
|
| 499 |
+
### <a name="blk_5">Block 5 Tensor Group : ~218M Elements</a>
|
| 500 |
+
|
| 501 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 502 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:--------|
|
| 503 |
+
| 49 | blk.5.attn_k.weight | Block 5 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 504 |
+
| 50 | blk.5.attn_norm.weight | Block 5 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 505 |
+
| 51 | blk.5.attn_output.weight | Block 5 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 506 |
+
| 52 | blk.5.attn_q.weight | Block 5 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 507 |
+
| 53 | blk.5.attn_v.weight | Block 5 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 508 |
+
| 54 | blk.5.ffn_down.weight | Block 5 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 509 |
+
| 55 | blk.5.ffn_gate.weight | Block 5 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 510 |
+
| 56 | blk.5.ffn_norm.weight | Block 5 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 511 |
+
| 57 | blk.5.ffn_up.weight | Block 5 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 512 |
+
|
| 513 |
+
- Total elements in blk.5: (~218M) 218112000
|
| 514 |
+
- Percentage of total elements: 2.72%
|
| 515 |
+
|
| 516 |
+
|
| 517 |
+
### <a name="blk_6">Block 6 Tensor Group : ~218M Elements</a>
|
| 518 |
+
|
| 519 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 520 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:--------|
|
| 521 |
+
| 58 | blk.6.attn_k.weight | Block 6 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 522 |
+
| 59 | blk.6.attn_norm.weight | Block 6 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 523 |
+
| 60 | blk.6.attn_output.weight | Block 6 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 524 |
+
| 61 | blk.6.attn_q.weight | Block 6 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 525 |
+
| 62 | blk.6.attn_v.weight | Block 6 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 526 |
+
| 63 | blk.6.ffn_down.weight | Block 6 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 527 |
+
| 64 | blk.6.ffn_gate.weight | Block 6 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 528 |
+
| 65 | blk.6.ffn_norm.weight | Block 6 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 529 |
+
| 66 | blk.6.ffn_up.weight | Block 6 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 530 |
+
|
| 531 |
+
- Total elements in blk.6: (~218M) 218112000
|
| 532 |
+
- Percentage of total elements: 2.72%
|
| 533 |
+
|
| 534 |
+
|
| 535 |
+
### <a name="blk_7">Block 7 Tensor Group : ~218M Elements</a>
|
| 536 |
+
|
| 537 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 538 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:--------|
|
| 539 |
+
| 67 | blk.7.attn_k.weight | Block 7 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 540 |
+
| 68 | blk.7.attn_norm.weight | Block 7 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 541 |
+
| 69 | blk.7.attn_output.weight | Block 7 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 542 |
+
| 70 | blk.7.attn_q.weight | Block 7 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 543 |
+
| 71 | blk.7.attn_v.weight | Block 7 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 544 |
+
| 72 | blk.7.ffn_down.weight | Block 7 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 545 |
+
| 73 | blk.7.ffn_gate.weight | Block 7 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 546 |
+
| 74 | blk.7.ffn_norm.weight | Block 7 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 547 |
+
| 75 | blk.7.ffn_up.weight | Block 7 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 548 |
+
|
| 549 |
+
- Total elements in blk.7: (~218M) 218112000
|
| 550 |
+
- Percentage of total elements: 2.72%
|
| 551 |
+
|
| 552 |
+
|
| 553 |
+
### <a name="blk_8">Block 8 Tensor Group : ~218M Elements</a>
|
| 554 |
+
|
| 555 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 556 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:--------|
|
| 557 |
+
| 76 | blk.8.attn_k.weight | Block 8 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 558 |
+
| 77 | blk.8.attn_norm.weight | Block 8 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 559 |
+
| 78 | blk.8.attn_output.weight | Block 8 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 560 |
+
| 79 | blk.8.attn_q.weight | Block 8 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 561 |
+
| 80 | blk.8.attn_v.weight | Block 8 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 562 |
+
| 81 | blk.8.ffn_down.weight | Block 8 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 563 |
+
| 82 | blk.8.ffn_gate.weight | Block 8 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 564 |
+
| 83 | blk.8.ffn_norm.weight | Block 8 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 565 |
+
| 84 | blk.8.ffn_up.weight | Block 8 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 566 |
+
|
| 567 |
+
- Total elements in blk.8: (~218M) 218112000
|
| 568 |
+
- Percentage of total elements: 2.72%
|
| 569 |
+
|
| 570 |
+
|
| 571 |
+
### <a name="blk_9">Block 9 Tensor Group : ~218M Elements</a>
|
| 572 |
+
|
| 573 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 574 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:--------|
|
| 575 |
+
| 85 | blk.9.attn_k.weight | Block 9 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 576 |
+
| 86 | blk.9.attn_norm.weight | Block 9 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 577 |
+
| 87 | blk.9.attn_output.weight | Block 9 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 578 |
+
| 88 | blk.9.attn_q.weight | Block 9 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 579 |
+
| 89 | blk.9.attn_v.weight | Block 9 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 580 |
+
| 90 | blk.9.ffn_down.weight | Block 9 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 581 |
+
| 91 | blk.9.ffn_gate.weight | Block 9 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 582 |
+
| 92 | blk.9.ffn_norm.weight | Block 9 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 583 |
+
| 93 | blk.9.ffn_up.weight | Block 9 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 584 |
+
|
| 585 |
+
- Total elements in blk.9: (~218M) 218112000
|
| 586 |
+
- Percentage of total elements: 2.72%
|
| 587 |
+
|
| 588 |
+
|
| 589 |
+
### <a name="blk_10">Block 10 Tensor Group : ~218M Elements</a>
|
| 590 |
+
|
| 591 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 592 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:--------|
|
| 593 |
+
| 94 | blk.10.attn_k.weight | Block 10 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 594 |
+
| 95 | blk.10.attn_norm.weight | Block 10 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 595 |
+
| 96 | blk.10.attn_output.weight | Block 10 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 596 |
+
| 97 | blk.10.attn_q.weight | Block 10 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 597 |
+
| 98 | blk.10.attn_v.weight | Block 10 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 598 |
+
| 99 | blk.10.ffn_down.weight | Block 10 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 599 |
+
| 100 | blk.10.ffn_gate.weight | Block 10 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 600 |
+
| 101 | blk.10.ffn_norm.weight | Block 10 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 601 |
+
| 102 | blk.10.ffn_up.weight | Block 10 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 602 |
+
|
| 603 |
+
- Total elements in blk.10: (~218M) 218112000
|
| 604 |
+
- Percentage of total elements: 2.72%
|
| 605 |
+
|
| 606 |
+
|
| 607 |
+
### <a name="blk_11">Block 11 Tensor Group : ~218M Elements</a>
|
| 608 |
+
|
| 609 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 610 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:--------|
|
| 611 |
+
| 103 | blk.11.attn_k.weight | Block 11 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 612 |
+
| 104 | blk.11.attn_norm.weight | Block 11 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 613 |
+
| 105 | blk.11.attn_output.weight | Block 11 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 614 |
+
| 106 | blk.11.attn_q.weight | Block 11 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 615 |
+
| 107 | blk.11.attn_v.weight | Block 11 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 616 |
+
| 108 | blk.11.ffn_down.weight | Block 11 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 617 |
+
| 109 | blk.11.ffn_gate.weight | Block 11 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 618 |
+
| 110 | blk.11.ffn_norm.weight | Block 11 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 619 |
+
| 111 | blk.11.ffn_up.weight | Block 11 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 620 |
+
|
| 621 |
+
- Total elements in blk.11: (~218M) 218112000
|
| 622 |
+
- Percentage of total elements: 2.72%
|
| 623 |
+
|
| 624 |
+
|
| 625 |
+
### <a name="blk_12">Block 12 Tensor Group : ~218M Elements</a>
|
| 626 |
+
|
| 627 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 628 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:--------|
|
| 629 |
+
| 112 | blk.12.attn_k.weight | Block 12 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 630 |
+
| 113 | blk.12.attn_norm.weight | Block 12 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 631 |
+
| 114 | blk.12.attn_output.weight | Block 12 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 632 |
+
| 115 | blk.12.attn_q.weight | Block 12 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 633 |
+
| 116 | blk.12.attn_v.weight | Block 12 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 634 |
+
| 117 | blk.12.ffn_down.weight | Block 12 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 635 |
+
| 118 | blk.12.ffn_gate.weight | Block 12 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 636 |
+
| 119 | blk.12.ffn_norm.weight | Block 12 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 637 |
+
| 120 | blk.12.ffn_up.weight | Block 12 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 638 |
+
|
| 639 |
+
- Total elements in blk.12: (~218M) 218112000
|
| 640 |
+
- Percentage of total elements: 2.72%
|
| 641 |
+
|
| 642 |
+
|
| 643 |
+
### <a name="blk_13">Block 13 Tensor Group : ~218M Elements</a>
|
| 644 |
+
|
| 645 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 646 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:--------|
|
| 647 |
+
| 121 | blk.13.attn_k.weight | Block 13 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 648 |
+
| 122 | blk.13.attn_norm.weight | Block 13 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 649 |
+
| 123 | blk.13.attn_output.weight | Block 13 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 650 |
+
| 124 | blk.13.attn_q.weight | Block 13 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 651 |
+
| 125 | blk.13.attn_v.weight | Block 13 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 652 |
+
| 126 | blk.13.ffn_down.weight | Block 13 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 653 |
+
| 127 | blk.13.ffn_gate.weight | Block 13 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 654 |
+
| 128 | blk.13.ffn_norm.weight | Block 13 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 655 |
+
| 129 | blk.13.ffn_up.weight | Block 13 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 656 |
+
|
| 657 |
+
- Total elements in blk.13: (~218M) 218112000
|
| 658 |
+
- Percentage of total elements: 2.72%
|
| 659 |
+
|
| 660 |
+
|
| 661 |
+
### <a name="blk_14">Block 14 Tensor Group : ~218M Elements</a>
|
| 662 |
+
|
| 663 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 664 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:--------|
|
| 665 |
+
| 130 | blk.14.attn_k.weight | Block 14 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 666 |
+
| 131 | blk.14.attn_norm.weight | Block 14 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 667 |
+
| 132 | blk.14.attn_output.weight | Block 14 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 668 |
+
| 133 | blk.14.attn_q.weight | Block 14 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 669 |
+
| 134 | blk.14.attn_v.weight | Block 14 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 670 |
+
| 135 | blk.14.ffn_down.weight | Block 14 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 671 |
+
| 136 | blk.14.ffn_gate.weight | Block 14 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 672 |
+
| 137 | blk.14.ffn_norm.weight | Block 14 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 673 |
+
| 138 | blk.14.ffn_up.weight | Block 14 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 674 |
+
|
| 675 |
+
- Total elements in blk.14: (~218M) 218112000
|
| 676 |
+
- Percentage of total elements: 2.72%
|
| 677 |
+
|
| 678 |
+
|
| 679 |
+
### <a name="blk_15">Block 15 Tensor Group : ~218M Elements</a>
|
| 680 |
+
|
| 681 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 682 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:--------|
|
| 683 |
+
| 139 | blk.15.attn_k.weight | Block 15 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 684 |
+
| 140 | blk.15.attn_norm.weight | Block 15 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 685 |
+
| 141 | blk.15.attn_output.weight | Block 15 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 686 |
+
| 142 | blk.15.attn_q.weight | Block 15 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 687 |
+
| 143 | blk.15.attn_v.weight | Block 15 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 688 |
+
| 144 | blk.15.ffn_down.weight | Block 15 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 689 |
+
| 145 | blk.15.ffn_gate.weight | Block 15 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 690 |
+
| 146 | blk.15.ffn_norm.weight | Block 15 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 691 |
+
| 147 | blk.15.ffn_up.weight | Block 15 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 692 |
+
|
| 693 |
+
- Total elements in blk.15: (~218M) 218112000
|
| 694 |
+
- Percentage of total elements: 2.72%
|
| 695 |
+
|
| 696 |
+
|
| 697 |
+
### <a name="blk_16">Block 16 Tensor Group : ~218M Elements</a>
|
| 698 |
+
|
| 699 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 700 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 701 |
+
| 148 | blk.16.attn_k.weight | Block 16 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 702 |
+
| 149 | blk.16.attn_norm.weight | Block 16 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 703 |
+
| 150 | blk.16.attn_output.weight | Block 16 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 704 |
+
| 151 | blk.16.attn_q.weight | Block 16 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 705 |
+
| 152 | blk.16.attn_v.weight | Block 16 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 706 |
+
| 153 | blk.16.ffn_down.weight | Block 16 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 707 |
+
| 154 | blk.16.ffn_gate.weight | Block 16 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 708 |
+
| 155 | blk.16.ffn_norm.weight | Block 16 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 709 |
+
| 156 | blk.16.ffn_up.weight | Block 16 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 710 |
+
|
| 711 |
+
- Total elements in blk.16: (~218M) 218112000
|
| 712 |
+
- Percentage of total elements: 2.72%
|
| 713 |
+
|
| 714 |
+
|
| 715 |
+
### <a name="blk_17">Block 17 Tensor Group : ~218M Elements</a>
|
| 716 |
+
|
| 717 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 718 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:--------|
|
| 719 |
+
| 157 | blk.17.attn_k.weight | Block 17 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 720 |
+
| 158 | blk.17.attn_norm.weight | Block 17 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 721 |
+
| 159 | blk.17.attn_output.weight | Block 17 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 722 |
+
| 160 | blk.17.attn_q.weight | Block 17 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 723 |
+
| 161 | blk.17.attn_v.weight | Block 17 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 724 |
+
| 162 | blk.17.ffn_down.weight | Block 17 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 725 |
+
| 163 | blk.17.ffn_gate.weight | Block 17 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 726 |
+
| 164 | blk.17.ffn_norm.weight | Block 17 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 727 |
+
| 165 | blk.17.ffn_up.weight | Block 17 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 728 |
+
|
| 729 |
+
- Total elements in blk.17: (~218M) 218112000
|
| 730 |
+
- Percentage of total elements: 2.72%
|
| 731 |
+
|
| 732 |
+
|
| 733 |
+
### <a name="blk_18">Block 18 Tensor Group : ~218M Elements</a>
|
| 734 |
+
|
| 735 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 736 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 737 |
+
| 166 | blk.18.attn_k.weight | Block 18 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 738 |
+
| 167 | blk.18.attn_norm.weight | Block 18 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 739 |
+
| 168 | blk.18.attn_output.weight | Block 18 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 740 |
+
| 169 | blk.18.attn_q.weight | Block 18 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 741 |
+
| 170 | blk.18.attn_v.weight | Block 18 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 742 |
+
| 171 | blk.18.ffn_down.weight | Block 18 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 743 |
+
| 172 | blk.18.ffn_gate.weight | Block 18 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 744 |
+
| 173 | blk.18.ffn_norm.weight | Block 18 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 745 |
+
| 174 | blk.18.ffn_up.weight | Block 18 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 746 |
+
|
| 747 |
+
- Total elements in blk.18: (~218M) 218112000
|
| 748 |
+
- Percentage of total elements: 2.72%
|
| 749 |
+
|
| 750 |
+
|
| 751 |
+
### <a name="blk_19">Block 19 Tensor Group : ~218M Elements</a>
|
| 752 |
+
|
| 753 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 754 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 755 |
+
| 175 | blk.19.attn_k.weight | Block 19 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 756 |
+
| 176 | blk.19.attn_norm.weight | Block 19 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 757 |
+
| 177 | blk.19.attn_output.weight | Block 19 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 758 |
+
| 178 | blk.19.attn_q.weight | Block 19 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 759 |
+
| 179 | blk.19.attn_v.weight | Block 19 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 760 |
+
| 180 | blk.19.ffn_down.weight | Block 19 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 761 |
+
| 181 | blk.19.ffn_gate.weight | Block 19 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 762 |
+
| 182 | blk.19.ffn_norm.weight | Block 19 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 763 |
+
| 183 | blk.19.ffn_up.weight | Block 19 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 764 |
+
|
| 765 |
+
- Total elements in blk.19: (~218M) 218112000
|
| 766 |
+
- Percentage of total elements: 2.72%
|
| 767 |
+
|
| 768 |
+
|
| 769 |
+
### <a name="blk_20">Block 20 Tensor Group : ~218M Elements</a>
|
| 770 |
+
|
| 771 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 772 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 773 |
+
| 184 | blk.20.attn_k.weight | Block 20 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 774 |
+
| 185 | blk.20.attn_norm.weight | Block 20 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 775 |
+
| 186 | blk.20.attn_output.weight | Block 20 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 776 |
+
| 187 | blk.20.attn_q.weight | Block 20 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 777 |
+
| 188 | blk.20.attn_v.weight | Block 20 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 778 |
+
| 189 | blk.20.ffn_down.weight | Block 20 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 779 |
+
| 190 | blk.20.ffn_gate.weight | Block 20 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 780 |
+
| 191 | blk.20.ffn_norm.weight | Block 20 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 781 |
+
| 192 | blk.20.ffn_up.weight | Block 20 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 782 |
+
|
| 783 |
+
- Total elements in blk.20: (~218M) 218112000
|
| 784 |
+
- Percentage of total elements: 2.72%
|
| 785 |
+
|
| 786 |
+
|
| 787 |
+
### <a name="blk_21">Block 21 Tensor Group : ~218M Elements</a>
|
| 788 |
+
|
| 789 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 790 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 791 |
+
| 193 | blk.21.attn_k.weight | Block 21 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 792 |
+
| 194 | blk.21.attn_norm.weight | Block 21 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 793 |
+
| 195 | blk.21.attn_output.weight | Block 21 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 794 |
+
| 196 | blk.21.attn_q.weight | Block 21 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 795 |
+
| 197 | blk.21.attn_v.weight | Block 21 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 796 |
+
| 198 | blk.21.ffn_down.weight | Block 21 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 797 |
+
| 199 | blk.21.ffn_gate.weight | Block 21 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 798 |
+
| 200 | blk.21.ffn_norm.weight | Block 21 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 799 |
+
| 201 | blk.21.ffn_up.weight | Block 21 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 800 |
+
|
| 801 |
+
- Total elements in blk.21: (~218M) 218112000
|
| 802 |
+
- Percentage of total elements: 2.72%
|
| 803 |
+
|
| 804 |
+
|
| 805 |
+
### <a name="blk_22">Block 22 Tensor Group : ~218M Elements</a>
|
| 806 |
+
|
| 807 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 808 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 809 |
+
| 202 | blk.22.attn_k.weight | Block 22 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 810 |
+
| 203 | blk.22.attn_norm.weight | Block 22 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 811 |
+
| 204 | blk.22.attn_output.weight | Block 22 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 812 |
+
| 205 | blk.22.attn_q.weight | Block 22 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 813 |
+
| 206 | blk.22.attn_v.weight | Block 22 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 814 |
+
| 207 | blk.22.ffn_down.weight | Block 22 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 815 |
+
| 208 | blk.22.ffn_gate.weight | Block 22 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 816 |
+
| 209 | blk.22.ffn_norm.weight | Block 22 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 817 |
+
| 210 | blk.22.ffn_up.weight | Block 22 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 818 |
+
|
| 819 |
+
- Total elements in blk.22: (~218M) 218112000
|
| 820 |
+
- Percentage of total elements: 2.72%
|
| 821 |
+
|
| 822 |
+
|
| 823 |
+
### <a name="blk_23">Block 23 Tensor Group : ~218M Elements</a>
|
| 824 |
+
|
| 825 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 826 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 827 |
+
| 211 | blk.23.attn_k.weight | Block 23 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 828 |
+
| 212 | blk.23.attn_norm.weight | Block 23 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 829 |
+
| 213 | blk.23.attn_output.weight | Block 23 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 830 |
+
| 214 | blk.23.attn_q.weight | Block 23 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 831 |
+
| 215 | blk.23.attn_v.weight | Block 23 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 832 |
+
| 216 | blk.23.ffn_down.weight | Block 23 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 833 |
+
| 217 | blk.23.ffn_gate.weight | Block 23 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 834 |
+
| 218 | blk.23.ffn_norm.weight | Block 23 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 835 |
+
| 219 | blk.23.ffn_up.weight | Block 23 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 836 |
+
|
| 837 |
+
- Total elements in blk.23: (~218M) 218112000
|
| 838 |
+
- Percentage of total elements: 2.72%
|
| 839 |
+
|
| 840 |
+
|
| 841 |
+
### <a name="blk_24">Block 24 Tensor Group : ~218M Elements</a>
|
| 842 |
+
|
| 843 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 844 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 845 |
+
| 220 | blk.24.attn_k.weight | Block 24 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 846 |
+
| 221 | blk.24.attn_norm.weight | Block 24 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 847 |
+
| 222 | blk.24.attn_output.weight | Block 24 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 848 |
+
| 223 | blk.24.attn_q.weight | Block 24 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 849 |
+
| 224 | blk.24.attn_v.weight | Block 24 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 850 |
+
| 225 | blk.24.ffn_down.weight | Block 24 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 851 |
+
| 226 | blk.24.ffn_gate.weight | Block 24 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 852 |
+
| 227 | blk.24.ffn_norm.weight | Block 24 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 853 |
+
| 228 | blk.24.ffn_up.weight | Block 24 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 854 |
+
|
| 855 |
+
- Total elements in blk.24: (~218M) 218112000
|
| 856 |
+
- Percentage of total elements: 2.72%
|
| 857 |
+
|
| 858 |
+
|
| 859 |
+
### <a name="blk_25">Block 25 Tensor Group : ~218M Elements</a>
|
| 860 |
+
|
| 861 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 862 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 863 |
+
| 229 | blk.25.attn_k.weight | Block 25 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 864 |
+
| 230 | blk.25.attn_norm.weight | Block 25 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 865 |
+
| 231 | blk.25.attn_output.weight | Block 25 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 866 |
+
| 232 | blk.25.attn_q.weight | Block 25 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 867 |
+
| 233 | blk.25.attn_v.weight | Block 25 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 868 |
+
| 234 | blk.25.ffn_down.weight | Block 25 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 869 |
+
| 235 | blk.25.ffn_gate.weight | Block 25 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 870 |
+
| 236 | blk.25.ffn_norm.weight | Block 25 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 871 |
+
| 237 | blk.25.ffn_up.weight | Block 25 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 872 |
+
|
| 873 |
+
- Total elements in blk.25: (~218M) 218112000
|
| 874 |
+
- Percentage of total elements: 2.72%
|
| 875 |
+
|
| 876 |
+
|
| 877 |
+
### <a name="blk_26">Block 26 Tensor Group : ~218M Elements</a>
|
| 878 |
+
|
| 879 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 880 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 881 |
+
| 238 | blk.26.attn_k.weight | Block 26 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 882 |
+
| 239 | blk.26.attn_norm.weight | Block 26 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 883 |
+
| 240 | blk.26.attn_output.weight | Block 26 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 884 |
+
| 241 | blk.26.attn_q.weight | Block 26 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 885 |
+
| 242 | blk.26.attn_v.weight | Block 26 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 886 |
+
| 243 | blk.26.ffn_down.weight | Block 26 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 887 |
+
| 244 | blk.26.ffn_gate.weight | Block 26 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 888 |
+
| 245 | blk.26.ffn_norm.weight | Block 26 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 889 |
+
| 246 | blk.26.ffn_up.weight | Block 26 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 890 |
+
|
| 891 |
+
- Total elements in blk.26: (~218M) 218112000
|
| 892 |
+
- Percentage of total elements: 2.72%
|
| 893 |
+
|
| 894 |
+
|
| 895 |
+
### <a name="blk_27">Block 27 Tensor Group : ~218M Elements</a>
|
| 896 |
+
|
| 897 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 898 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 899 |
+
| 247 | blk.27.attn_k.weight | Block 27 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 900 |
+
| 248 | blk.27.attn_norm.weight | Block 27 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 901 |
+
| 249 | blk.27.attn_output.weight | Block 27 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 902 |
+
| 250 | blk.27.attn_q.weight | Block 27 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 903 |
+
| 251 | blk.27.attn_v.weight | Block 27 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 904 |
+
| 252 | blk.27.ffn_down.weight | Block 27 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 905 |
+
| 253 | blk.27.ffn_gate.weight | Block 27 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 906 |
+
| 254 | blk.27.ffn_norm.weight | Block 27 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 907 |
+
| 255 | blk.27.ffn_up.weight | Block 27 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 908 |
+
|
| 909 |
+
- Total elements in blk.27: (~218M) 218112000
|
| 910 |
+
- Percentage of total elements: 2.72%
|
| 911 |
+
|
| 912 |
+
|
| 913 |
+
### <a name="blk_28">Block 28 Tensor Group : ~218M Elements</a>
|
| 914 |
+
|
| 915 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 916 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 917 |
+
| 256 | blk.28.attn_k.weight | Block 28 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 918 |
+
| 257 | blk.28.attn_norm.weight | Block 28 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 919 |
+
| 258 | blk.28.attn_output.weight | Block 28 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 920 |
+
| 259 | blk.28.attn_q.weight | Block 28 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 921 |
+
| 260 | blk.28.attn_v.weight | Block 28 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 922 |
+
| 261 | blk.28.ffn_down.weight | Block 28 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 923 |
+
| 262 | blk.28.ffn_gate.weight | Block 28 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 924 |
+
| 263 | blk.28.ffn_norm.weight | Block 28 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 925 |
+
| 264 | blk.28.ffn_up.weight | Block 28 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 926 |
+
|
| 927 |
+
- Total elements in blk.28: (~218M) 218112000
|
| 928 |
+
- Percentage of total elements: 2.72%
|
| 929 |
+
|
| 930 |
+
|
| 931 |
+
### <a name="blk_29">Block 29 Tensor Group : ~218M Elements</a>
|
| 932 |
+
|
| 933 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 934 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 935 |
+
| 265 | blk.29.attn_k.weight | Block 29 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 936 |
+
| 266 | blk.29.attn_norm.weight | Block 29 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 937 |
+
| 267 | blk.29.attn_output.weight | Block 29 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 938 |
+
| 268 | blk.29.attn_q.weight | Block 29 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 939 |
+
| 269 | blk.29.attn_v.weight | Block 29 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 940 |
+
| 270 | blk.29.ffn_down.weight | Block 29 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 941 |
+
| 271 | blk.29.ffn_gate.weight | Block 29 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 942 |
+
| 272 | blk.29.ffn_norm.weight | Block 29 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 943 |
+
| 273 | blk.29.ffn_up.weight | Block 29 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 944 |
+
|
| 945 |
+
- Total elements in blk.29: (~218M) 218112000
|
| 946 |
+
- Percentage of total elements: 2.72%
|
| 947 |
+
|
| 948 |
+
|
| 949 |
+
### <a name="blk_30">Block 30 Tensor Group : ~218M Elements</a>
|
| 950 |
+
|
| 951 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 952 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 953 |
+
| 274 | blk.30.attn_k.weight | Block 30 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 954 |
+
| 275 | blk.30.attn_norm.weight | Block 30 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 955 |
+
| 276 | blk.30.attn_output.weight | Block 30 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 956 |
+
| 277 | blk.30.attn_q.weight | Block 30 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 957 |
+
| 278 | blk.30.attn_v.weight | Block 30 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 958 |
+
| 279 | blk.30.ffn_down.weight | Block 30 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 959 |
+
| 280 | blk.30.ffn_gate.weight | Block 30 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 960 |
+
| 281 | blk.30.ffn_norm.weight | Block 30 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 961 |
+
| 282 | blk.30.ffn_up.weight | Block 30 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 962 |
+
|
| 963 |
+
- Total elements in blk.30: (~218M) 218112000
|
| 964 |
+
- Percentage of total elements: 2.72%
|
| 965 |
+
|
| 966 |
+
|
| 967 |
+
### <a name="blk_31">Block 31 Tensor Group : ~218M Elements</a>
|
| 968 |
+
|
| 969 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 970 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:--------|
|
| 971 |
+
| 283 | blk.31.attn_k.weight | Block 31 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 972 |
+
| 284 | blk.31.attn_norm.weight | Block 31 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 973 |
+
| 285 | blk.31.attn_output.weight | Block 31 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 974 |
+
| 286 | blk.31.attn_q.weight | Block 31 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 975 |
+
| 287 | blk.31.attn_v.weight | Block 31 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 976 |
+
| 288 | blk.31.ffn_down.weight | Block 31 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 977 |
+
| 289 | blk.31.ffn_gate.weight | Block 31 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 978 |
+
| 290 | blk.31.ffn_norm.weight | Block 31 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 979 |
+
| 291 | blk.31.ffn_up.weight | Block 31 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 980 |
+
|
| 981 |
+
- Total elements in blk.31: (~218M) 218112000
|
| 982 |
+
- Percentage of total elements: 2.72%
|
scores/Watt-Tool-8B-IQ3_S.md
ADDED
|
@@ -0,0 +1,982 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Watt-Tool-8B-IQ3_S.gguf - GGUF Internal File Dump
|
| 2 |
+
|
| 3 |
+
- Endian: LITTLE endian
|
| 4 |
+
|
| 5 |
+
## Key Value Metadata Store
|
| 6 |
+
|
| 7 |
+
There are 43 key-value pairs in this file
|
| 8 |
+
|
| 9 |
+
| POS | TYPE | Count | Key | Value |
|
| 10 |
+
|----:|:---------|-------:|:---------------------------------------|:--------------------------------------------------------------------|
|
| 11 |
+
| 1 | UINT32 | 1 | GGUF.version | 3 |
|
| 12 |
+
| 2 | UINT64 | 1 | GGUF.tensor_count | 292 |
|
| 13 |
+
| 3 | UINT64 | 1 | GGUF.kv_count | 40 |
|
| 14 |
+
| 4 | STRING | 1 | general.architecture | `llama` |
|
| 15 |
+
| 5 | STRING | 1 | general.type | `model` |
|
| 16 |
+
| 6 | STRING | 1 | general.name | `Watt Tool 8B GGUF` |
|
| 17 |
+
| 7 | STRING | 1 | general.finetune | `GGUF` |
|
| 18 |
+
| 8 | STRING | 1 | general.basename | `Watt-Tool` |
|
| 19 |
+
| 9 | STRING | 1 | general.size_label | `8B` |
|
| 20 |
+
| 10 | STRING | 1 | general.license | `apache-2.0` |
|
| 21 |
+
| 11 | UINT32 | 1 | general.base_model.count | 1 |
|
| 22 |
+
| 12 | STRING | 1 | general.base_model.0.name | `Llama 3.1 8B Instruct` |
|
| 23 |
+
| 13 | STRING | 1 | general.base_model.0.organization | `Meta Llama` |
|
| 24 |
+
| 14 | STRING | 1 | general.base_model.0.repo_url | `https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct` |
|
| 25 |
+
| 15 | [STRING] | 4 | general.tags | [ `function-calling`, `tool-use`, `llama`, `bfcl` ] |
|
| 26 |
+
| 16 | [STRING] | 1 | general.languages | [ `en` ] |
|
| 27 |
+
| 17 | UINT32 | 1 | llama.block_count | 32 |
|
| 28 |
+
| 18 | UINT32 | 1 | llama.context_length | 131072 |
|
| 29 |
+
| 19 | UINT32 | 1 | llama.embedding_length | 4096 |
|
| 30 |
+
| 20 | UINT32 | 1 | llama.feed_forward_length | 14336 |
|
| 31 |
+
| 21 | UINT32 | 1 | llama.attention.head_count | 32 |
|
| 32 |
+
| 22 | UINT32 | 1 | llama.attention.head_count_kv | 8 |
|
| 33 |
+
| 23 | FLOAT32 | 1 | llama.rope.freq_base | 500000.0 |
|
| 34 |
+
| 24 | FLOAT32 | 1 | llama.attention.layer_norm_rms_epsilon | 1e-05 |
|
| 35 |
+
| 25 | UINT32 | 1 | llama.attention.key_length | 128 |
|
| 36 |
+
| 26 | UINT32 | 1 | llama.attention.value_length | 128 |
|
| 37 |
+
| 27 | UINT32 | 1 | llama.vocab_size | 128256 |
|
| 38 |
+
| 28 | UINT32 | 1 | llama.rope.dimension_count | 128 |
|
| 39 |
+
| 29 | STRING | 1 | tokenizer.ggml.model | `gpt2` |
|
| 40 |
+
| 30 | STRING | 1 | tokenizer.ggml.pre | `llama-bpe` |
|
| 41 |
+
| 31 | [STRING] | 128256 | tokenizer.ggml.tokens | [ `!`, `"`, `#`, `$`, `%`, ... ] |
|
| 42 |
+
| 32 | [INT32] | 128256 | tokenizer.ggml.token_type | [ 1, 1, 1, 1, 1, 1, 1, ... ] |
|
| 43 |
+
| 33 | [STRING] | 280147 | tokenizer.ggml.merges | [ `Ġ Ġ`, `Ġ ĠĠĠ`, `ĠĠ ĠĠ`, `ĠĠĠ Ġ`, `i n`, ... ] |
|
| 44 |
+
| 34 | UINT32 | 1 | tokenizer.ggml.bos_token_id | 128000 |
|
| 45 |
+
| 35 | UINT32 | 1 | tokenizer.ggml.eos_token_id | 128009 |
|
| 46 |
+
| 36 | UINT32 | 1 | tokenizer.ggml.padding_token_id | 128009 |
|
| 47 |
+
| 37 | STRING | 1 | tokenizer.chat_template | `{{ '<|begin_of_text|>' }}{% if`...`d|>' }}{% endif %}{% endfor %}` |
|
| 48 |
+
| 38 | UINT32 | 1 | general.quantization_version | 2 |
|
| 49 |
+
| 39 | UINT32 | 1 | general.file_type | 26 |
|
| 50 |
+
| 40 | STRING | 1 | quantize.imatrix.file | `./imatrix/imatrix-Watt-Tool-8B-small.dat` |
|
| 51 |
+
| 41 | STRING | 1 | quantize.imatrix.dataset | `../../datasets/imatrix/calibration_eur_small.txt` |
|
| 52 |
+
| 42 | INT32 | 1 | quantize.imatrix.entries_count | 225 |
|
| 53 |
+
| 43 | INT32 | 1 | quantize.imatrix.chunks_count | 962 |
|
| 54 |
+
|
| 55 |
+
## Tensors Overview ~8B Elements
|
| 56 |
+
|
| 57 |
+
Total number of elements in all tensors: 8030261312 Elements
|
| 58 |
+
|
| 59 |
+
- [Watt-Tool-8B-IQ3\_S.gguf - GGUF Internal File Dump](#watt-tool-8b-iq3_sgguf---gguf-internal-file-dump)
|
| 60 |
+
- [Key Value Metadata Store](#key-value-metadata-store)
|
| 61 |
+
- [Tensors Overview ~8B Elements](#tensors-overview-8b-elements)
|
| 62 |
+
- [Tensor Data Offset](#tensor-data-offset)
|
| 63 |
+
- [Base Tensor Group : ~1B Elements](#base-tensor-group--1b-elements)
|
| 64 |
+
- [Block 0 Tensor Group : ~218M Elements](#block-0-tensor-group--218m-elements)
|
| 65 |
+
- [Block 1 Tensor Group : ~218M Elements](#block-1-tensor-group--218m-elements)
|
| 66 |
+
- [Block 2 Tensor Group : ~218M Elements](#block-2-tensor-group--218m-elements)
|
| 67 |
+
- [Block 3 Tensor Group : ~218M Elements](#block-3-tensor-group--218m-elements)
|
| 68 |
+
- [Block 4 Tensor Group : ~218M Elements](#block-4-tensor-group--218m-elements)
|
| 69 |
+
- [Block 5 Tensor Group : ~218M Elements](#block-5-tensor-group--218m-elements)
|
| 70 |
+
- [Block 6 Tensor Group : ~218M Elements](#block-6-tensor-group--218m-elements)
|
| 71 |
+
- [Block 7 Tensor Group : ~218M Elements](#block-7-tensor-group--218m-elements)
|
| 72 |
+
- [Block 8 Tensor Group : ~218M Elements](#block-8-tensor-group--218m-elements)
|
| 73 |
+
- [Block 9 Tensor Group : ~218M Elements](#block-9-tensor-group--218m-elements)
|
| 74 |
+
- [Block 10 Tensor Group : ~218M Elements](#block-10-tensor-group--218m-elements)
|
| 75 |
+
- [Block 11 Tensor Group : ~218M Elements](#block-11-tensor-group--218m-elements)
|
| 76 |
+
- [Block 12 Tensor Group : ~218M Elements](#block-12-tensor-group--218m-elements)
|
| 77 |
+
- [Block 13 Tensor Group : ~218M Elements](#block-13-tensor-group--218m-elements)
|
| 78 |
+
- [Block 14 Tensor Group : ~218M Elements](#block-14-tensor-group--218m-elements)
|
| 79 |
+
- [Block 15 Tensor Group : ~218M Elements](#block-15-tensor-group--218m-elements)
|
| 80 |
+
- [Block 16 Tensor Group : ~218M Elements](#block-16-tensor-group--218m-elements)
|
| 81 |
+
- [Block 17 Tensor Group : ~218M Elements](#block-17-tensor-group--218m-elements)
|
| 82 |
+
- [Block 18 Tensor Group : ~218M Elements](#block-18-tensor-group--218m-elements)
|
| 83 |
+
- [Block 19 Tensor Group : ~218M Elements](#block-19-tensor-group--218m-elements)
|
| 84 |
+
- [Block 20 Tensor Group : ~218M Elements](#block-20-tensor-group--218m-elements)
|
| 85 |
+
- [Block 21 Tensor Group : ~218M Elements](#block-21-tensor-group--218m-elements)
|
| 86 |
+
- [Block 22 Tensor Group : ~218M Elements](#block-22-tensor-group--218m-elements)
|
| 87 |
+
- [Block 23 Tensor Group : ~218M Elements](#block-23-tensor-group--218m-elements)
|
| 88 |
+
- [Block 24 Tensor Group : ~218M Elements](#block-24-tensor-group--218m-elements)
|
| 89 |
+
- [Block 25 Tensor Group : ~218M Elements](#block-25-tensor-group--218m-elements)
|
| 90 |
+
- [Block 26 Tensor Group : ~218M Elements](#block-26-tensor-group--218m-elements)
|
| 91 |
+
- [Block 27 Tensor Group : ~218M Elements](#block-27-tensor-group--218m-elements)
|
| 92 |
+
- [Block 28 Tensor Group : ~218M Elements](#block-28-tensor-group--218m-elements)
|
| 93 |
+
- [Block 29 Tensor Group : ~218M Elements](#block-29-tensor-group--218m-elements)
|
| 94 |
+
- [Block 30 Tensor Group : ~218M Elements](#block-30-tensor-group--218m-elements)
|
| 95 |
+
- [Block 31 Tensor Group : ~218M Elements](#block-31-tensor-group--218m-elements)
|
| 96 |
+
|
| 97 |
+
### Tensor Data Offset
|
| 98 |
+
|
| 99 |
+
This table contains the offset and data segment relative to start of file
|
| 100 |
+
|
| 101 |
+
| T_ID | Tensor Layer Name | Data Offset (B) | Data Size (B) |
|
| 102 |
+
|-----:|:--------------------------|-----------------:|-----------------:|
|
| 103 |
+
| 0 | output.weight | 0x779620 | 0xbfca000 |
|
| 104 |
+
| 1 | output_norm.weight | 0xc743620 | 0x4000 |
|
| 105 |
+
| 2 | rope_freqs.weight | 0xc747620 | 0x100 |
|
| 106 |
+
| 3 | token_embd.weight | 0xc747720 | 0xbfca000 |
|
| 107 |
+
| 4 | blk.0.attn_k.weight | 0x18711720 | 0x188000 |
|
| 108 |
+
| 5 | blk.0.attn_norm.weight | 0x18899720 | 0x4000 |
|
| 109 |
+
| 6 | blk.0.attn_output.weight | 0x1889d720 | 0x6e0000 |
|
| 110 |
+
| 7 | blk.0.attn_q.weight | 0x18f7d720 | 0x620000 |
|
| 111 |
+
| 8 | blk.0.attn_v.weight | 0x1959d720 | 0x1b8000 |
|
| 112 |
+
| 9 | blk.0.ffn_down.weight | 0x19755720 | 0x1810000 |
|
| 113 |
+
| 10 | blk.0.ffn_gate.weight | 0x1af65720 | 0x1570000 |
|
| 114 |
+
| 11 | blk.0.ffn_norm.weight | 0x1c4d5720 | 0x4000 |
|
| 115 |
+
| 12 | blk.0.ffn_up.weight | 0x1c4d9720 | 0x1570000 |
|
| 116 |
+
| 13 | blk.1.attn_k.weight | 0x1da49720 | 0x188000 |
|
| 117 |
+
| 14 | blk.1.attn_norm.weight | 0x1dbd1720 | 0x4000 |
|
| 118 |
+
| 15 | blk.1.attn_output.weight | 0x1dbd5720 | 0x6e0000 |
|
| 119 |
+
| 16 | blk.1.attn_q.weight | 0x1e2b5720 | 0x620000 |
|
| 120 |
+
| 17 | blk.1.attn_v.weight | 0x1e8d5720 | 0x1b8000 |
|
| 121 |
+
| 18 | blk.1.ffn_down.weight | 0x1ea8d720 | 0x1f80000 |
|
| 122 |
+
| 19 | blk.1.ffn_gate.weight | 0x20a0d720 | 0x1570000 |
|
| 123 |
+
| 20 | blk.1.ffn_norm.weight | 0x21f7d720 | 0x4000 |
|
| 124 |
+
| 21 | blk.1.ffn_up.weight | 0x21f81720 | 0x1570000 |
|
| 125 |
+
| 22 | blk.2.attn_k.weight | 0x234f1720 | 0x188000 |
|
| 126 |
+
| 23 | blk.2.attn_norm.weight | 0x23679720 | 0x4000 |
|
| 127 |
+
| 24 | blk.2.attn_output.weight | 0x2367d720 | 0x6e0000 |
|
| 128 |
+
| 25 | blk.2.attn_q.weight | 0x23d5d720 | 0x620000 |
|
| 129 |
+
| 26 | blk.2.attn_v.weight | 0x2437d720 | 0x1b8000 |
|
| 130 |
+
| 27 | blk.2.ffn_down.weight | 0x24535720 | 0x1810000 |
|
| 131 |
+
| 28 | blk.2.ffn_gate.weight | 0x25d45720 | 0x1570000 |
|
| 132 |
+
| 29 | blk.2.ffn_norm.weight | 0x272b5720 | 0x4000 |
|
| 133 |
+
| 30 | blk.2.ffn_up.weight | 0x272b9720 | 0x1570000 |
|
| 134 |
+
| 31 | blk.3.attn_k.weight | 0x28829720 | 0x188000 |
|
| 135 |
+
| 32 | blk.3.attn_norm.weight | 0x289b1720 | 0x4000 |
|
| 136 |
+
| 33 | blk.3.attn_output.weight | 0x289b5720 | 0x6e0000 |
|
| 137 |
+
| 34 | blk.3.attn_q.weight | 0x29095720 | 0x620000 |
|
| 138 |
+
| 35 | blk.3.attn_v.weight | 0x296b5720 | 0x1b8000 |
|
| 139 |
+
| 36 | blk.3.ffn_down.weight | 0x2986d720 | 0x1810000 |
|
| 140 |
+
| 37 | blk.3.ffn_gate.weight | 0x2b07d720 | 0x1570000 |
|
| 141 |
+
| 38 | blk.3.ffn_norm.weight | 0x2c5ed720 | 0x4000 |
|
| 142 |
+
| 39 | blk.3.ffn_up.weight | 0x2c5f1720 | 0x1570000 |
|
| 143 |
+
| 40 | blk.4.attn_k.weight | 0x2db61720 | 0x188000 |
|
| 144 |
+
| 41 | blk.4.attn_norm.weight | 0x2dce9720 | 0x4000 |
|
| 145 |
+
| 42 | blk.4.attn_output.weight | 0x2dced720 | 0x6e0000 |
|
| 146 |
+
| 43 | blk.4.attn_q.weight | 0x2e3cd720 | 0x620000 |
|
| 147 |
+
| 44 | blk.4.attn_v.weight | 0x2e9ed720 | 0x1b8000 |
|
| 148 |
+
| 45 | blk.4.ffn_down.weight | 0x2eba5720 | 0x1810000 |
|
| 149 |
+
| 46 | blk.4.ffn_gate.weight | 0x303b5720 | 0x1570000 |
|
| 150 |
+
| 47 | blk.4.ffn_norm.weight | 0x31925720 | 0x4000 |
|
| 151 |
+
| 48 | blk.4.ffn_up.weight | 0x31929720 | 0x1570000 |
|
| 152 |
+
| 49 | blk.5.attn_k.weight | 0x32e99720 | 0x188000 |
|
| 153 |
+
| 50 | blk.5.attn_norm.weight | 0x33021720 | 0x4000 |
|
| 154 |
+
| 51 | blk.5.attn_output.weight | 0x33025720 | 0x6e0000 |
|
| 155 |
+
| 52 | blk.5.attn_q.weight | 0x33705720 | 0x620000 |
|
| 156 |
+
| 53 | blk.5.attn_v.weight | 0x33d25720 | 0x1b8000 |
|
| 157 |
+
| 54 | blk.5.ffn_down.weight | 0x33edd720 | 0x1810000 |
|
| 158 |
+
| 55 | blk.5.ffn_gate.weight | 0x356ed720 | 0x1570000 |
|
| 159 |
+
| 56 | blk.5.ffn_norm.weight | 0x36c5d720 | 0x4000 |
|
| 160 |
+
| 57 | blk.5.ffn_up.weight | 0x36c61720 | 0x1570000 |
|
| 161 |
+
| 58 | blk.6.attn_k.weight | 0x381d1720 | 0x188000 |
|
| 162 |
+
| 59 | blk.6.attn_norm.weight | 0x38359720 | 0x4000 |
|
| 163 |
+
| 60 | blk.6.attn_output.weight | 0x3835d720 | 0x6e0000 |
|
| 164 |
+
| 61 | blk.6.attn_q.weight | 0x38a3d720 | 0x620000 |
|
| 165 |
+
| 62 | blk.6.attn_v.weight | 0x3905d720 | 0x1b8000 |
|
| 166 |
+
| 63 | blk.6.ffn_down.weight | 0x39215720 | 0x1810000 |
|
| 167 |
+
| 64 | blk.6.ffn_gate.weight | 0x3aa25720 | 0x1570000 |
|
| 168 |
+
| 65 | blk.6.ffn_norm.weight | 0x3bf95720 | 0x4000 |
|
| 169 |
+
| 66 | blk.6.ffn_up.weight | 0x3bf99720 | 0x1570000 |
|
| 170 |
+
| 67 | blk.7.attn_k.weight | 0x3d509720 | 0x188000 |
|
| 171 |
+
| 68 | blk.7.attn_norm.weight | 0x3d691720 | 0x4000 |
|
| 172 |
+
| 69 | blk.7.attn_output.weight | 0x3d695720 | 0x6e0000 |
|
| 173 |
+
| 70 | blk.7.attn_q.weight | 0x3dd75720 | 0x620000 |
|
| 174 |
+
| 71 | blk.7.attn_v.weight | 0x3e395720 | 0x1b8000 |
|
| 175 |
+
| 72 | blk.7.ffn_down.weight | 0x3e54d720 | 0x1810000 |
|
| 176 |
+
| 73 | blk.7.ffn_gate.weight | 0x3fd5d720 | 0x1570000 |
|
| 177 |
+
| 74 | blk.7.ffn_norm.weight | 0x412cd720 | 0x4000 |
|
| 178 |
+
| 75 | blk.7.ffn_up.weight | 0x412d1720 | 0x1570000 |
|
| 179 |
+
| 76 | blk.8.attn_k.weight | 0x42841720 | 0x188000 |
|
| 180 |
+
| 77 | blk.8.attn_norm.weight | 0x429c9720 | 0x4000 |
|
| 181 |
+
| 78 | blk.8.attn_output.weight | 0x429cd720 | 0x6e0000 |
|
| 182 |
+
| 79 | blk.8.attn_q.weight | 0x430ad720 | 0x620000 |
|
| 183 |
+
| 80 | blk.8.attn_v.weight | 0x436cd720 | 0x1b8000 |
|
| 184 |
+
| 81 | blk.8.ffn_down.weight | 0x43885720 | 0x1810000 |
|
| 185 |
+
| 82 | blk.8.ffn_gate.weight | 0x45095720 | 0x1570000 |
|
| 186 |
+
| 83 | blk.8.ffn_norm.weight | 0x46605720 | 0x4000 |
|
| 187 |
+
| 84 | blk.8.ffn_up.weight | 0x46609720 | 0x1570000 |
|
| 188 |
+
| 85 | blk.9.attn_k.weight | 0x47b79720 | 0x188000 |
|
| 189 |
+
| 86 | blk.9.attn_norm.weight | 0x47d01720 | 0x4000 |
|
| 190 |
+
| 87 | blk.9.attn_output.weight | 0x47d05720 | 0x6e0000 |
|
| 191 |
+
| 88 | blk.9.attn_q.weight | 0x483e5720 | 0x620000 |
|
| 192 |
+
| 89 | blk.9.attn_v.weight | 0x48a05720 | 0x1b8000 |
|
| 193 |
+
| 90 | blk.9.ffn_down.weight | 0x48bbd720 | 0x1810000 |
|
| 194 |
+
| 91 | blk.9.ffn_gate.weight | 0x4a3cd720 | 0x1570000 |
|
| 195 |
+
| 92 | blk.9.ffn_norm.weight | 0x4b93d720 | 0x4000 |
|
| 196 |
+
| 93 | blk.9.ffn_up.weight | 0x4b941720 | 0x1570000 |
|
| 197 |
+
| 94 | blk.10.attn_k.weight | 0x4ceb1720 | 0x188000 |
|
| 198 |
+
| 95 | blk.10.attn_norm.weight | 0x4d039720 | 0x4000 |
|
| 199 |
+
| 96 | blk.10.attn_output.weight | 0x4d03d720 | 0x6e0000 |
|
| 200 |
+
| 97 | blk.10.attn_q.weight | 0x4d71d720 | 0x620000 |
|
| 201 |
+
| 98 | blk.10.attn_v.weight | 0x4dd3d720 | 0x1b8000 |
|
| 202 |
+
| 99 | blk.10.ffn_down.weight | 0x4def5720 | 0x1810000 |
|
| 203 |
+
| 100 | blk.10.ffn_gate.weight | 0x4f705720 | 0x1570000 |
|
| 204 |
+
| 101 | blk.10.ffn_norm.weight | 0x50c75720 | 0x4000 |
|
| 205 |
+
| 102 | blk.10.ffn_up.weight | 0x50c79720 | 0x1570000 |
|
| 206 |
+
| 103 | blk.11.attn_k.weight | 0x521e9720 | 0x188000 |
|
| 207 |
+
| 104 | blk.11.attn_norm.weight | 0x52371720 | 0x4000 |
|
| 208 |
+
| 105 | blk.11.attn_output.weight | 0x52375720 | 0x6e0000 |
|
| 209 |
+
| 106 | blk.11.attn_q.weight | 0x52a55720 | 0x620000 |
|
| 210 |
+
| 107 | blk.11.attn_v.weight | 0x53075720 | 0x1b8000 |
|
| 211 |
+
| 108 | blk.11.ffn_down.weight | 0x5322d720 | 0x1810000 |
|
| 212 |
+
| 109 | blk.11.ffn_gate.weight | 0x54a3d720 | 0x1570000 |
|
| 213 |
+
| 110 | blk.11.ffn_norm.weight | 0x55fad720 | 0x4000 |
|
| 214 |
+
| 111 | blk.11.ffn_up.weight | 0x55fb1720 | 0x1570000 |
|
| 215 |
+
| 112 | blk.12.attn_k.weight | 0x57521720 | 0x188000 |
|
| 216 |
+
| 113 | blk.12.attn_norm.weight | 0x576a9720 | 0x4000 |
|
| 217 |
+
| 114 | blk.12.attn_output.weight | 0x576ad720 | 0x6e0000 |
|
| 218 |
+
| 115 | blk.12.attn_q.weight | 0x57d8d720 | 0x620000 |
|
| 219 |
+
| 116 | blk.12.attn_v.weight | 0x583ad720 | 0x1b8000 |
|
| 220 |
+
| 117 | blk.12.ffn_down.weight | 0x58565720 | 0x1810000 |
|
| 221 |
+
| 118 | blk.12.ffn_gate.weight | 0x59d75720 | 0x1570000 |
|
| 222 |
+
| 119 | blk.12.ffn_norm.weight | 0x5b2e5720 | 0x4000 |
|
| 223 |
+
| 120 | blk.12.ffn_up.weight | 0x5b2e9720 | 0x1570000 |
|
| 224 |
+
| 121 | blk.13.attn_k.weight | 0x5c859720 | 0x1b8000 |
|
| 225 |
+
| 122 | blk.13.attn_norm.weight | 0x5ca11720 | 0x4000 |
|
| 226 |
+
| 123 | blk.13.attn_output.weight | 0x5ca15720 | 0x6e0000 |
|
| 227 |
+
| 124 | blk.13.attn_q.weight | 0x5d0f5720 | 0x6e0000 |
|
| 228 |
+
| 125 | blk.13.attn_v.weight | 0x5d7d5720 | 0x1b8000 |
|
| 229 |
+
| 126 | blk.13.ffn_down.weight | 0x5d98d720 | 0x1810000 |
|
| 230 |
+
| 127 | blk.13.ffn_gate.weight | 0x5f19d720 | 0x1570000 |
|
| 231 |
+
| 128 | blk.13.ffn_norm.weight | 0x6070d720 | 0x4000 |
|
| 232 |
+
| 129 | blk.13.ffn_up.weight | 0x60711720 | 0x1570000 |
|
| 233 |
+
| 130 | blk.14.attn_k.weight | 0x61c81720 | 0x1b8000 |
|
| 234 |
+
| 131 | blk.14.attn_norm.weight | 0x61e39720 | 0x4000 |
|
| 235 |
+
| 132 | blk.14.attn_output.weight | 0x61e3d720 | 0x6e0000 |
|
| 236 |
+
| 133 | blk.14.attn_q.weight | 0x6251d720 | 0x6e0000 |
|
| 237 |
+
| 134 | blk.14.attn_v.weight | 0x62bfd720 | 0x1b8000 |
|
| 238 |
+
| 135 | blk.14.ffn_down.weight | 0x62db5720 | 0x1810000 |
|
| 239 |
+
| 136 | blk.14.ffn_gate.weight | 0x645c5720 | 0x1570000 |
|
| 240 |
+
| 137 | blk.14.ffn_norm.weight | 0x65b35720 | 0x4000 |
|
| 241 |
+
| 138 | blk.14.ffn_up.weight | 0x65b39720 | 0x1570000 |
|
| 242 |
+
| 139 | blk.15.attn_k.weight | 0x670a9720 | 0x188000 |
|
| 243 |
+
| 140 | blk.15.attn_norm.weight | 0x67231720 | 0x4000 |
|
| 244 |
+
| 141 | blk.15.attn_output.weight | 0x67235720 | 0x6e0000 |
|
| 245 |
+
| 142 | blk.15.attn_q.weight | 0x67915720 | 0x620000 |
|
| 246 |
+
| 143 | blk.15.attn_v.weight | 0x67f35720 | 0x1b8000 |
|
| 247 |
+
| 144 | blk.15.ffn_down.weight | 0x680ed720 | 0x1810000 |
|
| 248 |
+
| 145 | blk.15.ffn_gate.weight | 0x698fd720 | 0x1570000 |
|
| 249 |
+
| 146 | blk.15.ffn_norm.weight | 0x6ae6d720 | 0x4000 |
|
| 250 |
+
| 147 | blk.15.ffn_up.weight | 0x6ae71720 | 0x1570000 |
|
| 251 |
+
| 148 | blk.16.attn_k.weight | 0x6c3e1720 | 0x1b8000 |
|
| 252 |
+
| 149 | blk.16.attn_norm.weight | 0x6c599720 | 0x4000 |
|
| 253 |
+
| 150 | blk.16.attn_output.weight | 0x6c59d720 | 0x6e0000 |
|
| 254 |
+
| 151 | blk.16.attn_q.weight | 0x6cc7d720 | 0x6e0000 |
|
| 255 |
+
| 152 | blk.16.attn_v.weight | 0x6d35d720 | 0x1b8000 |
|
| 256 |
+
| 153 | blk.16.ffn_down.weight | 0x6d515720 | 0x1810000 |
|
| 257 |
+
| 154 | blk.16.ffn_gate.weight | 0x6ed25720 | 0x1810000 |
|
| 258 |
+
| 155 | blk.16.ffn_norm.weight | 0x70535720 | 0x4000 |
|
| 259 |
+
| 156 | blk.16.ffn_up.weight | 0x70539720 | 0x1810000 |
|
| 260 |
+
| 157 | blk.17.attn_k.weight | 0x71d49720 | 0x188000 |
|
| 261 |
+
| 158 | blk.17.attn_norm.weight | 0x71ed1720 | 0x4000 |
|
| 262 |
+
| 159 | blk.17.attn_output.weight | 0x71ed5720 | 0x6e0000 |
|
| 263 |
+
| 160 | blk.17.attn_q.weight | 0x725b5720 | 0x620000 |
|
| 264 |
+
| 161 | blk.17.attn_v.weight | 0x72bd5720 | 0x1b8000 |
|
| 265 |
+
| 162 | blk.17.ffn_down.weight | 0x72d8d720 | 0x1f80000 |
|
| 266 |
+
| 163 | blk.17.ffn_gate.weight | 0x74d0d720 | 0x1810000 |
|
| 267 |
+
| 164 | blk.17.ffn_norm.weight | 0x7651d720 | 0x4000 |
|
| 268 |
+
| 165 | blk.17.ffn_up.weight | 0x76521720 | 0x1810000 |
|
| 269 |
+
| 166 | blk.18.attn_k.weight | 0x77d31720 | 0x1b8000 |
|
| 270 |
+
| 167 | blk.18.attn_norm.weight | 0x77ee9720 | 0x4000 |
|
| 271 |
+
| 168 | blk.18.attn_output.weight | 0x77eed720 | 0x6e0000 |
|
| 272 |
+
| 169 | blk.18.attn_q.weight | 0x785cd720 | 0x6e0000 |
|
| 273 |
+
| 170 | blk.18.attn_v.weight | 0x78cad720 | 0x1b8000 |
|
| 274 |
+
| 171 | blk.18.ffn_down.weight | 0x78e65720 | 0x1f80000 |
|
| 275 |
+
| 172 | blk.18.ffn_gate.weight | 0x7ade5720 | 0x1810000 |
|
| 276 |
+
| 173 | blk.18.ffn_norm.weight | 0x7c5f5720 | 0x4000 |
|
| 277 |
+
| 174 | blk.18.ffn_up.weight | 0x7c5f9720 | 0x1810000 |
|
| 278 |
+
| 175 | blk.19.attn_k.weight | 0x7de09720 | 0x1b8000 |
|
| 279 |
+
| 176 | blk.19.attn_norm.weight | 0x7dfc1720 | 0x4000 |
|
| 280 |
+
| 177 | blk.19.attn_output.weight | 0x7dfc5720 | 0x6e0000 |
|
| 281 |
+
| 178 | blk.19.attn_q.weight | 0x7e6a5720 | 0x6e0000 |
|
| 282 |
+
| 179 | blk.19.attn_v.weight | 0x7ed85720 | 0x1b8000 |
|
| 283 |
+
| 180 | blk.19.ffn_down.weight | 0x7ef3d720 | 0x1f80000 |
|
| 284 |
+
| 181 | blk.19.ffn_gate.weight | 0x80ebd720 | 0x1810000 |
|
| 285 |
+
| 182 | blk.19.ffn_norm.weight | 0x826cd720 | 0x4000 |
|
| 286 |
+
| 183 | blk.19.ffn_up.weight | 0x826d1720 | 0x1810000 |
|
| 287 |
+
| 184 | blk.20.attn_k.weight | 0x83ee1720 | 0x1b8000 |
|
| 288 |
+
| 185 | blk.20.attn_norm.weight | 0x84099720 | 0x4000 |
|
| 289 |
+
| 186 | blk.20.attn_output.weight | 0x8409d720 | 0x6e0000 |
|
| 290 |
+
| 187 | blk.20.attn_q.weight | 0x8477d720 | 0x6e0000 |
|
| 291 |
+
| 188 | blk.20.attn_v.weight | 0x84e5d720 | 0x1b8000 |
|
| 292 |
+
| 189 | blk.20.ffn_down.weight | 0x85015720 | 0x1f80000 |
|
| 293 |
+
| 190 | blk.20.ffn_gate.weight | 0x86f95720 | 0x1810000 |
|
| 294 |
+
| 191 | blk.20.ffn_norm.weight | 0x887a5720 | 0x4000 |
|
| 295 |
+
| 192 | blk.20.ffn_up.weight | 0x887a9720 | 0x1810000 |
|
| 296 |
+
| 193 | blk.21.attn_k.weight | 0x89fb9720 | 0x1b8000 |
|
| 297 |
+
| 194 | blk.21.attn_norm.weight | 0x8a171720 | 0x4000 |
|
| 298 |
+
| 195 | blk.21.attn_output.weight | 0x8a175720 | 0x6e0000 |
|
| 299 |
+
| 196 | blk.21.attn_q.weight | 0x8a855720 | 0x6e0000 |
|
| 300 |
+
| 197 | blk.21.attn_v.weight | 0x8af35720 | 0x1b8000 |
|
| 301 |
+
| 198 | blk.21.ffn_down.weight | 0x8b0ed720 | 0x1f80000 |
|
| 302 |
+
| 199 | blk.21.ffn_gate.weight | 0x8d06d720 | 0x1810000 |
|
| 303 |
+
| 200 | blk.21.ffn_norm.weight | 0x8e87d720 | 0x4000 |
|
| 304 |
+
| 201 | blk.21.ffn_up.weight | 0x8e881720 | 0x1810000 |
|
| 305 |
+
| 202 | blk.22.attn_k.weight | 0x90091720 | 0x1b8000 |
|
| 306 |
+
| 203 | blk.22.attn_norm.weight | 0x90249720 | 0x4000 |
|
| 307 |
+
| 204 | blk.22.attn_output.weight | 0x9024d720 | 0x6e0000 |
|
| 308 |
+
| 205 | blk.22.attn_q.weight | 0x9092d720 | 0x6e0000 |
|
| 309 |
+
| 206 | blk.22.attn_v.weight | 0x9100d720 | 0x1b8000 |
|
| 310 |
+
| 207 | blk.22.ffn_down.weight | 0x911c5720 | 0x1f80000 |
|
| 311 |
+
| 208 | blk.22.ffn_gate.weight | 0x93145720 | 0x1810000 |
|
| 312 |
+
| 209 | blk.22.ffn_norm.weight | 0x94955720 | 0x4000 |
|
| 313 |
+
| 210 | blk.22.ffn_up.weight | 0x94959720 | 0x1810000 |
|
| 314 |
+
| 211 | blk.23.attn_k.weight | 0x96169720 | 0x1b8000 |
|
| 315 |
+
| 212 | blk.23.attn_norm.weight | 0x96321720 | 0x4000 |
|
| 316 |
+
| 213 | blk.23.attn_output.weight | 0x96325720 | 0x6e0000 |
|
| 317 |
+
| 214 | blk.23.attn_q.weight | 0x96a05720 | 0x6e0000 |
|
| 318 |
+
| 215 | blk.23.attn_v.weight | 0x970e5720 | 0x1b8000 |
|
| 319 |
+
| 216 | blk.23.ffn_down.weight | 0x9729d720 | 0x1f80000 |
|
| 320 |
+
| 217 | blk.23.ffn_gate.weight | 0x9921d720 | 0x1810000 |
|
| 321 |
+
| 218 | blk.23.ffn_norm.weight | 0x9aa2d720 | 0x4000 |
|
| 322 |
+
| 219 | blk.23.ffn_up.weight | 0x9aa31720 | 0x1810000 |
|
| 323 |
+
| 220 | blk.24.attn_k.weight | 0x9c241720 | 0x1b8000 |
|
| 324 |
+
| 221 | blk.24.attn_norm.weight | 0x9c3f9720 | 0x4000 |
|
| 325 |
+
| 222 | blk.24.attn_output.weight | 0x9c3fd720 | 0x6e0000 |
|
| 326 |
+
| 223 | blk.24.attn_q.weight | 0x9cadd720 | 0x6e0000 |
|
| 327 |
+
| 224 | blk.24.attn_v.weight | 0x9d1bd720 | 0x1b8000 |
|
| 328 |
+
| 225 | blk.24.ffn_down.weight | 0x9d375720 | 0x1f80000 |
|
| 329 |
+
| 226 | blk.24.ffn_gate.weight | 0x9f2f5720 | 0x1810000 |
|
| 330 |
+
| 227 | blk.24.ffn_norm.weight | 0xa0b05720 | 0x4000 |
|
| 331 |
+
| 228 | blk.24.ffn_up.weight | 0xa0b09720 | 0x1810000 |
|
| 332 |
+
| 229 | blk.25.attn_k.weight | 0xa2319720 | 0x1b8000 |
|
| 333 |
+
| 230 | blk.25.attn_norm.weight | 0xa24d1720 | 0x4000 |
|
| 334 |
+
| 231 | blk.25.attn_output.weight | 0xa24d5720 | 0x6e0000 |
|
| 335 |
+
| 232 | blk.25.attn_q.weight | 0xa2bb5720 | 0x6e0000 |
|
| 336 |
+
| 233 | blk.25.attn_v.weight | 0xa3295720 | 0x1b8000 |
|
| 337 |
+
| 234 | blk.25.ffn_down.weight | 0xa344d720 | 0x1f80000 |
|
| 338 |
+
| 235 | blk.25.ffn_gate.weight | 0xa53cd720 | 0x1810000 |
|
| 339 |
+
| 236 | blk.25.ffn_norm.weight | 0xa6bdd720 | 0x4000 |
|
| 340 |
+
| 237 | blk.25.ffn_up.weight | 0xa6be1720 | 0x1810000 |
|
| 341 |
+
| 238 | blk.26.attn_k.weight | 0xa83f1720 | 0x1b8000 |
|
| 342 |
+
| 239 | blk.26.attn_norm.weight | 0xa85a9720 | 0x4000 |
|
| 343 |
+
| 240 | blk.26.attn_output.weight | 0xa85ad720 | 0x6e0000 |
|
| 344 |
+
| 241 | blk.26.attn_q.weight | 0xa8c8d720 | 0x6e0000 |
|
| 345 |
+
| 242 | blk.26.attn_v.weight | 0xa936d720 | 0x1b8000 |
|
| 346 |
+
| 243 | blk.26.ffn_down.weight | 0xa9525720 | 0x1f80000 |
|
| 347 |
+
| 244 | blk.26.ffn_gate.weight | 0xab4a5720 | 0x1810000 |
|
| 348 |
+
| 245 | blk.26.ffn_norm.weight | 0xaccb5720 | 0x4000 |
|
| 349 |
+
| 246 | blk.26.ffn_up.weight | 0xaccb9720 | 0x1810000 |
|
| 350 |
+
| 247 | blk.27.attn_k.weight | 0xae4c9720 | 0x1b8000 |
|
| 351 |
+
| 248 | blk.27.attn_norm.weight | 0xae681720 | 0x4000 |
|
| 352 |
+
| 249 | blk.27.attn_output.weight | 0xae685720 | 0x6e0000 |
|
| 353 |
+
| 250 | blk.27.attn_q.weight | 0xaed65720 | 0x6e0000 |
|
| 354 |
+
| 251 | blk.27.attn_v.weight | 0xaf445720 | 0x1b8000 |
|
| 355 |
+
| 252 | blk.27.ffn_down.weight | 0xaf5fd720 | 0x1f80000 |
|
| 356 |
+
| 253 | blk.27.ffn_gate.weight | 0xb157d720 | 0x1810000 |
|
| 357 |
+
| 254 | blk.27.ffn_norm.weight | 0xb2d8d720 | 0x4000 |
|
| 358 |
+
| 255 | blk.27.ffn_up.weight | 0xb2d91720 | 0x1810000 |
|
| 359 |
+
| 256 | blk.28.attn_k.weight | 0xb45a1720 | 0x1b8000 |
|
| 360 |
+
| 257 | blk.28.attn_norm.weight | 0xb4759720 | 0x4000 |
|
| 361 |
+
| 258 | blk.28.attn_output.weight | 0xb475d720 | 0x6e0000 |
|
| 362 |
+
| 259 | blk.28.attn_q.weight | 0xb4e3d720 | 0x6e0000 |
|
| 363 |
+
| 260 | blk.28.attn_v.weight | 0xb551d720 | 0x1b8000 |
|
| 364 |
+
| 261 | blk.28.ffn_down.weight | 0xb56d5720 | 0x1f80000 |
|
| 365 |
+
| 262 | blk.28.ffn_gate.weight | 0xb7655720 | 0x1810000 |
|
| 366 |
+
| 263 | blk.28.ffn_norm.weight | 0xb8e65720 | 0x4000 |
|
| 367 |
+
| 264 | blk.28.ffn_up.weight | 0xb8e69720 | 0x1810000 |
|
| 368 |
+
| 265 | blk.29.attn_k.weight | 0xba679720 | 0x1b8000 |
|
| 369 |
+
| 266 | blk.29.attn_norm.weight | 0xba831720 | 0x4000 |
|
| 370 |
+
| 267 | blk.29.attn_output.weight | 0xba835720 | 0x6e0000 |
|
| 371 |
+
| 268 | blk.29.attn_q.weight | 0xbaf15720 | 0x6e0000 |
|
| 372 |
+
| 269 | blk.29.attn_v.weight | 0xbb5f5720 | 0x1b8000 |
|
| 373 |
+
| 270 | blk.29.ffn_down.weight | 0xbb7ad720 | 0x1f80000 |
|
| 374 |
+
| 271 | blk.29.ffn_gate.weight | 0xbd72d720 | 0x1810000 |
|
| 375 |
+
| 272 | blk.29.ffn_norm.weight | 0xbef3d720 | 0x4000 |
|
| 376 |
+
| 273 | blk.29.ffn_up.weight | 0xbef41720 | 0x1810000 |
|
| 377 |
+
| 274 | blk.30.attn_k.weight | 0xc0751720 | 0x1b8000 |
|
| 378 |
+
| 275 | blk.30.attn_norm.weight | 0xc0909720 | 0x4000 |
|
| 379 |
+
| 276 | blk.30.attn_output.weight | 0xc090d720 | 0x6e0000 |
|
| 380 |
+
| 277 | blk.30.attn_q.weight | 0xc0fed720 | 0x6e0000 |
|
| 381 |
+
| 278 | blk.30.attn_v.weight | 0xc16cd720 | 0x1b8000 |
|
| 382 |
+
| 279 | blk.30.ffn_down.weight | 0xc1885720 | 0x1f80000 |
|
| 383 |
+
| 280 | blk.30.ffn_gate.weight | 0xc3805720 | 0x1810000 |
|
| 384 |
+
| 281 | blk.30.ffn_norm.weight | 0xc5015720 | 0x4000 |
|
| 385 |
+
| 282 | blk.30.ffn_up.weight | 0xc5019720 | 0x1810000 |
|
| 386 |
+
| 283 | blk.31.attn_k.weight | 0xc6829720 | 0x188000 |
|
| 387 |
+
| 284 | blk.31.attn_norm.weight | 0xc69b1720 | 0x4000 |
|
| 388 |
+
| 285 | blk.31.attn_output.weight | 0xc69b5720 | 0x6e0000 |
|
| 389 |
+
| 286 | blk.31.attn_q.weight | 0xc7095720 | 0x620000 |
|
| 390 |
+
| 287 | blk.31.attn_v.weight | 0xc76b5720 | 0x1b8000 |
|
| 391 |
+
| 288 | blk.31.ffn_down.weight | 0xc786d720 | 0x1f80000 |
|
| 392 |
+
| 289 | blk.31.ffn_gate.weight | 0xc97ed720 | 0x1810000 |
|
| 393 |
+
| 290 | blk.31.ffn_norm.weight | 0xcaffd720 | 0x4000 |
|
| 394 |
+
| 291 | blk.31.ffn_up.weight | 0xcb001720 | 0x1810000 |
|
| 395 |
+
|
| 396 |
+
### <a name="base">Base Tensor Group : ~1B Elements</a>
|
| 397 |
+
|
| 398 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 399 |
+
|-----:|:-------------------|:---------------------------------|:------------------|:----------------------|:--------|
|
| 400 |
+
| 0 | output.weight | Output (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | IQ3_XXS |
|
| 401 |
+
| 1 | output_norm.weight | Output Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 402 |
+
| 2 | rope_freqs.weight | Rope_Freqs (W) | ( 64) 64 | 64 x 1 x 1 x 1 | F32 |
|
| 403 |
+
| 3 | token_embd.weight | Token Embedding (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | IQ3_XXS |
|
| 404 |
+
|
| 405 |
+
- Total elements in base: ( ~1B) 1050677312
|
| 406 |
+
- Percentage of total elements: 13.08%
|
| 407 |
+
|
| 408 |
+
|
| 409 |
+
### <a name="blk_0">Block 0 Tensor Group : ~218M Elements</a>
|
| 410 |
+
|
| 411 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 412 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:--------|
|
| 413 |
+
| 4 | blk.0.attn_k.weight | Block 0 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 414 |
+
| 5 | blk.0.attn_norm.weight | Block 0 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 415 |
+
| 6 | blk.0.attn_output.weight | Block 0 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 416 |
+
| 7 | blk.0.attn_q.weight | Block 0 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 417 |
+
| 8 | blk.0.attn_v.weight | Block 0 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 418 |
+
| 9 | blk.0.ffn_down.weight | Block 0 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ3_S |
|
| 419 |
+
| 10 | blk.0.ffn_gate.weight | Block 0 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 420 |
+
| 11 | blk.0.ffn_norm.weight | Block 0 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 421 |
+
| 12 | blk.0.ffn_up.weight | Block 0 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 422 |
+
|
| 423 |
+
- Total elements in blk.0: (~218M) 218112000
|
| 424 |
+
- Percentage of total elements: 2.72%
|
| 425 |
+
|
| 426 |
+
|
| 427 |
+
### <a name="blk_1">Block 1 Tensor Group : ~218M Elements</a>
|
| 428 |
+
|
| 429 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 430 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:--------|
|
| 431 |
+
| 13 | blk.1.attn_k.weight | Block 1 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 432 |
+
| 14 | blk.1.attn_norm.weight | Block 1 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 433 |
+
| 15 | blk.1.attn_output.weight | Block 1 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 434 |
+
| 16 | blk.1.attn_q.weight | Block 1 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 435 |
+
| 17 | blk.1.attn_v.weight | Block 1 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 436 |
+
| 18 | blk.1.ffn_down.weight | Block 1 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 437 |
+
| 19 | blk.1.ffn_gate.weight | Block 1 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 438 |
+
| 20 | blk.1.ffn_norm.weight | Block 1 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 439 |
+
| 21 | blk.1.ffn_up.weight | Block 1 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 440 |
+
|
| 441 |
+
- Total elements in blk.1: (~218M) 218112000
|
| 442 |
+
- Percentage of total elements: 2.72%
|
| 443 |
+
|
| 444 |
+
|
| 445 |
+
### <a name="blk_2">Block 2 Tensor Group : ~218M Elements</a>
|
| 446 |
+
|
| 447 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 448 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:--------|
|
| 449 |
+
| 22 | blk.2.attn_k.weight | Block 2 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 450 |
+
| 23 | blk.2.attn_norm.weight | Block 2 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 451 |
+
| 24 | blk.2.attn_output.weight | Block 2 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 452 |
+
| 25 | blk.2.attn_q.weight | Block 2 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 453 |
+
| 26 | blk.2.attn_v.weight | Block 2 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 454 |
+
| 27 | blk.2.ffn_down.weight | Block 2 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ3_S |
|
| 455 |
+
| 28 | blk.2.ffn_gate.weight | Block 2 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 456 |
+
| 29 | blk.2.ffn_norm.weight | Block 2 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 457 |
+
| 30 | blk.2.ffn_up.weight | Block 2 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 458 |
+
|
| 459 |
+
- Total elements in blk.2: (~218M) 218112000
|
| 460 |
+
- Percentage of total elements: 2.72%
|
| 461 |
+
|
| 462 |
+
|
| 463 |
+
### <a name="blk_3">Block 3 Tensor Group : ~218M Elements</a>
|
| 464 |
+
|
| 465 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 466 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:--------|
|
| 467 |
+
| 31 | blk.3.attn_k.weight | Block 3 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 468 |
+
| 32 | blk.3.attn_norm.weight | Block 3 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 469 |
+
| 33 | blk.3.attn_output.weight | Block 3 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 470 |
+
| 34 | blk.3.attn_q.weight | Block 3 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 471 |
+
| 35 | blk.3.attn_v.weight | Block 3 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 472 |
+
| 36 | blk.3.ffn_down.weight | Block 3 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ3_S |
|
| 473 |
+
| 37 | blk.3.ffn_gate.weight | Block 3 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 474 |
+
| 38 | blk.3.ffn_norm.weight | Block 3 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 475 |
+
| 39 | blk.3.ffn_up.weight | Block 3 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 476 |
+
|
| 477 |
+
- Total elements in blk.3: (~218M) 218112000
|
| 478 |
+
- Percentage of total elements: 2.72%
|
| 479 |
+
|
| 480 |
+
|
| 481 |
+
### <a name="blk_4">Block 4 Tensor Group : ~218M Elements</a>
|
| 482 |
+
|
| 483 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 484 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:--------|
|
| 485 |
+
| 40 | blk.4.attn_k.weight | Block 4 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 486 |
+
| 41 | blk.4.attn_norm.weight | Block 4 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 487 |
+
| 42 | blk.4.attn_output.weight | Block 4 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 488 |
+
| 43 | blk.4.attn_q.weight | Block 4 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 489 |
+
| 44 | blk.4.attn_v.weight | Block 4 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 490 |
+
| 45 | blk.4.ffn_down.weight | Block 4 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ3_S |
|
| 491 |
+
| 46 | blk.4.ffn_gate.weight | Block 4 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 492 |
+
| 47 | blk.4.ffn_norm.weight | Block 4 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 493 |
+
| 48 | blk.4.ffn_up.weight | Block 4 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 494 |
+
|
| 495 |
+
- Total elements in blk.4: (~218M) 218112000
|
| 496 |
+
- Percentage of total elements: 2.72%
|
| 497 |
+
|
| 498 |
+
|
| 499 |
+
### <a name="blk_5">Block 5 Tensor Group : ~218M Elements</a>
|
| 500 |
+
|
| 501 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 502 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:--------|
|
| 503 |
+
| 49 | blk.5.attn_k.weight | Block 5 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 504 |
+
| 50 | blk.5.attn_norm.weight | Block 5 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 505 |
+
| 51 | blk.5.attn_output.weight | Block 5 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 506 |
+
| 52 | blk.5.attn_q.weight | Block 5 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 507 |
+
| 53 | blk.5.attn_v.weight | Block 5 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 508 |
+
| 54 | blk.5.ffn_down.weight | Block 5 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ3_S |
|
| 509 |
+
| 55 | blk.5.ffn_gate.weight | Block 5 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 510 |
+
| 56 | blk.5.ffn_norm.weight | Block 5 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 511 |
+
| 57 | blk.5.ffn_up.weight | Block 5 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 512 |
+
|
| 513 |
+
- Total elements in blk.5: (~218M) 218112000
|
| 514 |
+
- Percentage of total elements: 2.72%
|
| 515 |
+
|
| 516 |
+
|
| 517 |
+
### <a name="blk_6">Block 6 Tensor Group : ~218M Elements</a>
|
| 518 |
+
|
| 519 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 520 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:--------|
|
| 521 |
+
| 58 | blk.6.attn_k.weight | Block 6 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 522 |
+
| 59 | blk.6.attn_norm.weight | Block 6 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 523 |
+
| 60 | blk.6.attn_output.weight | Block 6 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 524 |
+
| 61 | blk.6.attn_q.weight | Block 6 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 525 |
+
| 62 | blk.6.attn_v.weight | Block 6 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 526 |
+
| 63 | blk.6.ffn_down.weight | Block 6 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ3_S |
|
| 527 |
+
| 64 | blk.6.ffn_gate.weight | Block 6 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 528 |
+
| 65 | blk.6.ffn_norm.weight | Block 6 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 529 |
+
| 66 | blk.6.ffn_up.weight | Block 6 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 530 |
+
|
| 531 |
+
- Total elements in blk.6: (~218M) 218112000
|
| 532 |
+
- Percentage of total elements: 2.72%
|
| 533 |
+
|
| 534 |
+
|
| 535 |
+
### <a name="blk_7">Block 7 Tensor Group : ~218M Elements</a>
|
| 536 |
+
|
| 537 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 538 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:--------|
|
| 539 |
+
| 67 | blk.7.attn_k.weight | Block 7 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 540 |
+
| 68 | blk.7.attn_norm.weight | Block 7 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 541 |
+
| 69 | blk.7.attn_output.weight | Block 7 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 542 |
+
| 70 | blk.7.attn_q.weight | Block 7 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 543 |
+
| 71 | blk.7.attn_v.weight | Block 7 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 544 |
+
| 72 | blk.7.ffn_down.weight | Block 7 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ3_S |
|
| 545 |
+
| 73 | blk.7.ffn_gate.weight | Block 7 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 546 |
+
| 74 | blk.7.ffn_norm.weight | Block 7 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 547 |
+
| 75 | blk.7.ffn_up.weight | Block 7 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 548 |
+
|
| 549 |
+
- Total elements in blk.7: (~218M) 218112000
|
| 550 |
+
- Percentage of total elements: 2.72%
|
| 551 |
+
|
| 552 |
+
|
| 553 |
+
### <a name="blk_8">Block 8 Tensor Group : ~218M Elements</a>
|
| 554 |
+
|
| 555 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 556 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:--------|
|
| 557 |
+
| 76 | blk.8.attn_k.weight | Block 8 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 558 |
+
| 77 | blk.8.attn_norm.weight | Block 8 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 559 |
+
| 78 | blk.8.attn_output.weight | Block 8 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 560 |
+
| 79 | blk.8.attn_q.weight | Block 8 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 561 |
+
| 80 | blk.8.attn_v.weight | Block 8 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 562 |
+
| 81 | blk.8.ffn_down.weight | Block 8 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ3_S |
|
| 563 |
+
| 82 | blk.8.ffn_gate.weight | Block 8 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 564 |
+
| 83 | blk.8.ffn_norm.weight | Block 8 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 565 |
+
| 84 | blk.8.ffn_up.weight | Block 8 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 566 |
+
|
| 567 |
+
- Total elements in blk.8: (~218M) 218112000
|
| 568 |
+
- Percentage of total elements: 2.72%
|
| 569 |
+
|
| 570 |
+
|
| 571 |
+
### <a name="blk_9">Block 9 Tensor Group : ~218M Elements</a>
|
| 572 |
+
|
| 573 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 574 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:--------|
|
| 575 |
+
| 85 | blk.9.attn_k.weight | Block 9 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 576 |
+
| 86 | blk.9.attn_norm.weight | Block 9 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 577 |
+
| 87 | blk.9.attn_output.weight | Block 9 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 578 |
+
| 88 | blk.9.attn_q.weight | Block 9 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 579 |
+
| 89 | blk.9.attn_v.weight | Block 9 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 580 |
+
| 90 | blk.9.ffn_down.weight | Block 9 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ3_S |
|
| 581 |
+
| 91 | blk.9.ffn_gate.weight | Block 9 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 582 |
+
| 92 | blk.9.ffn_norm.weight | Block 9 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 583 |
+
| 93 | blk.9.ffn_up.weight | Block 9 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 584 |
+
|
| 585 |
+
- Total elements in blk.9: (~218M) 218112000
|
| 586 |
+
- Percentage of total elements: 2.72%
|
| 587 |
+
|
| 588 |
+
|
| 589 |
+
### <a name="blk_10">Block 10 Tensor Group : ~218M Elements</a>
|
| 590 |
+
|
| 591 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 592 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:--------|
|
| 593 |
+
| 94 | blk.10.attn_k.weight | Block 10 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 594 |
+
| 95 | blk.10.attn_norm.weight | Block 10 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 595 |
+
| 96 | blk.10.attn_output.weight | Block 10 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 596 |
+
| 97 | blk.10.attn_q.weight | Block 10 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 597 |
+
| 98 | blk.10.attn_v.weight | Block 10 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 598 |
+
| 99 | blk.10.ffn_down.weight | Block 10 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ3_S |
|
| 599 |
+
| 100 | blk.10.ffn_gate.weight | Block 10 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 600 |
+
| 101 | blk.10.ffn_norm.weight | Block 10 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 601 |
+
| 102 | blk.10.ffn_up.weight | Block 10 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 602 |
+
|
| 603 |
+
- Total elements in blk.10: (~218M) 218112000
|
| 604 |
+
- Percentage of total elements: 2.72%
|
| 605 |
+
|
| 606 |
+
|
| 607 |
+
### <a name="blk_11">Block 11 Tensor Group : ~218M Elements</a>
|
| 608 |
+
|
| 609 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 610 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:--------|
|
| 611 |
+
| 103 | blk.11.attn_k.weight | Block 11 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 612 |
+
| 104 | blk.11.attn_norm.weight | Block 11 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 613 |
+
| 105 | blk.11.attn_output.weight | Block 11 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 614 |
+
| 106 | blk.11.attn_q.weight | Block 11 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 615 |
+
| 107 | blk.11.attn_v.weight | Block 11 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 616 |
+
| 108 | blk.11.ffn_down.weight | Block 11 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ3_S |
|
| 617 |
+
| 109 | blk.11.ffn_gate.weight | Block 11 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 618 |
+
| 110 | blk.11.ffn_norm.weight | Block 11 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 619 |
+
| 111 | blk.11.ffn_up.weight | Block 11 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 620 |
+
|
| 621 |
+
- Total elements in blk.11: (~218M) 218112000
|
| 622 |
+
- Percentage of total elements: 2.72%
|
| 623 |
+
|
| 624 |
+
|
| 625 |
+
### <a name="blk_12">Block 12 Tensor Group : ~218M Elements</a>
|
| 626 |
+
|
| 627 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 628 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:--------|
|
| 629 |
+
| 112 | blk.12.attn_k.weight | Block 12 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 630 |
+
| 113 | blk.12.attn_norm.weight | Block 12 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 631 |
+
| 114 | blk.12.attn_output.weight | Block 12 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 632 |
+
| 115 | blk.12.attn_q.weight | Block 12 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 633 |
+
| 116 | blk.12.attn_v.weight | Block 12 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 634 |
+
| 117 | blk.12.ffn_down.weight | Block 12 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ3_S |
|
| 635 |
+
| 118 | blk.12.ffn_gate.weight | Block 12 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 636 |
+
| 119 | blk.12.ffn_norm.weight | Block 12 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 637 |
+
| 120 | blk.12.ffn_up.weight | Block 12 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 638 |
+
|
| 639 |
+
- Total elements in blk.12: (~218M) 218112000
|
| 640 |
+
- Percentage of total elements: 2.72%
|
| 641 |
+
|
| 642 |
+
|
| 643 |
+
### <a name="blk_13">Block 13 Tensor Group : ~218M Elements</a>
|
| 644 |
+
|
| 645 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 646 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:--------|
|
| 647 |
+
| 121 | blk.13.attn_k.weight | Block 13 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 648 |
+
| 122 | blk.13.attn_norm.weight | Block 13 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 649 |
+
| 123 | blk.13.attn_output.weight | Block 13 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 650 |
+
| 124 | blk.13.attn_q.weight | Block 13 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 651 |
+
| 125 | blk.13.attn_v.weight | Block 13 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 652 |
+
| 126 | blk.13.ffn_down.weight | Block 13 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ3_S |
|
| 653 |
+
| 127 | blk.13.ffn_gate.weight | Block 13 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 654 |
+
| 128 | blk.13.ffn_norm.weight | Block 13 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 655 |
+
| 129 | blk.13.ffn_up.weight | Block 13 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 656 |
+
|
| 657 |
+
- Total elements in blk.13: (~218M) 218112000
|
| 658 |
+
- Percentage of total elements: 2.72%
|
| 659 |
+
|
| 660 |
+
|
| 661 |
+
### <a name="blk_14">Block 14 Tensor Group : ~218M Elements</a>
|
| 662 |
+
|
| 663 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 664 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:--------|
|
| 665 |
+
| 130 | blk.14.attn_k.weight | Block 14 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 666 |
+
| 131 | blk.14.attn_norm.weight | Block 14 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 667 |
+
| 132 | blk.14.attn_output.weight | Block 14 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 668 |
+
| 133 | blk.14.attn_q.weight | Block 14 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 669 |
+
| 134 | blk.14.attn_v.weight | Block 14 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 670 |
+
| 135 | blk.14.ffn_down.weight | Block 14 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ3_S |
|
| 671 |
+
| 136 | blk.14.ffn_gate.weight | Block 14 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 672 |
+
| 137 | blk.14.ffn_norm.weight | Block 14 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 673 |
+
| 138 | blk.14.ffn_up.weight | Block 14 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 674 |
+
|
| 675 |
+
- Total elements in blk.14: (~218M) 218112000
|
| 676 |
+
- Percentage of total elements: 2.72%
|
| 677 |
+
|
| 678 |
+
|
| 679 |
+
### <a name="blk_15">Block 15 Tensor Group : ~218M Elements</a>
|
| 680 |
+
|
| 681 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 682 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:--------|
|
| 683 |
+
| 139 | blk.15.attn_k.weight | Block 15 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 684 |
+
| 140 | blk.15.attn_norm.weight | Block 15 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 685 |
+
| 141 | blk.15.attn_output.weight | Block 15 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 686 |
+
| 142 | blk.15.attn_q.weight | Block 15 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 687 |
+
| 143 | blk.15.attn_v.weight | Block 15 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 688 |
+
| 144 | blk.15.ffn_down.weight | Block 15 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ3_S |
|
| 689 |
+
| 145 | blk.15.ffn_gate.weight | Block 15 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 690 |
+
| 146 | blk.15.ffn_norm.weight | Block 15 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 691 |
+
| 147 | blk.15.ffn_up.weight | Block 15 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_XXS |
|
| 692 |
+
|
| 693 |
+
- Total elements in blk.15: (~218M) 218112000
|
| 694 |
+
- Percentage of total elements: 2.72%
|
| 695 |
+
|
| 696 |
+
|
| 697 |
+
### <a name="blk_16">Block 16 Tensor Group : ~218M Elements</a>
|
| 698 |
+
|
| 699 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 700 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:------|
|
| 701 |
+
| 148 | blk.16.attn_k.weight | Block 16 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 702 |
+
| 149 | blk.16.attn_norm.weight | Block 16 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 703 |
+
| 150 | blk.16.attn_output.weight | Block 16 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 704 |
+
| 151 | blk.16.attn_q.weight | Block 16 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 705 |
+
| 152 | blk.16.attn_v.weight | Block 16 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 706 |
+
| 153 | blk.16.ffn_down.weight | Block 16 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ3_S |
|
| 707 |
+
| 154 | blk.16.ffn_gate.weight | Block 16 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 708 |
+
| 155 | blk.16.ffn_norm.weight | Block 16 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 709 |
+
| 156 | blk.16.ffn_up.weight | Block 16 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 710 |
+
|
| 711 |
+
- Total elements in blk.16: (~218M) 218112000
|
| 712 |
+
- Percentage of total elements: 2.72%
|
| 713 |
+
|
| 714 |
+
|
| 715 |
+
### <a name="blk_17">Block 17 Tensor Group : ~218M Elements</a>
|
| 716 |
+
|
| 717 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 718 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:--------|
|
| 719 |
+
| 157 | blk.17.attn_k.weight | Block 17 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 720 |
+
| 158 | blk.17.attn_norm.weight | Block 17 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 721 |
+
| 159 | blk.17.attn_output.weight | Block 17 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 722 |
+
| 160 | blk.17.attn_q.weight | Block 17 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 723 |
+
| 161 | blk.17.attn_v.weight | Block 17 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 724 |
+
| 162 | blk.17.ffn_down.weight | Block 17 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 725 |
+
| 163 | blk.17.ffn_gate.weight | Block 17 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 726 |
+
| 164 | blk.17.ffn_norm.weight | Block 17 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 727 |
+
| 165 | blk.17.ffn_up.weight | Block 17 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 728 |
+
|
| 729 |
+
- Total elements in blk.17: (~218M) 218112000
|
| 730 |
+
- Percentage of total elements: 2.72%
|
| 731 |
+
|
| 732 |
+
|
| 733 |
+
### <a name="blk_18">Block 18 Tensor Group : ~218M Elements</a>
|
| 734 |
+
|
| 735 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 736 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 737 |
+
| 166 | blk.18.attn_k.weight | Block 18 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 738 |
+
| 167 | blk.18.attn_norm.weight | Block 18 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 739 |
+
| 168 | blk.18.attn_output.weight | Block 18 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 740 |
+
| 169 | blk.18.attn_q.weight | Block 18 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 741 |
+
| 170 | blk.18.attn_v.weight | Block 18 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 742 |
+
| 171 | blk.18.ffn_down.weight | Block 18 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 743 |
+
| 172 | blk.18.ffn_gate.weight | Block 18 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 744 |
+
| 173 | blk.18.ffn_norm.weight | Block 18 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 745 |
+
| 174 | blk.18.ffn_up.weight | Block 18 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 746 |
+
|
| 747 |
+
- Total elements in blk.18: (~218M) 218112000
|
| 748 |
+
- Percentage of total elements: 2.72%
|
| 749 |
+
|
| 750 |
+
|
| 751 |
+
### <a name="blk_19">Block 19 Tensor Group : ~218M Elements</a>
|
| 752 |
+
|
| 753 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 754 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 755 |
+
| 175 | blk.19.attn_k.weight | Block 19 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 756 |
+
| 176 | blk.19.attn_norm.weight | Block 19 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 757 |
+
| 177 | blk.19.attn_output.weight | Block 19 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 758 |
+
| 178 | blk.19.attn_q.weight | Block 19 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 759 |
+
| 179 | blk.19.attn_v.weight | Block 19 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 760 |
+
| 180 | blk.19.ffn_down.weight | Block 19 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 761 |
+
| 181 | blk.19.ffn_gate.weight | Block 19 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 762 |
+
| 182 | blk.19.ffn_norm.weight | Block 19 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 763 |
+
| 183 | blk.19.ffn_up.weight | Block 19 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 764 |
+
|
| 765 |
+
- Total elements in blk.19: (~218M) 218112000
|
| 766 |
+
- Percentage of total elements: 2.72%
|
| 767 |
+
|
| 768 |
+
|
| 769 |
+
### <a name="blk_20">Block 20 Tensor Group : ~218M Elements</a>
|
| 770 |
+
|
| 771 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 772 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 773 |
+
| 184 | blk.20.attn_k.weight | Block 20 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 774 |
+
| 185 | blk.20.attn_norm.weight | Block 20 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 775 |
+
| 186 | blk.20.attn_output.weight | Block 20 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 776 |
+
| 187 | blk.20.attn_q.weight | Block 20 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 777 |
+
| 188 | blk.20.attn_v.weight | Block 20 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 778 |
+
| 189 | blk.20.ffn_down.weight | Block 20 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 779 |
+
| 190 | blk.20.ffn_gate.weight | Block 20 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 780 |
+
| 191 | blk.20.ffn_norm.weight | Block 20 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 781 |
+
| 192 | blk.20.ffn_up.weight | Block 20 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 782 |
+
|
| 783 |
+
- Total elements in blk.20: (~218M) 218112000
|
| 784 |
+
- Percentage of total elements: 2.72%
|
| 785 |
+
|
| 786 |
+
|
| 787 |
+
### <a name="blk_21">Block 21 Tensor Group : ~218M Elements</a>
|
| 788 |
+
|
| 789 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 790 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 791 |
+
| 193 | blk.21.attn_k.weight | Block 21 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 792 |
+
| 194 | blk.21.attn_norm.weight | Block 21 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 793 |
+
| 195 | blk.21.attn_output.weight | Block 21 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 794 |
+
| 196 | blk.21.attn_q.weight | Block 21 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 795 |
+
| 197 | blk.21.attn_v.weight | Block 21 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 796 |
+
| 198 | blk.21.ffn_down.weight | Block 21 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 797 |
+
| 199 | blk.21.ffn_gate.weight | Block 21 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 798 |
+
| 200 | blk.21.ffn_norm.weight | Block 21 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 799 |
+
| 201 | blk.21.ffn_up.weight | Block 21 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 800 |
+
|
| 801 |
+
- Total elements in blk.21: (~218M) 218112000
|
| 802 |
+
- Percentage of total elements: 2.72%
|
| 803 |
+
|
| 804 |
+
|
| 805 |
+
### <a name="blk_22">Block 22 Tensor Group : ~218M Elements</a>
|
| 806 |
+
|
| 807 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 808 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 809 |
+
| 202 | blk.22.attn_k.weight | Block 22 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 810 |
+
| 203 | blk.22.attn_norm.weight | Block 22 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 811 |
+
| 204 | blk.22.attn_output.weight | Block 22 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 812 |
+
| 205 | blk.22.attn_q.weight | Block 22 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 813 |
+
| 206 | blk.22.attn_v.weight | Block 22 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 814 |
+
| 207 | blk.22.ffn_down.weight | Block 22 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 815 |
+
| 208 | blk.22.ffn_gate.weight | Block 22 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 816 |
+
| 209 | blk.22.ffn_norm.weight | Block 22 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 817 |
+
| 210 | blk.22.ffn_up.weight | Block 22 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 818 |
+
|
| 819 |
+
- Total elements in blk.22: (~218M) 218112000
|
| 820 |
+
- Percentage of total elements: 2.72%
|
| 821 |
+
|
| 822 |
+
|
| 823 |
+
### <a name="blk_23">Block 23 Tensor Group : ~218M Elements</a>
|
| 824 |
+
|
| 825 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 826 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 827 |
+
| 211 | blk.23.attn_k.weight | Block 23 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 828 |
+
| 212 | blk.23.attn_norm.weight | Block 23 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 829 |
+
| 213 | blk.23.attn_output.weight | Block 23 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 830 |
+
| 214 | blk.23.attn_q.weight | Block 23 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 831 |
+
| 215 | blk.23.attn_v.weight | Block 23 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 832 |
+
| 216 | blk.23.ffn_down.weight | Block 23 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 833 |
+
| 217 | blk.23.ffn_gate.weight | Block 23 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 834 |
+
| 218 | blk.23.ffn_norm.weight | Block 23 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 835 |
+
| 219 | blk.23.ffn_up.weight | Block 23 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 836 |
+
|
| 837 |
+
- Total elements in blk.23: (~218M) 218112000
|
| 838 |
+
- Percentage of total elements: 2.72%
|
| 839 |
+
|
| 840 |
+
|
| 841 |
+
### <a name="blk_24">Block 24 Tensor Group : ~218M Elements</a>
|
| 842 |
+
|
| 843 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 844 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 845 |
+
| 220 | blk.24.attn_k.weight | Block 24 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 846 |
+
| 221 | blk.24.attn_norm.weight | Block 24 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 847 |
+
| 222 | blk.24.attn_output.weight | Block 24 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 848 |
+
| 223 | blk.24.attn_q.weight | Block 24 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 849 |
+
| 224 | blk.24.attn_v.weight | Block 24 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 850 |
+
| 225 | blk.24.ffn_down.weight | Block 24 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 851 |
+
| 226 | blk.24.ffn_gate.weight | Block 24 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 852 |
+
| 227 | blk.24.ffn_norm.weight | Block 24 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 853 |
+
| 228 | blk.24.ffn_up.weight | Block 24 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 854 |
+
|
| 855 |
+
- Total elements in blk.24: (~218M) 218112000
|
| 856 |
+
- Percentage of total elements: 2.72%
|
| 857 |
+
|
| 858 |
+
|
| 859 |
+
### <a name="blk_25">Block 25 Tensor Group : ~218M Elements</a>
|
| 860 |
+
|
| 861 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 862 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 863 |
+
| 229 | blk.25.attn_k.weight | Block 25 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 864 |
+
| 230 | blk.25.attn_norm.weight | Block 25 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 865 |
+
| 231 | blk.25.attn_output.weight | Block 25 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 866 |
+
| 232 | blk.25.attn_q.weight | Block 25 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 867 |
+
| 233 | blk.25.attn_v.weight | Block 25 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 868 |
+
| 234 | blk.25.ffn_down.weight | Block 25 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 869 |
+
| 235 | blk.25.ffn_gate.weight | Block 25 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 870 |
+
| 236 | blk.25.ffn_norm.weight | Block 25 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 871 |
+
| 237 | blk.25.ffn_up.weight | Block 25 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 872 |
+
|
| 873 |
+
- Total elements in blk.25: (~218M) 218112000
|
| 874 |
+
- Percentage of total elements: 2.72%
|
| 875 |
+
|
| 876 |
+
|
| 877 |
+
### <a name="blk_26">Block 26 Tensor Group : ~218M Elements</a>
|
| 878 |
+
|
| 879 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 880 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 881 |
+
| 238 | blk.26.attn_k.weight | Block 26 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 882 |
+
| 239 | blk.26.attn_norm.weight | Block 26 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 883 |
+
| 240 | blk.26.attn_output.weight | Block 26 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 884 |
+
| 241 | blk.26.attn_q.weight | Block 26 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 885 |
+
| 242 | blk.26.attn_v.weight | Block 26 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 886 |
+
| 243 | blk.26.ffn_down.weight | Block 26 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 887 |
+
| 244 | blk.26.ffn_gate.weight | Block 26 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 888 |
+
| 245 | blk.26.ffn_norm.weight | Block 26 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 889 |
+
| 246 | blk.26.ffn_up.weight | Block 26 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 890 |
+
|
| 891 |
+
- Total elements in blk.26: (~218M) 218112000
|
| 892 |
+
- Percentage of total elements: 2.72%
|
| 893 |
+
|
| 894 |
+
|
| 895 |
+
### <a name="blk_27">Block 27 Tensor Group : ~218M Elements</a>
|
| 896 |
+
|
| 897 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 898 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 899 |
+
| 247 | blk.27.attn_k.weight | Block 27 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 900 |
+
| 248 | blk.27.attn_norm.weight | Block 27 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 901 |
+
| 249 | blk.27.attn_output.weight | Block 27 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 902 |
+
| 250 | blk.27.attn_q.weight | Block 27 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 903 |
+
| 251 | blk.27.attn_v.weight | Block 27 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 904 |
+
| 252 | blk.27.ffn_down.weight | Block 27 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 905 |
+
| 253 | blk.27.ffn_gate.weight | Block 27 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 906 |
+
| 254 | blk.27.ffn_norm.weight | Block 27 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 907 |
+
| 255 | blk.27.ffn_up.weight | Block 27 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 908 |
+
|
| 909 |
+
- Total elements in blk.27: (~218M) 218112000
|
| 910 |
+
- Percentage of total elements: 2.72%
|
| 911 |
+
|
| 912 |
+
|
| 913 |
+
### <a name="blk_28">Block 28 Tensor Group : ~218M Elements</a>
|
| 914 |
+
|
| 915 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 916 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 917 |
+
| 256 | blk.28.attn_k.weight | Block 28 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 918 |
+
| 257 | blk.28.attn_norm.weight | Block 28 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 919 |
+
| 258 | blk.28.attn_output.weight | Block 28 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 920 |
+
| 259 | blk.28.attn_q.weight | Block 28 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 921 |
+
| 260 | blk.28.attn_v.weight | Block 28 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 922 |
+
| 261 | blk.28.ffn_down.weight | Block 28 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 923 |
+
| 262 | blk.28.ffn_gate.weight | Block 28 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 924 |
+
| 263 | blk.28.ffn_norm.weight | Block 28 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 925 |
+
| 264 | blk.28.ffn_up.weight | Block 28 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 926 |
+
|
| 927 |
+
- Total elements in blk.28: (~218M) 218112000
|
| 928 |
+
- Percentage of total elements: 2.72%
|
| 929 |
+
|
| 930 |
+
|
| 931 |
+
### <a name="blk_29">Block 29 Tensor Group : ~218M Elements</a>
|
| 932 |
+
|
| 933 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 934 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 935 |
+
| 265 | blk.29.attn_k.weight | Block 29 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 936 |
+
| 266 | blk.29.attn_norm.weight | Block 29 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 937 |
+
| 267 | blk.29.attn_output.weight | Block 29 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 938 |
+
| 268 | blk.29.attn_q.weight | Block 29 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 939 |
+
| 269 | blk.29.attn_v.weight | Block 29 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 940 |
+
| 270 | blk.29.ffn_down.weight | Block 29 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 941 |
+
| 271 | blk.29.ffn_gate.weight | Block 29 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 942 |
+
| 272 | blk.29.ffn_norm.weight | Block 29 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 943 |
+
| 273 | blk.29.ffn_up.weight | Block 29 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 944 |
+
|
| 945 |
+
- Total elements in blk.29: (~218M) 218112000
|
| 946 |
+
- Percentage of total elements: 2.72%
|
| 947 |
+
|
| 948 |
+
|
| 949 |
+
### <a name="blk_30">Block 30 Tensor Group : ~218M Elements</a>
|
| 950 |
+
|
| 951 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 952 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 953 |
+
| 274 | blk.30.attn_k.weight | Block 30 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 954 |
+
| 275 | blk.30.attn_norm.weight | Block 30 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 955 |
+
| 276 | blk.30.attn_output.weight | Block 30 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 956 |
+
| 277 | blk.30.attn_q.weight | Block 30 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 957 |
+
| 278 | blk.30.attn_v.weight | Block 30 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 958 |
+
| 279 | blk.30.ffn_down.weight | Block 30 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 959 |
+
| 280 | blk.30.ffn_gate.weight | Block 30 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 960 |
+
| 281 | blk.30.ffn_norm.weight | Block 30 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 961 |
+
| 282 | blk.30.ffn_up.weight | Block 30 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 962 |
+
|
| 963 |
+
- Total elements in blk.30: (~218M) 218112000
|
| 964 |
+
- Percentage of total elements: 2.72%
|
| 965 |
+
|
| 966 |
+
|
| 967 |
+
### <a name="blk_31">Block 31 Tensor Group : ~218M Elements</a>
|
| 968 |
+
|
| 969 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 970 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:--------|
|
| 971 |
+
| 283 | blk.31.attn_k.weight | Block 31 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_XXS |
|
| 972 |
+
| 284 | blk.31.attn_norm.weight | Block 31 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 973 |
+
| 285 | blk.31.attn_output.weight | Block 31 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 974 |
+
| 286 | blk.31.attn_q.weight | Block 31 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_XXS |
|
| 975 |
+
| 287 | blk.31.attn_v.weight | Block 31 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 976 |
+
| 288 | blk.31.ffn_down.weight | Block 31 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | IQ4_NL |
|
| 977 |
+
| 289 | blk.31.ffn_gate.weight | Block 31 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 978 |
+
| 290 | blk.31.ffn_norm.weight | Block 31 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 979 |
+
| 291 | blk.31.ffn_up.weight | Block 31 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 980 |
+
|
| 981 |
+
- Total elements in blk.31: (~218M) 218112000
|
| 982 |
+
- Percentage of total elements: 2.72%
|
scores/Watt-Tool-8B-IQ4_NL.md
ADDED
|
@@ -0,0 +1,982 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Watt-Tool-8B-IQ4_NL.gguf - GGUF Internal File Dump
|
| 2 |
+
|
| 3 |
+
- Endian: LITTLE endian
|
| 4 |
+
|
| 5 |
+
## Key Value Metadata Store
|
| 6 |
+
|
| 7 |
+
There are 43 key-value pairs in this file
|
| 8 |
+
|
| 9 |
+
| POS | TYPE | Count | Key | Value |
|
| 10 |
+
|----:|:---------|-------:|:---------------------------------------|:--------------------------------------------------------------------|
|
| 11 |
+
| 1 | UINT32 | 1 | GGUF.version | 3 |
|
| 12 |
+
| 2 | UINT64 | 1 | GGUF.tensor_count | 292 |
|
| 13 |
+
| 3 | UINT64 | 1 | GGUF.kv_count | 40 |
|
| 14 |
+
| 4 | STRING | 1 | general.architecture | `llama` |
|
| 15 |
+
| 5 | STRING | 1 | general.type | `model` |
|
| 16 |
+
| 6 | STRING | 1 | general.name | `Watt Tool 8B GGUF` |
|
| 17 |
+
| 7 | STRING | 1 | general.finetune | `GGUF` |
|
| 18 |
+
| 8 | STRING | 1 | general.basename | `Watt-Tool` |
|
| 19 |
+
| 9 | STRING | 1 | general.size_label | `8B` |
|
| 20 |
+
| 10 | STRING | 1 | general.license | `apache-2.0` |
|
| 21 |
+
| 11 | UINT32 | 1 | general.base_model.count | 1 |
|
| 22 |
+
| 12 | STRING | 1 | general.base_model.0.name | `Llama 3.1 8B Instruct` |
|
| 23 |
+
| 13 | STRING | 1 | general.base_model.0.organization | `Meta Llama` |
|
| 24 |
+
| 14 | STRING | 1 | general.base_model.0.repo_url | `https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct` |
|
| 25 |
+
| 15 | [STRING] | 4 | general.tags | [ `function-calling`, `tool-use`, `llama`, `bfcl` ] |
|
| 26 |
+
| 16 | [STRING] | 1 | general.languages | [ `en` ] |
|
| 27 |
+
| 17 | UINT32 | 1 | llama.block_count | 32 |
|
| 28 |
+
| 18 | UINT32 | 1 | llama.context_length | 131072 |
|
| 29 |
+
| 19 | UINT32 | 1 | llama.embedding_length | 4096 |
|
| 30 |
+
| 20 | UINT32 | 1 | llama.feed_forward_length | 14336 |
|
| 31 |
+
| 21 | UINT32 | 1 | llama.attention.head_count | 32 |
|
| 32 |
+
| 22 | UINT32 | 1 | llama.attention.head_count_kv | 8 |
|
| 33 |
+
| 23 | FLOAT32 | 1 | llama.rope.freq_base | 500000.0 |
|
| 34 |
+
| 24 | FLOAT32 | 1 | llama.attention.layer_norm_rms_epsilon | 1e-05 |
|
| 35 |
+
| 25 | UINT32 | 1 | llama.attention.key_length | 128 |
|
| 36 |
+
| 26 | UINT32 | 1 | llama.attention.value_length | 128 |
|
| 37 |
+
| 27 | UINT32 | 1 | llama.vocab_size | 128256 |
|
| 38 |
+
| 28 | UINT32 | 1 | llama.rope.dimension_count | 128 |
|
| 39 |
+
| 29 | STRING | 1 | tokenizer.ggml.model | `gpt2` |
|
| 40 |
+
| 30 | STRING | 1 | tokenizer.ggml.pre | `llama-bpe` |
|
| 41 |
+
| 31 | [STRING] | 128256 | tokenizer.ggml.tokens | [ `!`, `"`, `#`, `$`, `%`, ... ] |
|
| 42 |
+
| 32 | [INT32] | 128256 | tokenizer.ggml.token_type | [ 1, 1, 1, 1, 1, 1, 1, ... ] |
|
| 43 |
+
| 33 | [STRING] | 280147 | tokenizer.ggml.merges | [ `Ġ Ġ`, `Ġ ĠĠĠ`, `ĠĠ ĠĠ`, `ĠĠĠ Ġ`, `i n`, ... ] |
|
| 44 |
+
| 34 | UINT32 | 1 | tokenizer.ggml.bos_token_id | 128000 |
|
| 45 |
+
| 35 | UINT32 | 1 | tokenizer.ggml.eos_token_id | 128009 |
|
| 46 |
+
| 36 | UINT32 | 1 | tokenizer.ggml.padding_token_id | 128009 |
|
| 47 |
+
| 37 | STRING | 1 | tokenizer.chat_template | `{{ '<|begin_of_text|>' }}{% if`...`d|>' }}{% endif %}{% endfor %}` |
|
| 48 |
+
| 38 | UINT32 | 1 | general.quantization_version | 2 |
|
| 49 |
+
| 39 | UINT32 | 1 | general.file_type | 25 |
|
| 50 |
+
| 40 | STRING | 1 | quantize.imatrix.file | `./imatrix/imatrix-Watt-Tool-8B-small.dat` |
|
| 51 |
+
| 41 | STRING | 1 | quantize.imatrix.dataset | `../../datasets/imatrix/calibration_eur_small.txt` |
|
| 52 |
+
| 42 | INT32 | 1 | quantize.imatrix.entries_count | 225 |
|
| 53 |
+
| 43 | INT32 | 1 | quantize.imatrix.chunks_count | 962 |
|
| 54 |
+
|
| 55 |
+
## Tensors Overview ~8B Elements
|
| 56 |
+
|
| 57 |
+
Total number of elements in all tensors: 8030261312 Elements
|
| 58 |
+
|
| 59 |
+
- [Watt-Tool-8B-IQ4\_NL.gguf - GGUF Internal File Dump](#watt-tool-8b-iq4_nlgguf---gguf-internal-file-dump)
|
| 60 |
+
- [Key Value Metadata Store](#key-value-metadata-store)
|
| 61 |
+
- [Tensors Overview ~8B Elements](#tensors-overview-8b-elements)
|
| 62 |
+
- [Tensor Data Offset](#tensor-data-offset)
|
| 63 |
+
- [Base Tensor Group : ~1B Elements](#base-tensor-group--1b-elements)
|
| 64 |
+
- [Block 0 Tensor Group : ~218M Elements](#block-0-tensor-group--218m-elements)
|
| 65 |
+
- [Block 1 Tensor Group : ~218M Elements](#block-1-tensor-group--218m-elements)
|
| 66 |
+
- [Block 2 Tensor Group : ~218M Elements](#block-2-tensor-group--218m-elements)
|
| 67 |
+
- [Block 3 Tensor Group : ~218M Elements](#block-3-tensor-group--218m-elements)
|
| 68 |
+
- [Block 4 Tensor Group : ~218M Elements](#block-4-tensor-group--218m-elements)
|
| 69 |
+
- [Block 5 Tensor Group : ~218M Elements](#block-5-tensor-group--218m-elements)
|
| 70 |
+
- [Block 6 Tensor Group : ~218M Elements](#block-6-tensor-group--218m-elements)
|
| 71 |
+
- [Block 7 Tensor Group : ~218M Elements](#block-7-tensor-group--218m-elements)
|
| 72 |
+
- [Block 8 Tensor Group : ~218M Elements](#block-8-tensor-group--218m-elements)
|
| 73 |
+
- [Block 9 Tensor Group : ~218M Elements](#block-9-tensor-group--218m-elements)
|
| 74 |
+
- [Block 10 Tensor Group : ~218M Elements](#block-10-tensor-group--218m-elements)
|
| 75 |
+
- [Block 11 Tensor Group : ~218M Elements](#block-11-tensor-group--218m-elements)
|
| 76 |
+
- [Block 12 Tensor Group : ~218M Elements](#block-12-tensor-group--218m-elements)
|
| 77 |
+
- [Block 13 Tensor Group : ~218M Elements](#block-13-tensor-group--218m-elements)
|
| 78 |
+
- [Block 14 Tensor Group : ~218M Elements](#block-14-tensor-group--218m-elements)
|
| 79 |
+
- [Block 15 Tensor Group : ~218M Elements](#block-15-tensor-group--218m-elements)
|
| 80 |
+
- [Block 16 Tensor Group : ~218M Elements](#block-16-tensor-group--218m-elements)
|
| 81 |
+
- [Block 17 Tensor Group : ~218M Elements](#block-17-tensor-group--218m-elements)
|
| 82 |
+
- [Block 18 Tensor Group : ~218M Elements](#block-18-tensor-group--218m-elements)
|
| 83 |
+
- [Block 19 Tensor Group : ~218M Elements](#block-19-tensor-group--218m-elements)
|
| 84 |
+
- [Block 20 Tensor Group : ~218M Elements](#block-20-tensor-group--218m-elements)
|
| 85 |
+
- [Block 21 Tensor Group : ~218M Elements](#block-21-tensor-group--218m-elements)
|
| 86 |
+
- [Block 22 Tensor Group : ~218M Elements](#block-22-tensor-group--218m-elements)
|
| 87 |
+
- [Block 23 Tensor Group : ~218M Elements](#block-23-tensor-group--218m-elements)
|
| 88 |
+
- [Block 24 Tensor Group : ~218M Elements](#block-24-tensor-group--218m-elements)
|
| 89 |
+
- [Block 25 Tensor Group : ~218M Elements](#block-25-tensor-group--218m-elements)
|
| 90 |
+
- [Block 26 Tensor Group : ~218M Elements](#block-26-tensor-group--218m-elements)
|
| 91 |
+
- [Block 27 Tensor Group : ~218M Elements](#block-27-tensor-group--218m-elements)
|
| 92 |
+
- [Block 28 Tensor Group : ~218M Elements](#block-28-tensor-group--218m-elements)
|
| 93 |
+
- [Block 29 Tensor Group : ~218M Elements](#block-29-tensor-group--218m-elements)
|
| 94 |
+
- [Block 30 Tensor Group : ~218M Elements](#block-30-tensor-group--218m-elements)
|
| 95 |
+
- [Block 31 Tensor Group : ~218M Elements](#block-31-tensor-group--218m-elements)
|
| 96 |
+
|
| 97 |
+
### Tensor Data Offset
|
| 98 |
+
|
| 99 |
+
This table contains the offset and data segment relative to start of file
|
| 100 |
+
|
| 101 |
+
| T_ID | Tensor Layer Name | Data Offset (B) | Data Size (B) |
|
| 102 |
+
|-----:|:--------------------------|-----------------:|-----------------:|
|
| 103 |
+
| 0 | output.weight | 0x779620 | 0x119d0000 |
|
| 104 |
+
| 1 | output_norm.weight | 0x12149620 | 0x4000 |
|
| 105 |
+
| 2 | rope_freqs.weight | 0x1214d620 | 0x100 |
|
| 106 |
+
| 3 | token_embd.weight | 0x1214d720 | 0xd746000 |
|
| 107 |
+
| 4 | blk.0.attn_k.weight | 0x1f893720 | 0x1b8000 |
|
| 108 |
+
| 5 | blk.0.attn_norm.weight | 0x1fa4b720 | 0x4000 |
|
| 109 |
+
| 6 | blk.0.attn_output.weight | 0x1fa4f720 | 0x900000 |
|
| 110 |
+
| 7 | blk.0.attn_q.weight | 0x2034f720 | 0x6e0000 |
|
| 111 |
+
| 8 | blk.0.attn_v.weight | 0x20a2f720 | 0x220000 |
|
| 112 |
+
| 9 | blk.0.ffn_down.weight | 0x20c4f720 | 0x2680000 |
|
| 113 |
+
| 10 | blk.0.ffn_gate.weight | 0x232cf720 | 0x1810000 |
|
| 114 |
+
| 11 | blk.0.ffn_norm.weight | 0x24adf720 | 0x4000 |
|
| 115 |
+
| 12 | blk.0.ffn_up.weight | 0x24ae3720 | 0x1810000 |
|
| 116 |
+
| 13 | blk.1.attn_k.weight | 0x262f3720 | 0x1b8000 |
|
| 117 |
+
| 14 | blk.1.attn_norm.weight | 0x264ab720 | 0x4000 |
|
| 118 |
+
| 15 | blk.1.attn_output.weight | 0x264af720 | 0x900000 |
|
| 119 |
+
| 16 | blk.1.attn_q.weight | 0x26daf720 | 0x6e0000 |
|
| 120 |
+
| 17 | blk.1.attn_v.weight | 0x2748f720 | 0x220000 |
|
| 121 |
+
| 18 | blk.1.ffn_down.weight | 0x276af720 | 0x2680000 |
|
| 122 |
+
| 19 | blk.1.ffn_gate.weight | 0x29d2f720 | 0x1810000 |
|
| 123 |
+
| 20 | blk.1.ffn_norm.weight | 0x2b53f720 | 0x4000 |
|
| 124 |
+
| 21 | blk.1.ffn_up.weight | 0x2b543720 | 0x1810000 |
|
| 125 |
+
| 22 | blk.2.attn_k.weight | 0x2cd53720 | 0x1b8000 |
|
| 126 |
+
| 23 | blk.2.attn_norm.weight | 0x2cf0b720 | 0x4000 |
|
| 127 |
+
| 24 | blk.2.attn_output.weight | 0x2cf0f720 | 0x900000 |
|
| 128 |
+
| 25 | blk.2.attn_q.weight | 0x2d80f720 | 0x6e0000 |
|
| 129 |
+
| 26 | blk.2.attn_v.weight | 0x2deef720 | 0x220000 |
|
| 130 |
+
| 27 | blk.2.ffn_down.weight | 0x2e10f720 | 0x2680000 |
|
| 131 |
+
| 28 | blk.2.ffn_gate.weight | 0x3078f720 | 0x1810000 |
|
| 132 |
+
| 29 | blk.2.ffn_norm.weight | 0x31f9f720 | 0x4000 |
|
| 133 |
+
| 30 | blk.2.ffn_up.weight | 0x31fa3720 | 0x1810000 |
|
| 134 |
+
| 31 | blk.3.attn_k.weight | 0x337b3720 | 0x1b8000 |
|
| 135 |
+
| 32 | blk.3.attn_norm.weight | 0x3396b720 | 0x4000 |
|
| 136 |
+
| 33 | blk.3.attn_output.weight | 0x3396f720 | 0x900000 |
|
| 137 |
+
| 34 | blk.3.attn_q.weight | 0x3426f720 | 0x6e0000 |
|
| 138 |
+
| 35 | blk.3.attn_v.weight | 0x3494f720 | 0x220000 |
|
| 139 |
+
| 36 | blk.3.ffn_down.weight | 0x34b6f720 | 0x2680000 |
|
| 140 |
+
| 37 | blk.3.ffn_gate.weight | 0x371ef720 | 0x1810000 |
|
| 141 |
+
| 38 | blk.3.ffn_norm.weight | 0x389ff720 | 0x4000 |
|
| 142 |
+
| 39 | blk.3.ffn_up.weight | 0x38a03720 | 0x1810000 |
|
| 143 |
+
| 40 | blk.4.attn_k.weight | 0x3a213720 | 0x1b8000 |
|
| 144 |
+
| 41 | blk.4.attn_norm.weight | 0x3a3cb720 | 0x4000 |
|
| 145 |
+
| 42 | blk.4.attn_output.weight | 0x3a3cf720 | 0x900000 |
|
| 146 |
+
| 43 | blk.4.attn_q.weight | 0x3accf720 | 0x6e0000 |
|
| 147 |
+
| 44 | blk.4.attn_v.weight | 0x3b3af720 | 0x220000 |
|
| 148 |
+
| 45 | blk.4.ffn_down.weight | 0x3b5cf720 | 0x2680000 |
|
| 149 |
+
| 46 | blk.4.ffn_gate.weight | 0x3dc4f720 | 0x1810000 |
|
| 150 |
+
| 47 | blk.4.ffn_norm.weight | 0x3f45f720 | 0x4000 |
|
| 151 |
+
| 48 | blk.4.ffn_up.weight | 0x3f463720 | 0x1810000 |
|
| 152 |
+
| 49 | blk.5.attn_k.weight | 0x40c73720 | 0x1b8000 |
|
| 153 |
+
| 50 | blk.5.attn_norm.weight | 0x40e2b720 | 0x4000 |
|
| 154 |
+
| 51 | blk.5.attn_output.weight | 0x40e2f720 | 0x900000 |
|
| 155 |
+
| 52 | blk.5.attn_q.weight | 0x4172f720 | 0x6e0000 |
|
| 156 |
+
| 53 | blk.5.attn_v.weight | 0x41e0f720 | 0x220000 |
|
| 157 |
+
| 54 | blk.5.ffn_down.weight | 0x4202f720 | 0x2680000 |
|
| 158 |
+
| 55 | blk.5.ffn_gate.weight | 0x446af720 | 0x1810000 |
|
| 159 |
+
| 56 | blk.5.ffn_norm.weight | 0x45ebf720 | 0x4000 |
|
| 160 |
+
| 57 | blk.5.ffn_up.weight | 0x45ec3720 | 0x1810000 |
|
| 161 |
+
| 58 | blk.6.attn_k.weight | 0x476d3720 | 0x1b8000 |
|
| 162 |
+
| 59 | blk.6.attn_norm.weight | 0x4788b720 | 0x4000 |
|
| 163 |
+
| 60 | blk.6.attn_output.weight | 0x4788f720 | 0x900000 |
|
| 164 |
+
| 61 | blk.6.attn_q.weight | 0x4818f720 | 0x6e0000 |
|
| 165 |
+
| 62 | blk.6.attn_v.weight | 0x4886f720 | 0x220000 |
|
| 166 |
+
| 63 | blk.6.ffn_down.weight | 0x48a8f720 | 0x2680000 |
|
| 167 |
+
| 64 | blk.6.ffn_gate.weight | 0x4b10f720 | 0x1810000 |
|
| 168 |
+
| 65 | blk.6.ffn_norm.weight | 0x4c91f720 | 0x4000 |
|
| 169 |
+
| 66 | blk.6.ffn_up.weight | 0x4c923720 | 0x1810000 |
|
| 170 |
+
| 67 | blk.7.attn_k.weight | 0x4e133720 | 0x1b8000 |
|
| 171 |
+
| 68 | blk.7.attn_norm.weight | 0x4e2eb720 | 0x4000 |
|
| 172 |
+
| 69 | blk.7.attn_output.weight | 0x4e2ef720 | 0x900000 |
|
| 173 |
+
| 70 | blk.7.attn_q.weight | 0x4ebef720 | 0x6e0000 |
|
| 174 |
+
| 71 | blk.7.attn_v.weight | 0x4f2cf720 | 0x220000 |
|
| 175 |
+
| 72 | blk.7.ffn_down.weight | 0x4f4ef720 | 0x2680000 |
|
| 176 |
+
| 73 | blk.7.ffn_gate.weight | 0x51b6f720 | 0x1810000 |
|
| 177 |
+
| 74 | blk.7.ffn_norm.weight | 0x5337f720 | 0x4000 |
|
| 178 |
+
| 75 | blk.7.ffn_up.weight | 0x53383720 | 0x1810000 |
|
| 179 |
+
| 76 | blk.8.attn_k.weight | 0x54b93720 | 0x1b8000 |
|
| 180 |
+
| 77 | blk.8.attn_norm.weight | 0x54d4b720 | 0x4000 |
|
| 181 |
+
| 78 | blk.8.attn_output.weight | 0x54d4f720 | 0x900000 |
|
| 182 |
+
| 79 | blk.8.attn_q.weight | 0x5564f720 | 0x6e0000 |
|
| 183 |
+
| 80 | blk.8.attn_v.weight | 0x55d2f720 | 0x220000 |
|
| 184 |
+
| 81 | blk.8.ffn_down.weight | 0x55f4f720 | 0x2680000 |
|
| 185 |
+
| 82 | blk.8.ffn_gate.weight | 0x585cf720 | 0x1810000 |
|
| 186 |
+
| 83 | blk.8.ffn_norm.weight | 0x59ddf720 | 0x4000 |
|
| 187 |
+
| 84 | blk.8.ffn_up.weight | 0x59de3720 | 0x1810000 |
|
| 188 |
+
| 85 | blk.9.attn_k.weight | 0x5b5f3720 | 0x1b8000 |
|
| 189 |
+
| 86 | blk.9.attn_norm.weight | 0x5b7ab720 | 0x4000 |
|
| 190 |
+
| 87 | blk.9.attn_output.weight | 0x5b7af720 | 0x900000 |
|
| 191 |
+
| 88 | blk.9.attn_q.weight | 0x5c0af720 | 0x6e0000 |
|
| 192 |
+
| 89 | blk.9.attn_v.weight | 0x5c78f720 | 0x220000 |
|
| 193 |
+
| 90 | blk.9.ffn_down.weight | 0x5c9af720 | 0x2680000 |
|
| 194 |
+
| 91 | blk.9.ffn_gate.weight | 0x5f02f720 | 0x1810000 |
|
| 195 |
+
| 92 | blk.9.ffn_norm.weight | 0x6083f720 | 0x4000 |
|
| 196 |
+
| 93 | blk.9.ffn_up.weight | 0x60843720 | 0x1810000 |
|
| 197 |
+
| 94 | blk.10.attn_k.weight | 0x62053720 | 0x1b8000 |
|
| 198 |
+
| 95 | blk.10.attn_norm.weight | 0x6220b720 | 0x4000 |
|
| 199 |
+
| 96 | blk.10.attn_output.weight | 0x6220f720 | 0x900000 |
|
| 200 |
+
| 97 | blk.10.attn_q.weight | 0x62b0f720 | 0x6e0000 |
|
| 201 |
+
| 98 | blk.10.attn_v.weight | 0x631ef720 | 0x220000 |
|
| 202 |
+
| 99 | blk.10.ffn_down.weight | 0x6340f720 | 0x2680000 |
|
| 203 |
+
| 100 | blk.10.ffn_gate.weight | 0x65a8f720 | 0x1810000 |
|
| 204 |
+
| 101 | blk.10.ffn_norm.weight | 0x6729f720 | 0x4000 |
|
| 205 |
+
| 102 | blk.10.ffn_up.weight | 0x672a3720 | 0x1810000 |
|
| 206 |
+
| 103 | blk.11.attn_k.weight | 0x68ab3720 | 0x1b8000 |
|
| 207 |
+
| 104 | blk.11.attn_norm.weight | 0x68c6b720 | 0x4000 |
|
| 208 |
+
| 105 | blk.11.attn_output.weight | 0x68c6f720 | 0x900000 |
|
| 209 |
+
| 106 | blk.11.attn_q.weight | 0x6956f720 | 0x6e0000 |
|
| 210 |
+
| 107 | blk.11.attn_v.weight | 0x69c4f720 | 0x220000 |
|
| 211 |
+
| 108 | blk.11.ffn_down.weight | 0x69e6f720 | 0x2680000 |
|
| 212 |
+
| 109 | blk.11.ffn_gate.weight | 0x6c4ef720 | 0x1810000 |
|
| 213 |
+
| 110 | blk.11.ffn_norm.weight | 0x6dcff720 | 0x4000 |
|
| 214 |
+
| 111 | blk.11.ffn_up.weight | 0x6dd03720 | 0x1810000 |
|
| 215 |
+
| 112 | blk.12.attn_k.weight | 0x6f513720 | 0x1b8000 |
|
| 216 |
+
| 113 | blk.12.attn_norm.weight | 0x6f6cb720 | 0x4000 |
|
| 217 |
+
| 114 | blk.12.attn_output.weight | 0x6f6cf720 | 0x900000 |
|
| 218 |
+
| 115 | blk.12.attn_q.weight | 0x6ffcf720 | 0x6e0000 |
|
| 219 |
+
| 116 | blk.12.attn_v.weight | 0x706af720 | 0x220000 |
|
| 220 |
+
| 117 | blk.12.ffn_down.weight | 0x708cf720 | 0x2680000 |
|
| 221 |
+
| 118 | blk.12.ffn_gate.weight | 0x72f4f720 | 0x1810000 |
|
| 222 |
+
| 119 | blk.12.ffn_norm.weight | 0x7475f720 | 0x4000 |
|
| 223 |
+
| 120 | blk.12.ffn_up.weight | 0x74763720 | 0x1810000 |
|
| 224 |
+
| 121 | blk.13.attn_k.weight | 0x75f73720 | 0x240000 |
|
| 225 |
+
| 122 | blk.13.attn_norm.weight | 0x761b3720 | 0x4000 |
|
| 226 |
+
| 123 | blk.13.attn_output.weight | 0x761b7720 | 0x900000 |
|
| 227 |
+
| 124 | blk.13.attn_q.weight | 0x76ab7720 | 0x900000 |
|
| 228 |
+
| 125 | blk.13.attn_v.weight | 0x773b7720 | 0x240000 |
|
| 229 |
+
| 126 | blk.13.ffn_down.weight | 0x775f7720 | 0x2680000 |
|
| 230 |
+
| 127 | blk.13.ffn_gate.weight | 0x79c77720 | 0x1810000 |
|
| 231 |
+
| 128 | blk.13.ffn_norm.weight | 0x7b487720 | 0x4000 |
|
| 232 |
+
| 129 | blk.13.ffn_up.weight | 0x7b48b720 | 0x1810000 |
|
| 233 |
+
| 130 | blk.14.attn_k.weight | 0x7cc9b720 | 0x240000 |
|
| 234 |
+
| 131 | blk.14.attn_norm.weight | 0x7cedb720 | 0x4000 |
|
| 235 |
+
| 132 | blk.14.attn_output.weight | 0x7cedf720 | 0x900000 |
|
| 236 |
+
| 133 | blk.14.attn_q.weight | 0x7d7df720 | 0x900000 |
|
| 237 |
+
| 134 | blk.14.attn_v.weight | 0x7e0df720 | 0x240000 |
|
| 238 |
+
| 135 | blk.14.ffn_down.weight | 0x7e31f720 | 0x2680000 |
|
| 239 |
+
| 136 | blk.14.ffn_gate.weight | 0x8099f720 | 0x1810000 |
|
| 240 |
+
| 137 | blk.14.ffn_norm.weight | 0x821af720 | 0x4000 |
|
| 241 |
+
| 138 | blk.14.ffn_up.weight | 0x821b3720 | 0x1810000 |
|
| 242 |
+
| 139 | blk.15.attn_k.weight | 0x839c3720 | 0x1b8000 |
|
| 243 |
+
| 140 | blk.15.attn_norm.weight | 0x83b7b720 | 0x4000 |
|
| 244 |
+
| 141 | blk.15.attn_output.weight | 0x83b7f720 | 0x900000 |
|
| 245 |
+
| 142 | blk.15.attn_q.weight | 0x8447f720 | 0x6e0000 |
|
| 246 |
+
| 143 | blk.15.attn_v.weight | 0x84b5f720 | 0x220000 |
|
| 247 |
+
| 144 | blk.15.ffn_down.weight | 0x84d7f720 | 0x2680000 |
|
| 248 |
+
| 145 | blk.15.ffn_gate.weight | 0x873ff720 | 0x1810000 |
|
| 249 |
+
| 146 | blk.15.ffn_norm.weight | 0x88c0f720 | 0x4000 |
|
| 250 |
+
| 147 | blk.15.ffn_up.weight | 0x88c13720 | 0x1810000 |
|
| 251 |
+
| 148 | blk.16.attn_k.weight | 0x8a423720 | 0x240000 |
|
| 252 |
+
| 149 | blk.16.attn_norm.weight | 0x8a663720 | 0x4000 |
|
| 253 |
+
| 150 | blk.16.attn_output.weight | 0x8a667720 | 0x900000 |
|
| 254 |
+
| 151 | blk.16.attn_q.weight | 0x8af67720 | 0x900000 |
|
| 255 |
+
| 152 | blk.16.attn_v.weight | 0x8b867720 | 0x240000 |
|
| 256 |
+
| 153 | blk.16.ffn_down.weight | 0x8baa7720 | 0x2680000 |
|
| 257 |
+
| 154 | blk.16.ffn_gate.weight | 0x8e127720 | 0x1f80000 |
|
| 258 |
+
| 155 | blk.16.ffn_norm.weight | 0x900a7720 | 0x4000 |
|
| 259 |
+
| 156 | blk.16.ffn_up.weight | 0x900ab720 | 0x1f80000 |
|
| 260 |
+
| 157 | blk.17.attn_k.weight | 0x9202b720 | 0x1b8000 |
|
| 261 |
+
| 158 | blk.17.attn_norm.weight | 0x921e3720 | 0x4000 |
|
| 262 |
+
| 159 | blk.17.attn_output.weight | 0x921e7720 | 0x900000 |
|
| 263 |
+
| 160 | blk.17.attn_q.weight | 0x92ae7720 | 0x6e0000 |
|
| 264 |
+
| 161 | blk.17.attn_v.weight | 0x931c7720 | 0x220000 |
|
| 265 |
+
| 162 | blk.17.ffn_down.weight | 0x933e7720 | 0x2680000 |
|
| 266 |
+
| 163 | blk.17.ffn_gate.weight | 0x95a67720 | 0x1f80000 |
|
| 267 |
+
| 164 | blk.17.ffn_norm.weight | 0x979e7720 | 0x4000 |
|
| 268 |
+
| 165 | blk.17.ffn_up.weight | 0x979eb720 | 0x1f80000 |
|
| 269 |
+
| 166 | blk.18.attn_k.weight | 0x9996b720 | 0x240000 |
|
| 270 |
+
| 167 | blk.18.attn_norm.weight | 0x99bab720 | 0x4000 |
|
| 271 |
+
| 168 | blk.18.attn_output.weight | 0x99baf720 | 0x900000 |
|
| 272 |
+
| 169 | blk.18.attn_q.weight | 0x9a4af720 | 0x900000 |
|
| 273 |
+
| 170 | blk.18.attn_v.weight | 0x9adaf720 | 0x240000 |
|
| 274 |
+
| 171 | blk.18.ffn_down.weight | 0x9afef720 | 0x2680000 |
|
| 275 |
+
| 172 | blk.18.ffn_gate.weight | 0x9d66f720 | 0x1f80000 |
|
| 276 |
+
| 173 | blk.18.ffn_norm.weight | 0x9f5ef720 | 0x4000 |
|
| 277 |
+
| 174 | blk.18.ffn_up.weight | 0x9f5f3720 | 0x1f80000 |
|
| 278 |
+
| 175 | blk.19.attn_k.weight | 0xa1573720 | 0x240000 |
|
| 279 |
+
| 176 | blk.19.attn_norm.weight | 0xa17b3720 | 0x4000 |
|
| 280 |
+
| 177 | blk.19.attn_output.weight | 0xa17b7720 | 0x900000 |
|
| 281 |
+
| 178 | blk.19.attn_q.weight | 0xa20b7720 | 0x900000 |
|
| 282 |
+
| 179 | blk.19.attn_v.weight | 0xa29b7720 | 0x240000 |
|
| 283 |
+
| 180 | blk.19.ffn_down.weight | 0xa2bf7720 | 0x2680000 |
|
| 284 |
+
| 181 | blk.19.ffn_gate.weight | 0xa5277720 | 0x1f80000 |
|
| 285 |
+
| 182 | blk.19.ffn_norm.weight | 0xa71f7720 | 0x4000 |
|
| 286 |
+
| 183 | blk.19.ffn_up.weight | 0xa71fb720 | 0x1f80000 |
|
| 287 |
+
| 184 | blk.20.attn_k.weight | 0xa917b720 | 0x240000 |
|
| 288 |
+
| 185 | blk.20.attn_norm.weight | 0xa93bb720 | 0x4000 |
|
| 289 |
+
| 186 | blk.20.attn_output.weight | 0xa93bf720 | 0x900000 |
|
| 290 |
+
| 187 | blk.20.attn_q.weight | 0xa9cbf720 | 0x900000 |
|
| 291 |
+
| 188 | blk.20.attn_v.weight | 0xaa5bf720 | 0x240000 |
|
| 292 |
+
| 189 | blk.20.ffn_down.weight | 0xaa7ff720 | 0x2680000 |
|
| 293 |
+
| 190 | blk.20.ffn_gate.weight | 0xace7f720 | 0x1f80000 |
|
| 294 |
+
| 191 | blk.20.ffn_norm.weight | 0xaedff720 | 0x4000 |
|
| 295 |
+
| 192 | blk.20.ffn_up.weight | 0xaee03720 | 0x1f80000 |
|
| 296 |
+
| 193 | blk.21.attn_k.weight | 0xb0d83720 | 0x240000 |
|
| 297 |
+
| 194 | blk.21.attn_norm.weight | 0xb0fc3720 | 0x4000 |
|
| 298 |
+
| 195 | blk.21.attn_output.weight | 0xb0fc7720 | 0x900000 |
|
| 299 |
+
| 196 | blk.21.attn_q.weight | 0xb18c7720 | 0x900000 |
|
| 300 |
+
| 197 | blk.21.attn_v.weight | 0xb21c7720 | 0x240000 |
|
| 301 |
+
| 198 | blk.21.ffn_down.weight | 0xb2407720 | 0x2680000 |
|
| 302 |
+
| 199 | blk.21.ffn_gate.weight | 0xb4a87720 | 0x1f80000 |
|
| 303 |
+
| 200 | blk.21.ffn_norm.weight | 0xb6a07720 | 0x4000 |
|
| 304 |
+
| 201 | blk.21.ffn_up.weight | 0xb6a0b720 | 0x1f80000 |
|
| 305 |
+
| 202 | blk.22.attn_k.weight | 0xb898b720 | 0x240000 |
|
| 306 |
+
| 203 | blk.22.attn_norm.weight | 0xb8bcb720 | 0x4000 |
|
| 307 |
+
| 204 | blk.22.attn_output.weight | 0xb8bcf720 | 0x900000 |
|
| 308 |
+
| 205 | blk.22.attn_q.weight | 0xb94cf720 | 0x900000 |
|
| 309 |
+
| 206 | blk.22.attn_v.weight | 0xb9dcf720 | 0x240000 |
|
| 310 |
+
| 207 | blk.22.ffn_down.weight | 0xba00f720 | 0x2680000 |
|
| 311 |
+
| 208 | blk.22.ffn_gate.weight | 0xbc68f720 | 0x1f80000 |
|
| 312 |
+
| 209 | blk.22.ffn_norm.weight | 0xbe60f720 | 0x4000 |
|
| 313 |
+
| 210 | blk.22.ffn_up.weight | 0xbe613720 | 0x1f80000 |
|
| 314 |
+
| 211 | blk.23.attn_k.weight | 0xc0593720 | 0x240000 |
|
| 315 |
+
| 212 | blk.23.attn_norm.weight | 0xc07d3720 | 0x4000 |
|
| 316 |
+
| 213 | blk.23.attn_output.weight | 0xc07d7720 | 0x900000 |
|
| 317 |
+
| 214 | blk.23.attn_q.weight | 0xc10d7720 | 0x900000 |
|
| 318 |
+
| 215 | blk.23.attn_v.weight | 0xc19d7720 | 0x240000 |
|
| 319 |
+
| 216 | blk.23.ffn_down.weight | 0xc1c17720 | 0x2680000 |
|
| 320 |
+
| 217 | blk.23.ffn_gate.weight | 0xc4297720 | 0x1f80000 |
|
| 321 |
+
| 218 | blk.23.ffn_norm.weight | 0xc6217720 | 0x4000 |
|
| 322 |
+
| 219 | blk.23.ffn_up.weight | 0xc621b720 | 0x1f80000 |
|
| 323 |
+
| 220 | blk.24.attn_k.weight | 0xc819b720 | 0x240000 |
|
| 324 |
+
| 221 | blk.24.attn_norm.weight | 0xc83db720 | 0x4000 |
|
| 325 |
+
| 222 | blk.24.attn_output.weight | 0xc83df720 | 0x900000 |
|
| 326 |
+
| 223 | blk.24.attn_q.weight | 0xc8cdf720 | 0x900000 |
|
| 327 |
+
| 224 | blk.24.attn_v.weight | 0xc95df720 | 0x240000 |
|
| 328 |
+
| 225 | blk.24.ffn_down.weight | 0xc981f720 | 0x2680000 |
|
| 329 |
+
| 226 | blk.24.ffn_gate.weight | 0xcbe9f720 | 0x1f80000 |
|
| 330 |
+
| 227 | blk.24.ffn_norm.weight | 0xcde1f720 | 0x4000 |
|
| 331 |
+
| 228 | blk.24.ffn_up.weight | 0xcde23720 | 0x1f80000 |
|
| 332 |
+
| 229 | blk.25.attn_k.weight | 0xcfda3720 | 0x240000 |
|
| 333 |
+
| 230 | blk.25.attn_norm.weight | 0xcffe3720 | 0x4000 |
|
| 334 |
+
| 231 | blk.25.attn_output.weight | 0xcffe7720 | 0x900000 |
|
| 335 |
+
| 232 | blk.25.attn_q.weight | 0xd08e7720 | 0x900000 |
|
| 336 |
+
| 233 | blk.25.attn_v.weight | 0xd11e7720 | 0x240000 |
|
| 337 |
+
| 234 | blk.25.ffn_down.weight | 0xd1427720 | 0x2680000 |
|
| 338 |
+
| 235 | blk.25.ffn_gate.weight | 0xd3aa7720 | 0x1f80000 |
|
| 339 |
+
| 236 | blk.25.ffn_norm.weight | 0xd5a27720 | 0x4000 |
|
| 340 |
+
| 237 | blk.25.ffn_up.weight | 0xd5a2b720 | 0x1f80000 |
|
| 341 |
+
| 238 | blk.26.attn_k.weight | 0xd79ab720 | 0x240000 |
|
| 342 |
+
| 239 | blk.26.attn_norm.weight | 0xd7beb720 | 0x4000 |
|
| 343 |
+
| 240 | blk.26.attn_output.weight | 0xd7bef720 | 0x900000 |
|
| 344 |
+
| 241 | blk.26.attn_q.weight | 0xd84ef720 | 0x900000 |
|
| 345 |
+
| 242 | blk.26.attn_v.weight | 0xd8def720 | 0x240000 |
|
| 346 |
+
| 243 | blk.26.ffn_down.weight | 0xd902f720 | 0x2680000 |
|
| 347 |
+
| 244 | blk.26.ffn_gate.weight | 0xdb6af720 | 0x1f80000 |
|
| 348 |
+
| 245 | blk.26.ffn_norm.weight | 0xdd62f720 | 0x4000 |
|
| 349 |
+
| 246 | blk.26.ffn_up.weight | 0xdd633720 | 0x1f80000 |
|
| 350 |
+
| 247 | blk.27.attn_k.weight | 0xdf5b3720 | 0x240000 |
|
| 351 |
+
| 248 | blk.27.attn_norm.weight | 0xdf7f3720 | 0x4000 |
|
| 352 |
+
| 249 | blk.27.attn_output.weight | 0xdf7f7720 | 0x900000 |
|
| 353 |
+
| 250 | blk.27.attn_q.weight | 0xe00f7720 | 0x900000 |
|
| 354 |
+
| 251 | blk.27.attn_v.weight | 0xe09f7720 | 0x240000 |
|
| 355 |
+
| 252 | blk.27.ffn_down.weight | 0xe0c37720 | 0x2680000 |
|
| 356 |
+
| 253 | blk.27.ffn_gate.weight | 0xe32b7720 | 0x1f80000 |
|
| 357 |
+
| 254 | blk.27.ffn_norm.weight | 0xe5237720 | 0x4000 |
|
| 358 |
+
| 255 | blk.27.ffn_up.weight | 0xe523b720 | 0x1f80000 |
|
| 359 |
+
| 256 | blk.28.attn_k.weight | 0xe71bb720 | 0x240000 |
|
| 360 |
+
| 257 | blk.28.attn_norm.weight | 0xe73fb720 | 0x4000 |
|
| 361 |
+
| 258 | blk.28.attn_output.weight | 0xe73ff720 | 0x900000 |
|
| 362 |
+
| 259 | blk.28.attn_q.weight | 0xe7cff720 | 0x900000 |
|
| 363 |
+
| 260 | blk.28.attn_v.weight | 0xe85ff720 | 0x240000 |
|
| 364 |
+
| 261 | blk.28.ffn_down.weight | 0xe883f720 | 0x2680000 |
|
| 365 |
+
| 262 | blk.28.ffn_gate.weight | 0xeaebf720 | 0x1f80000 |
|
| 366 |
+
| 263 | blk.28.ffn_norm.weight | 0xece3f720 | 0x4000 |
|
| 367 |
+
| 264 | blk.28.ffn_up.weight | 0xece43720 | 0x1f80000 |
|
| 368 |
+
| 265 | blk.29.attn_k.weight | 0xeedc3720 | 0x240000 |
|
| 369 |
+
| 266 | blk.29.attn_norm.weight | 0xef003720 | 0x4000 |
|
| 370 |
+
| 267 | blk.29.attn_output.weight | 0xef007720 | 0x900000 |
|
| 371 |
+
| 268 | blk.29.attn_q.weight | 0xef907720 | 0x900000 |
|
| 372 |
+
| 269 | blk.29.attn_v.weight | 0xf0207720 | 0x240000 |
|
| 373 |
+
| 270 | blk.29.ffn_down.weight | 0xf0447720 | 0x2680000 |
|
| 374 |
+
| 271 | blk.29.ffn_gate.weight | 0xf2ac7720 | 0x1f80000 |
|
| 375 |
+
| 272 | blk.29.ffn_norm.weight | 0xf4a47720 | 0x4000 |
|
| 376 |
+
| 273 | blk.29.ffn_up.weight | 0xf4a4b720 | 0x1f80000 |
|
| 377 |
+
| 274 | blk.30.attn_k.weight | 0xf69cb720 | 0x240000 |
|
| 378 |
+
| 275 | blk.30.attn_norm.weight | 0xf6c0b720 | 0x4000 |
|
| 379 |
+
| 276 | blk.30.attn_output.weight | 0xf6c0f720 | 0x900000 |
|
| 380 |
+
| 277 | blk.30.attn_q.weight | 0xf750f720 | 0x900000 |
|
| 381 |
+
| 278 | blk.30.attn_v.weight | 0xf7e0f720 | 0x240000 |
|
| 382 |
+
| 279 | blk.30.ffn_down.weight | 0xf804f720 | 0x2680000 |
|
| 383 |
+
| 280 | blk.30.ffn_gate.weight | 0xfa6cf720 | 0x1f80000 |
|
| 384 |
+
| 281 | blk.30.ffn_norm.weight | 0xfc64f720 | 0x4000 |
|
| 385 |
+
| 282 | blk.30.ffn_up.weight | 0xfc653720 | 0x1f80000 |
|
| 386 |
+
| 283 | blk.31.attn_k.weight | 0xfe5d3720 | 0x1b8000 |
|
| 387 |
+
| 284 | blk.31.attn_norm.weight | 0xfe78b720 | 0x4000 |
|
| 388 |
+
| 285 | blk.31.attn_output.weight | 0xfe78f720 | 0x900000 |
|
| 389 |
+
| 286 | blk.31.attn_q.weight | 0xff08f720 | 0x6e0000 |
|
| 390 |
+
| 287 | blk.31.attn_v.weight | 0xff76f720 | 0x220000 |
|
| 391 |
+
| 288 | blk.31.ffn_down.weight | 0xff98f720 | 0x2680000 |
|
| 392 |
+
| 289 | blk.31.ffn_gate.weight | 0x10200f720 | 0x1f80000 |
|
| 393 |
+
| 290 | blk.31.ffn_norm.weight | 0x103f8f720 | 0x4000 |
|
| 394 |
+
| 291 | blk.31.ffn_up.weight | 0x103f93720 | 0x1f80000 |
|
| 395 |
+
|
| 396 |
+
### <a name="base">Base Tensor Group : ~1B Elements</a>
|
| 397 |
+
|
| 398 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 399 |
+
|-----:|:-------------------|:---------------------------------|:------------------|:----------------------|:-------|
|
| 400 |
+
| 0 | output.weight | Output (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | IQ4_NL |
|
| 401 |
+
| 1 | output_norm.weight | Output Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 402 |
+
| 2 | rope_freqs.weight | Rope_Freqs (W) | ( 64) 64 | 64 x 1 x 1 x 1 | F32 |
|
| 403 |
+
| 3 | token_embd.weight | Token Embedding (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | IQ3_S |
|
| 404 |
+
|
| 405 |
+
- Total elements in base: ( ~1B) 1050677312
|
| 406 |
+
- Percentage of total elements: 13.08%
|
| 407 |
+
|
| 408 |
+
|
| 409 |
+
### <a name="blk_0">Block 0 Tensor Group : ~218M Elements</a>
|
| 410 |
+
|
| 411 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 412 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-------|
|
| 413 |
+
| 4 | blk.0.attn_k.weight | Block 0 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 414 |
+
| 5 | blk.0.attn_norm.weight | Block 0 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 415 |
+
| 6 | blk.0.attn_output.weight | Block 0 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 416 |
+
| 7 | blk.0.attn_q.weight | Block 0 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 417 |
+
| 8 | blk.0.attn_v.weight | Block 0 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_XS |
|
| 418 |
+
| 9 | blk.0.ffn_down.weight | Block 0 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 419 |
+
| 10 | blk.0.ffn_gate.weight | Block 0 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 420 |
+
| 11 | blk.0.ffn_norm.weight | Block 0 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 421 |
+
| 12 | blk.0.ffn_up.weight | Block 0 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 422 |
+
|
| 423 |
+
- Total elements in blk.0: (~218M) 218112000
|
| 424 |
+
- Percentage of total elements: 2.72%
|
| 425 |
+
|
| 426 |
+
|
| 427 |
+
### <a name="blk_1">Block 1 Tensor Group : ~218M Elements</a>
|
| 428 |
+
|
| 429 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 430 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-------|
|
| 431 |
+
| 13 | blk.1.attn_k.weight | Block 1 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 432 |
+
| 14 | blk.1.attn_norm.weight | Block 1 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 433 |
+
| 15 | blk.1.attn_output.weight | Block 1 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 434 |
+
| 16 | blk.1.attn_q.weight | Block 1 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 435 |
+
| 17 | blk.1.attn_v.weight | Block 1 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_XS |
|
| 436 |
+
| 18 | blk.1.ffn_down.weight | Block 1 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 437 |
+
| 19 | blk.1.ffn_gate.weight | Block 1 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 438 |
+
| 20 | blk.1.ffn_norm.weight | Block 1 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 439 |
+
| 21 | blk.1.ffn_up.weight | Block 1 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 440 |
+
|
| 441 |
+
- Total elements in blk.1: (~218M) 218112000
|
| 442 |
+
- Percentage of total elements: 2.72%
|
| 443 |
+
|
| 444 |
+
|
| 445 |
+
### <a name="blk_2">Block 2 Tensor Group : ~218M Elements</a>
|
| 446 |
+
|
| 447 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 448 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-------|
|
| 449 |
+
| 22 | blk.2.attn_k.weight | Block 2 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 450 |
+
| 23 | blk.2.attn_norm.weight | Block 2 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 451 |
+
| 24 | blk.2.attn_output.weight | Block 2 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 452 |
+
| 25 | blk.2.attn_q.weight | Block 2 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 453 |
+
| 26 | blk.2.attn_v.weight | Block 2 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_XS |
|
| 454 |
+
| 27 | blk.2.ffn_down.weight | Block 2 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 455 |
+
| 28 | blk.2.ffn_gate.weight | Block 2 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 456 |
+
| 29 | blk.2.ffn_norm.weight | Block 2 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 457 |
+
| 30 | blk.2.ffn_up.weight | Block 2 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 458 |
+
|
| 459 |
+
- Total elements in blk.2: (~218M) 218112000
|
| 460 |
+
- Percentage of total elements: 2.72%
|
| 461 |
+
|
| 462 |
+
|
| 463 |
+
### <a name="blk_3">Block 3 Tensor Group : ~218M Elements</a>
|
| 464 |
+
|
| 465 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 466 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-------|
|
| 467 |
+
| 31 | blk.3.attn_k.weight | Block 3 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 468 |
+
| 32 | blk.3.attn_norm.weight | Block 3 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 469 |
+
| 33 | blk.3.attn_output.weight | Block 3 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 470 |
+
| 34 | blk.3.attn_q.weight | Block 3 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 471 |
+
| 35 | blk.3.attn_v.weight | Block 3 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_XS |
|
| 472 |
+
| 36 | blk.3.ffn_down.weight | Block 3 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 473 |
+
| 37 | blk.3.ffn_gate.weight | Block 3 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 474 |
+
| 38 | blk.3.ffn_norm.weight | Block 3 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 475 |
+
| 39 | blk.3.ffn_up.weight | Block 3 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 476 |
+
|
| 477 |
+
- Total elements in blk.3: (~218M) 218112000
|
| 478 |
+
- Percentage of total elements: 2.72%
|
| 479 |
+
|
| 480 |
+
|
| 481 |
+
### <a name="blk_4">Block 4 Tensor Group : ~218M Elements</a>
|
| 482 |
+
|
| 483 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 484 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-------|
|
| 485 |
+
| 40 | blk.4.attn_k.weight | Block 4 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 486 |
+
| 41 | blk.4.attn_norm.weight | Block 4 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 487 |
+
| 42 | blk.4.attn_output.weight | Block 4 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 488 |
+
| 43 | blk.4.attn_q.weight | Block 4 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 489 |
+
| 44 | blk.4.attn_v.weight | Block 4 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_XS |
|
| 490 |
+
| 45 | blk.4.ffn_down.weight | Block 4 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 491 |
+
| 46 | blk.4.ffn_gate.weight | Block 4 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 492 |
+
| 47 | blk.4.ffn_norm.weight | Block 4 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 493 |
+
| 48 | blk.4.ffn_up.weight | Block 4 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 494 |
+
|
| 495 |
+
- Total elements in blk.4: (~218M) 218112000
|
| 496 |
+
- Percentage of total elements: 2.72%
|
| 497 |
+
|
| 498 |
+
|
| 499 |
+
### <a name="blk_5">Block 5 Tensor Group : ~218M Elements</a>
|
| 500 |
+
|
| 501 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 502 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-------|
|
| 503 |
+
| 49 | blk.5.attn_k.weight | Block 5 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 504 |
+
| 50 | blk.5.attn_norm.weight | Block 5 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 505 |
+
| 51 | blk.5.attn_output.weight | Block 5 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 506 |
+
| 52 | blk.5.attn_q.weight | Block 5 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 507 |
+
| 53 | blk.5.attn_v.weight | Block 5 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_XS |
|
| 508 |
+
| 54 | blk.5.ffn_down.weight | Block 5 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 509 |
+
| 55 | blk.5.ffn_gate.weight | Block 5 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 510 |
+
| 56 | blk.5.ffn_norm.weight | Block 5 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 511 |
+
| 57 | blk.5.ffn_up.weight | Block 5 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 512 |
+
|
| 513 |
+
- Total elements in blk.5: (~218M) 218112000
|
| 514 |
+
- Percentage of total elements: 2.72%
|
| 515 |
+
|
| 516 |
+
|
| 517 |
+
### <a name="blk_6">Block 6 Tensor Group : ~218M Elements</a>
|
| 518 |
+
|
| 519 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 520 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-------|
|
| 521 |
+
| 58 | blk.6.attn_k.weight | Block 6 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 522 |
+
| 59 | blk.6.attn_norm.weight | Block 6 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 523 |
+
| 60 | blk.6.attn_output.weight | Block 6 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 524 |
+
| 61 | blk.6.attn_q.weight | Block 6 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 525 |
+
| 62 | blk.6.attn_v.weight | Block 6 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_XS |
|
| 526 |
+
| 63 | blk.6.ffn_down.weight | Block 6 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 527 |
+
| 64 | blk.6.ffn_gate.weight | Block 6 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 528 |
+
| 65 | blk.6.ffn_norm.weight | Block 6 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 529 |
+
| 66 | blk.6.ffn_up.weight | Block 6 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 530 |
+
|
| 531 |
+
- Total elements in blk.6: (~218M) 218112000
|
| 532 |
+
- Percentage of total elements: 2.72%
|
| 533 |
+
|
| 534 |
+
|
| 535 |
+
### <a name="blk_7">Block 7 Tensor Group : ~218M Elements</a>
|
| 536 |
+
|
| 537 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 538 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-------|
|
| 539 |
+
| 67 | blk.7.attn_k.weight | Block 7 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 540 |
+
| 68 | blk.7.attn_norm.weight | Block 7 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 541 |
+
| 69 | blk.7.attn_output.weight | Block 7 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 542 |
+
| 70 | blk.7.attn_q.weight | Block 7 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 543 |
+
| 71 | blk.7.attn_v.weight | Block 7 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_XS |
|
| 544 |
+
| 72 | blk.7.ffn_down.weight | Block 7 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 545 |
+
| 73 | blk.7.ffn_gate.weight | Block 7 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 546 |
+
| 74 | blk.7.ffn_norm.weight | Block 7 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 547 |
+
| 75 | blk.7.ffn_up.weight | Block 7 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 548 |
+
|
| 549 |
+
- Total elements in blk.7: (~218M) 218112000
|
| 550 |
+
- Percentage of total elements: 2.72%
|
| 551 |
+
|
| 552 |
+
|
| 553 |
+
### <a name="blk_8">Block 8 Tensor Group : ~218M Elements</a>
|
| 554 |
+
|
| 555 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 556 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-------|
|
| 557 |
+
| 76 | blk.8.attn_k.weight | Block 8 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 558 |
+
| 77 | blk.8.attn_norm.weight | Block 8 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 559 |
+
| 78 | blk.8.attn_output.weight | Block 8 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 560 |
+
| 79 | blk.8.attn_q.weight | Block 8 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 561 |
+
| 80 | blk.8.attn_v.weight | Block 8 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_XS |
|
| 562 |
+
| 81 | blk.8.ffn_down.weight | Block 8 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 563 |
+
| 82 | blk.8.ffn_gate.weight | Block 8 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 564 |
+
| 83 | blk.8.ffn_norm.weight | Block 8 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 565 |
+
| 84 | blk.8.ffn_up.weight | Block 8 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 566 |
+
|
| 567 |
+
- Total elements in blk.8: (~218M) 218112000
|
| 568 |
+
- Percentage of total elements: 2.72%
|
| 569 |
+
|
| 570 |
+
|
| 571 |
+
### <a name="blk_9">Block 9 Tensor Group : ~218M Elements</a>
|
| 572 |
+
|
| 573 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 574 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-------|
|
| 575 |
+
| 85 | blk.9.attn_k.weight | Block 9 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 576 |
+
| 86 | blk.9.attn_norm.weight | Block 9 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 577 |
+
| 87 | blk.9.attn_output.weight | Block 9 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 578 |
+
| 88 | blk.9.attn_q.weight | Block 9 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 579 |
+
| 89 | blk.9.attn_v.weight | Block 9 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_XS |
|
| 580 |
+
| 90 | blk.9.ffn_down.weight | Block 9 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 581 |
+
| 91 | blk.9.ffn_gate.weight | Block 9 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 582 |
+
| 92 | blk.9.ffn_norm.weight | Block 9 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 583 |
+
| 93 | blk.9.ffn_up.weight | Block 9 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 584 |
+
|
| 585 |
+
- Total elements in blk.9: (~218M) 218112000
|
| 586 |
+
- Percentage of total elements: 2.72%
|
| 587 |
+
|
| 588 |
+
|
| 589 |
+
### <a name="blk_10">Block 10 Tensor Group : ~218M Elements</a>
|
| 590 |
+
|
| 591 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 592 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 593 |
+
| 94 | blk.10.attn_k.weight | Block 10 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 594 |
+
| 95 | blk.10.attn_norm.weight | Block 10 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 595 |
+
| 96 | blk.10.attn_output.weight | Block 10 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 596 |
+
| 97 | blk.10.attn_q.weight | Block 10 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 597 |
+
| 98 | blk.10.attn_v.weight | Block 10 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_XS |
|
| 598 |
+
| 99 | blk.10.ffn_down.weight | Block 10 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 599 |
+
| 100 | blk.10.ffn_gate.weight | Block 10 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 600 |
+
| 101 | blk.10.ffn_norm.weight | Block 10 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 601 |
+
| 102 | blk.10.ffn_up.weight | Block 10 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 602 |
+
|
| 603 |
+
- Total elements in blk.10: (~218M) 218112000
|
| 604 |
+
- Percentage of total elements: 2.72%
|
| 605 |
+
|
| 606 |
+
|
| 607 |
+
### <a name="blk_11">Block 11 Tensor Group : ~218M Elements</a>
|
| 608 |
+
|
| 609 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 610 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 611 |
+
| 103 | blk.11.attn_k.weight | Block 11 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 612 |
+
| 104 | blk.11.attn_norm.weight | Block 11 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 613 |
+
| 105 | blk.11.attn_output.weight | Block 11 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 614 |
+
| 106 | blk.11.attn_q.weight | Block 11 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 615 |
+
| 107 | blk.11.attn_v.weight | Block 11 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_XS |
|
| 616 |
+
| 108 | blk.11.ffn_down.weight | Block 11 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 617 |
+
| 109 | blk.11.ffn_gate.weight | Block 11 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 618 |
+
| 110 | blk.11.ffn_norm.weight | Block 11 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 619 |
+
| 111 | blk.11.ffn_up.weight | Block 11 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 620 |
+
|
| 621 |
+
- Total elements in blk.11: (~218M) 218112000
|
| 622 |
+
- Percentage of total elements: 2.72%
|
| 623 |
+
|
| 624 |
+
|
| 625 |
+
### <a name="blk_12">Block 12 Tensor Group : ~218M Elements</a>
|
| 626 |
+
|
| 627 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 628 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 629 |
+
| 112 | blk.12.attn_k.weight | Block 12 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 630 |
+
| 113 | blk.12.attn_norm.weight | Block 12 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 631 |
+
| 114 | blk.12.attn_output.weight | Block 12 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 632 |
+
| 115 | blk.12.attn_q.weight | Block 12 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 633 |
+
| 116 | blk.12.attn_v.weight | Block 12 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_XS |
|
| 634 |
+
| 117 | blk.12.ffn_down.weight | Block 12 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 635 |
+
| 118 | blk.12.ffn_gate.weight | Block 12 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 636 |
+
| 119 | blk.12.ffn_norm.weight | Block 12 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 637 |
+
| 120 | blk.12.ffn_up.weight | Block 12 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 638 |
+
|
| 639 |
+
- Total elements in blk.12: (~218M) 218112000
|
| 640 |
+
- Percentage of total elements: 2.72%
|
| 641 |
+
|
| 642 |
+
|
| 643 |
+
### <a name="blk_13">Block 13 Tensor Group : ~218M Elements</a>
|
| 644 |
+
|
| 645 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 646 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 647 |
+
| 121 | blk.13.attn_k.weight | Block 13 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 648 |
+
| 122 | blk.13.attn_norm.weight | Block 13 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 649 |
+
| 123 | blk.13.attn_output.weight | Block 13 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 650 |
+
| 124 | blk.13.attn_q.weight | Block 13 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 651 |
+
| 125 | blk.13.attn_v.weight | Block 13 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 652 |
+
| 126 | blk.13.ffn_down.weight | Block 13 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 653 |
+
| 127 | blk.13.ffn_gate.weight | Block 13 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 654 |
+
| 128 | blk.13.ffn_norm.weight | Block 13 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 655 |
+
| 129 | blk.13.ffn_up.weight | Block 13 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 656 |
+
|
| 657 |
+
- Total elements in blk.13: (~218M) 218112000
|
| 658 |
+
- Percentage of total elements: 2.72%
|
| 659 |
+
|
| 660 |
+
|
| 661 |
+
### <a name="blk_14">Block 14 Tensor Group : ~218M Elements</a>
|
| 662 |
+
|
| 663 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 664 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 665 |
+
| 130 | blk.14.attn_k.weight | Block 14 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 666 |
+
| 131 | blk.14.attn_norm.weight | Block 14 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 667 |
+
| 132 | blk.14.attn_output.weight | Block 14 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 668 |
+
| 133 | blk.14.attn_q.weight | Block 14 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 669 |
+
| 134 | blk.14.attn_v.weight | Block 14 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 670 |
+
| 135 | blk.14.ffn_down.weight | Block 14 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 671 |
+
| 136 | blk.14.ffn_gate.weight | Block 14 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 672 |
+
| 137 | blk.14.ffn_norm.weight | Block 14 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 673 |
+
| 138 | blk.14.ffn_up.weight | Block 14 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 674 |
+
|
| 675 |
+
- Total elements in blk.14: (~218M) 218112000
|
| 676 |
+
- Percentage of total elements: 2.72%
|
| 677 |
+
|
| 678 |
+
|
| 679 |
+
### <a name="blk_15">Block 15 Tensor Group : ~218M Elements</a>
|
| 680 |
+
|
| 681 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 682 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 683 |
+
| 139 | blk.15.attn_k.weight | Block 15 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 684 |
+
| 140 | blk.15.attn_norm.weight | Block 15 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 685 |
+
| 141 | blk.15.attn_output.weight | Block 15 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 686 |
+
| 142 | blk.15.attn_q.weight | Block 15 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 687 |
+
| 143 | blk.15.attn_v.weight | Block 15 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_XS |
|
| 688 |
+
| 144 | blk.15.ffn_down.weight | Block 15 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 689 |
+
| 145 | blk.15.ffn_gate.weight | Block 15 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 690 |
+
| 146 | blk.15.ffn_norm.weight | Block 15 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 691 |
+
| 147 | blk.15.ffn_up.weight | Block 15 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ3_S |
|
| 692 |
+
|
| 693 |
+
- Total elements in blk.15: (~218M) 218112000
|
| 694 |
+
- Percentage of total elements: 2.72%
|
| 695 |
+
|
| 696 |
+
|
| 697 |
+
### <a name="blk_16">Block 16 Tensor Group : ~218M Elements</a>
|
| 698 |
+
|
| 699 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 700 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 701 |
+
| 148 | blk.16.attn_k.weight | Block 16 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 702 |
+
| 149 | blk.16.attn_norm.weight | Block 16 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 703 |
+
| 150 | blk.16.attn_output.weight | Block 16 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 704 |
+
| 151 | blk.16.attn_q.weight | Block 16 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 705 |
+
| 152 | blk.16.attn_v.weight | Block 16 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 706 |
+
| 153 | blk.16.ffn_down.weight | Block 16 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 707 |
+
| 154 | blk.16.ffn_gate.weight | Block 16 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 708 |
+
| 155 | blk.16.ffn_norm.weight | Block 16 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 709 |
+
| 156 | blk.16.ffn_up.weight | Block 16 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 710 |
+
|
| 711 |
+
- Total elements in blk.16: (~218M) 218112000
|
| 712 |
+
- Percentage of total elements: 2.72%
|
| 713 |
+
|
| 714 |
+
|
| 715 |
+
### <a name="blk_17">Block 17 Tensor Group : ~218M Elements</a>
|
| 716 |
+
|
| 717 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 718 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 719 |
+
| 157 | blk.17.attn_k.weight | Block 17 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 720 |
+
| 158 | blk.17.attn_norm.weight | Block 17 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 721 |
+
| 159 | blk.17.attn_output.weight | Block 17 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 722 |
+
| 160 | blk.17.attn_q.weight | Block 17 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 723 |
+
| 161 | blk.17.attn_v.weight | Block 17 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_XS |
|
| 724 |
+
| 162 | blk.17.ffn_down.weight | Block 17 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 725 |
+
| 163 | blk.17.ffn_gate.weight | Block 17 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 726 |
+
| 164 | blk.17.ffn_norm.weight | Block 17 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 727 |
+
| 165 | blk.17.ffn_up.weight | Block 17 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 728 |
+
|
| 729 |
+
- Total elements in blk.17: (~218M) 218112000
|
| 730 |
+
- Percentage of total elements: 2.72%
|
| 731 |
+
|
| 732 |
+
|
| 733 |
+
### <a name="blk_18">Block 18 Tensor Group : ~218M Elements</a>
|
| 734 |
+
|
| 735 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 736 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 737 |
+
| 166 | blk.18.attn_k.weight | Block 18 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 738 |
+
| 167 | blk.18.attn_norm.weight | Block 18 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 739 |
+
| 168 | blk.18.attn_output.weight | Block 18 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 740 |
+
| 169 | blk.18.attn_q.weight | Block 18 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 741 |
+
| 170 | blk.18.attn_v.weight | Block 18 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 742 |
+
| 171 | blk.18.ffn_down.weight | Block 18 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 743 |
+
| 172 | blk.18.ffn_gate.weight | Block 18 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 744 |
+
| 173 | blk.18.ffn_norm.weight | Block 18 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 745 |
+
| 174 | blk.18.ffn_up.weight | Block 18 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 746 |
+
|
| 747 |
+
- Total elements in blk.18: (~218M) 218112000
|
| 748 |
+
- Percentage of total elements: 2.72%
|
| 749 |
+
|
| 750 |
+
|
| 751 |
+
### <a name="blk_19">Block 19 Tensor Group : ~218M Elements</a>
|
| 752 |
+
|
| 753 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 754 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 755 |
+
| 175 | blk.19.attn_k.weight | Block 19 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 756 |
+
| 176 | blk.19.attn_norm.weight | Block 19 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 757 |
+
| 177 | blk.19.attn_output.weight | Block 19 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 758 |
+
| 178 | blk.19.attn_q.weight | Block 19 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 759 |
+
| 179 | blk.19.attn_v.weight | Block 19 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 760 |
+
| 180 | blk.19.ffn_down.weight | Block 19 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 761 |
+
| 181 | blk.19.ffn_gate.weight | Block 19 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 762 |
+
| 182 | blk.19.ffn_norm.weight | Block 19 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 763 |
+
| 183 | blk.19.ffn_up.weight | Block 19 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 764 |
+
|
| 765 |
+
- Total elements in blk.19: (~218M) 218112000
|
| 766 |
+
- Percentage of total elements: 2.72%
|
| 767 |
+
|
| 768 |
+
|
| 769 |
+
### <a name="blk_20">Block 20 Tensor Group : ~218M Elements</a>
|
| 770 |
+
|
| 771 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 772 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 773 |
+
| 184 | blk.20.attn_k.weight | Block 20 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 774 |
+
| 185 | blk.20.attn_norm.weight | Block 20 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 775 |
+
| 186 | blk.20.attn_output.weight | Block 20 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 776 |
+
| 187 | blk.20.attn_q.weight | Block 20 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 777 |
+
| 188 | blk.20.attn_v.weight | Block 20 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 778 |
+
| 189 | blk.20.ffn_down.weight | Block 20 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 779 |
+
| 190 | blk.20.ffn_gate.weight | Block 20 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 780 |
+
| 191 | blk.20.ffn_norm.weight | Block 20 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 781 |
+
| 192 | blk.20.ffn_up.weight | Block 20 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 782 |
+
|
| 783 |
+
- Total elements in blk.20: (~218M) 218112000
|
| 784 |
+
- Percentage of total elements: 2.72%
|
| 785 |
+
|
| 786 |
+
|
| 787 |
+
### <a name="blk_21">Block 21 Tensor Group : ~218M Elements</a>
|
| 788 |
+
|
| 789 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 790 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 791 |
+
| 193 | blk.21.attn_k.weight | Block 21 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 792 |
+
| 194 | blk.21.attn_norm.weight | Block 21 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 793 |
+
| 195 | blk.21.attn_output.weight | Block 21 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 794 |
+
| 196 | blk.21.attn_q.weight | Block 21 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 795 |
+
| 197 | blk.21.attn_v.weight | Block 21 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 796 |
+
| 198 | blk.21.ffn_down.weight | Block 21 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 797 |
+
| 199 | blk.21.ffn_gate.weight | Block 21 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 798 |
+
| 200 | blk.21.ffn_norm.weight | Block 21 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 799 |
+
| 201 | blk.21.ffn_up.weight | Block 21 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 800 |
+
|
| 801 |
+
- Total elements in blk.21: (~218M) 218112000
|
| 802 |
+
- Percentage of total elements: 2.72%
|
| 803 |
+
|
| 804 |
+
|
| 805 |
+
### <a name="blk_22">Block 22 Tensor Group : ~218M Elements</a>
|
| 806 |
+
|
| 807 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 808 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 809 |
+
| 202 | blk.22.attn_k.weight | Block 22 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 810 |
+
| 203 | blk.22.attn_norm.weight | Block 22 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 811 |
+
| 204 | blk.22.attn_output.weight | Block 22 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 812 |
+
| 205 | blk.22.attn_q.weight | Block 22 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 813 |
+
| 206 | blk.22.attn_v.weight | Block 22 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 814 |
+
| 207 | blk.22.ffn_down.weight | Block 22 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 815 |
+
| 208 | blk.22.ffn_gate.weight | Block 22 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 816 |
+
| 209 | blk.22.ffn_norm.weight | Block 22 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 817 |
+
| 210 | blk.22.ffn_up.weight | Block 22 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 818 |
+
|
| 819 |
+
- Total elements in blk.22: (~218M) 218112000
|
| 820 |
+
- Percentage of total elements: 2.72%
|
| 821 |
+
|
| 822 |
+
|
| 823 |
+
### <a name="blk_23">Block 23 Tensor Group : ~218M Elements</a>
|
| 824 |
+
|
| 825 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 826 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 827 |
+
| 211 | blk.23.attn_k.weight | Block 23 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 828 |
+
| 212 | blk.23.attn_norm.weight | Block 23 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 829 |
+
| 213 | blk.23.attn_output.weight | Block 23 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 830 |
+
| 214 | blk.23.attn_q.weight | Block 23 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 831 |
+
| 215 | blk.23.attn_v.weight | Block 23 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 832 |
+
| 216 | blk.23.ffn_down.weight | Block 23 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 833 |
+
| 217 | blk.23.ffn_gate.weight | Block 23 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 834 |
+
| 218 | blk.23.ffn_norm.weight | Block 23 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 835 |
+
| 219 | blk.23.ffn_up.weight | Block 23 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 836 |
+
|
| 837 |
+
- Total elements in blk.23: (~218M) 218112000
|
| 838 |
+
- Percentage of total elements: 2.72%
|
| 839 |
+
|
| 840 |
+
|
| 841 |
+
### <a name="blk_24">Block 24 Tensor Group : ~218M Elements</a>
|
| 842 |
+
|
| 843 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 844 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 845 |
+
| 220 | blk.24.attn_k.weight | Block 24 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 846 |
+
| 221 | blk.24.attn_norm.weight | Block 24 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 847 |
+
| 222 | blk.24.attn_output.weight | Block 24 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 848 |
+
| 223 | blk.24.attn_q.weight | Block 24 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 849 |
+
| 224 | blk.24.attn_v.weight | Block 24 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 850 |
+
| 225 | blk.24.ffn_down.weight | Block 24 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 851 |
+
| 226 | blk.24.ffn_gate.weight | Block 24 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 852 |
+
| 227 | blk.24.ffn_norm.weight | Block 24 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 853 |
+
| 228 | blk.24.ffn_up.weight | Block 24 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 854 |
+
|
| 855 |
+
- Total elements in blk.24: (~218M) 218112000
|
| 856 |
+
- Percentage of total elements: 2.72%
|
| 857 |
+
|
| 858 |
+
|
| 859 |
+
### <a name="blk_25">Block 25 Tensor Group : ~218M Elements</a>
|
| 860 |
+
|
| 861 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 862 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 863 |
+
| 229 | blk.25.attn_k.weight | Block 25 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 864 |
+
| 230 | blk.25.attn_norm.weight | Block 25 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 865 |
+
| 231 | blk.25.attn_output.weight | Block 25 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 866 |
+
| 232 | blk.25.attn_q.weight | Block 25 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 867 |
+
| 233 | blk.25.attn_v.weight | Block 25 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 868 |
+
| 234 | blk.25.ffn_down.weight | Block 25 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 869 |
+
| 235 | blk.25.ffn_gate.weight | Block 25 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 870 |
+
| 236 | blk.25.ffn_norm.weight | Block 25 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 871 |
+
| 237 | blk.25.ffn_up.weight | Block 25 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 872 |
+
|
| 873 |
+
- Total elements in blk.25: (~218M) 218112000
|
| 874 |
+
- Percentage of total elements: 2.72%
|
| 875 |
+
|
| 876 |
+
|
| 877 |
+
### <a name="blk_26">Block 26 Tensor Group : ~218M Elements</a>
|
| 878 |
+
|
| 879 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 880 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 881 |
+
| 238 | blk.26.attn_k.weight | Block 26 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 882 |
+
| 239 | blk.26.attn_norm.weight | Block 26 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 883 |
+
| 240 | blk.26.attn_output.weight | Block 26 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 884 |
+
| 241 | blk.26.attn_q.weight | Block 26 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 885 |
+
| 242 | blk.26.attn_v.weight | Block 26 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 886 |
+
| 243 | blk.26.ffn_down.weight | Block 26 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 887 |
+
| 244 | blk.26.ffn_gate.weight | Block 26 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 888 |
+
| 245 | blk.26.ffn_norm.weight | Block 26 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 889 |
+
| 246 | blk.26.ffn_up.weight | Block 26 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 890 |
+
|
| 891 |
+
- Total elements in blk.26: (~218M) 218112000
|
| 892 |
+
- Percentage of total elements: 2.72%
|
| 893 |
+
|
| 894 |
+
|
| 895 |
+
### <a name="blk_27">Block 27 Tensor Group : ~218M Elements</a>
|
| 896 |
+
|
| 897 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 898 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 899 |
+
| 247 | blk.27.attn_k.weight | Block 27 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 900 |
+
| 248 | blk.27.attn_norm.weight | Block 27 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 901 |
+
| 249 | blk.27.attn_output.weight | Block 27 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 902 |
+
| 250 | blk.27.attn_q.weight | Block 27 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 903 |
+
| 251 | blk.27.attn_v.weight | Block 27 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 904 |
+
| 252 | blk.27.ffn_down.weight | Block 27 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 905 |
+
| 253 | blk.27.ffn_gate.weight | Block 27 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 906 |
+
| 254 | blk.27.ffn_norm.weight | Block 27 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 907 |
+
| 255 | blk.27.ffn_up.weight | Block 27 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 908 |
+
|
| 909 |
+
- Total elements in blk.27: (~218M) 218112000
|
| 910 |
+
- Percentage of total elements: 2.72%
|
| 911 |
+
|
| 912 |
+
|
| 913 |
+
### <a name="blk_28">Block 28 Tensor Group : ~218M Elements</a>
|
| 914 |
+
|
| 915 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 916 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 917 |
+
| 256 | blk.28.attn_k.weight | Block 28 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 918 |
+
| 257 | blk.28.attn_norm.weight | Block 28 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 919 |
+
| 258 | blk.28.attn_output.weight | Block 28 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 920 |
+
| 259 | blk.28.attn_q.weight | Block 28 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 921 |
+
| 260 | blk.28.attn_v.weight | Block 28 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 922 |
+
| 261 | blk.28.ffn_down.weight | Block 28 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 923 |
+
| 262 | blk.28.ffn_gate.weight | Block 28 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 924 |
+
| 263 | blk.28.ffn_norm.weight | Block 28 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 925 |
+
| 264 | blk.28.ffn_up.weight | Block 28 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 926 |
+
|
| 927 |
+
- Total elements in blk.28: (~218M) 218112000
|
| 928 |
+
- Percentage of total elements: 2.72%
|
| 929 |
+
|
| 930 |
+
|
| 931 |
+
### <a name="blk_29">Block 29 Tensor Group : ~218M Elements</a>
|
| 932 |
+
|
| 933 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 934 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 935 |
+
| 265 | blk.29.attn_k.weight | Block 29 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 936 |
+
| 266 | blk.29.attn_norm.weight | Block 29 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 937 |
+
| 267 | blk.29.attn_output.weight | Block 29 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 938 |
+
| 268 | blk.29.attn_q.weight | Block 29 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 939 |
+
| 269 | blk.29.attn_v.weight | Block 29 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 940 |
+
| 270 | blk.29.ffn_down.weight | Block 29 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 941 |
+
| 271 | blk.29.ffn_gate.weight | Block 29 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 942 |
+
| 272 | blk.29.ffn_norm.weight | Block 29 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 943 |
+
| 273 | blk.29.ffn_up.weight | Block 29 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 944 |
+
|
| 945 |
+
- Total elements in blk.29: (~218M) 218112000
|
| 946 |
+
- Percentage of total elements: 2.72%
|
| 947 |
+
|
| 948 |
+
|
| 949 |
+
### <a name="blk_30">Block 30 Tensor Group : ~218M Elements</a>
|
| 950 |
+
|
| 951 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 952 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 953 |
+
| 274 | blk.30.attn_k.weight | Block 30 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 954 |
+
| 275 | blk.30.attn_norm.weight | Block 30 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 955 |
+
| 276 | blk.30.attn_output.weight | Block 30 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 956 |
+
| 277 | blk.30.attn_q.weight | Block 30 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 957 |
+
| 278 | blk.30.attn_v.weight | Block 30 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_NL |
|
| 958 |
+
| 279 | blk.30.ffn_down.weight | Block 30 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 959 |
+
| 280 | blk.30.ffn_gate.weight | Block 30 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 960 |
+
| 281 | blk.30.ffn_norm.weight | Block 30 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 961 |
+
| 282 | blk.30.ffn_up.weight | Block 30 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 962 |
+
|
| 963 |
+
- Total elements in blk.30: (~218M) 218112000
|
| 964 |
+
- Percentage of total elements: 2.72%
|
| 965 |
+
|
| 966 |
+
|
| 967 |
+
### <a name="blk_31">Block 31 Tensor Group : ~218M Elements</a>
|
| 968 |
+
|
| 969 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 970 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-------|
|
| 971 |
+
| 283 | blk.31.attn_k.weight | Block 31 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ3_S |
|
| 972 |
+
| 284 | blk.31.attn_norm.weight | Block 31 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 973 |
+
| 285 | blk.31.attn_output.weight | Block 31 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ4_NL |
|
| 974 |
+
| 286 | blk.31.attn_q.weight | Block 31 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | IQ3_S |
|
| 975 |
+
| 287 | blk.31.attn_v.weight | Block 31 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | IQ4_XS |
|
| 976 |
+
| 288 | blk.31.ffn_down.weight | Block 31 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 977 |
+
| 289 | blk.31.ffn_gate.weight | Block 31 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 978 |
+
| 290 | blk.31.ffn_norm.weight | Block 31 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 979 |
+
| 291 | blk.31.ffn_up.weight | Block 31 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | IQ4_NL |
|
| 980 |
+
|
| 981 |
+
- Total elements in blk.31: (~218M) 218112000
|
| 982 |
+
- Percentage of total elements: 2.72%
|
scores/Watt-Tool-8B-Q3_K_L.md
ADDED
|
@@ -0,0 +1,982 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Watt-Tool-8B-Q3_K_L.gguf - GGUF Internal File Dump
|
| 2 |
+
|
| 3 |
+
- Endian: LITTLE endian
|
| 4 |
+
|
| 5 |
+
## Key Value Metadata Store
|
| 6 |
+
|
| 7 |
+
There are 43 key-value pairs in this file
|
| 8 |
+
|
| 9 |
+
| POS | TYPE | Count | Key | Value |
|
| 10 |
+
|----:|:---------|-------:|:---------------------------------------|:--------------------------------------------------------------------|
|
| 11 |
+
| 1 | UINT32 | 1 | GGUF.version | 3 |
|
| 12 |
+
| 2 | UINT64 | 1 | GGUF.tensor_count | 292 |
|
| 13 |
+
| 3 | UINT64 | 1 | GGUF.kv_count | 40 |
|
| 14 |
+
| 4 | STRING | 1 | general.architecture | `llama` |
|
| 15 |
+
| 5 | STRING | 1 | general.type | `model` |
|
| 16 |
+
| 6 | STRING | 1 | general.name | `Watt Tool 8B GGUF` |
|
| 17 |
+
| 7 | STRING | 1 | general.finetune | `GGUF` |
|
| 18 |
+
| 8 | STRING | 1 | general.basename | `Watt-Tool` |
|
| 19 |
+
| 9 | STRING | 1 | general.size_label | `8B` |
|
| 20 |
+
| 10 | STRING | 1 | general.license | `apache-2.0` |
|
| 21 |
+
| 11 | UINT32 | 1 | general.base_model.count | 1 |
|
| 22 |
+
| 12 | STRING | 1 | general.base_model.0.name | `Llama 3.1 8B Instruct` |
|
| 23 |
+
| 13 | STRING | 1 | general.base_model.0.organization | `Meta Llama` |
|
| 24 |
+
| 14 | STRING | 1 | general.base_model.0.repo_url | `https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct` |
|
| 25 |
+
| 15 | [STRING] | 4 | general.tags | [ `function-calling`, `tool-use`, `llama`, `bfcl` ] |
|
| 26 |
+
| 16 | [STRING] | 1 | general.languages | [ `en` ] |
|
| 27 |
+
| 17 | UINT32 | 1 | llama.block_count | 32 |
|
| 28 |
+
| 18 | UINT32 | 1 | llama.context_length | 131072 |
|
| 29 |
+
| 19 | UINT32 | 1 | llama.embedding_length | 4096 |
|
| 30 |
+
| 20 | UINT32 | 1 | llama.feed_forward_length | 14336 |
|
| 31 |
+
| 21 | UINT32 | 1 | llama.attention.head_count | 32 |
|
| 32 |
+
| 22 | UINT32 | 1 | llama.attention.head_count_kv | 8 |
|
| 33 |
+
| 23 | FLOAT32 | 1 | llama.rope.freq_base | 500000.0 |
|
| 34 |
+
| 24 | FLOAT32 | 1 | llama.attention.layer_norm_rms_epsilon | 1e-05 |
|
| 35 |
+
| 25 | UINT32 | 1 | llama.attention.key_length | 128 |
|
| 36 |
+
| 26 | UINT32 | 1 | llama.attention.value_length | 128 |
|
| 37 |
+
| 27 | UINT32 | 1 | llama.vocab_size | 128256 |
|
| 38 |
+
| 28 | UINT32 | 1 | llama.rope.dimension_count | 128 |
|
| 39 |
+
| 29 | STRING | 1 | tokenizer.ggml.model | `gpt2` |
|
| 40 |
+
| 30 | STRING | 1 | tokenizer.ggml.pre | `llama-bpe` |
|
| 41 |
+
| 31 | [STRING] | 128256 | tokenizer.ggml.tokens | [ `!`, `"`, `#`, `$`, `%`, ... ] |
|
| 42 |
+
| 32 | [INT32] | 128256 | tokenizer.ggml.token_type | [ 1, 1, 1, 1, 1, 1, 1, ... ] |
|
| 43 |
+
| 33 | [STRING] | 280147 | tokenizer.ggml.merges | [ `Ġ Ġ`, `Ġ ĠĠĠ`, `ĠĠ ĠĠ`, `ĠĠĠ Ġ`, `i n`, ... ] |
|
| 44 |
+
| 34 | UINT32 | 1 | tokenizer.ggml.bos_token_id | 128000 |
|
| 45 |
+
| 35 | UINT32 | 1 | tokenizer.ggml.eos_token_id | 128009 |
|
| 46 |
+
| 36 | UINT32 | 1 | tokenizer.ggml.padding_token_id | 128009 |
|
| 47 |
+
| 37 | STRING | 1 | tokenizer.chat_template | `{{ '<|begin_of_text|>' }}{% if`...`d|>' }}{% endif %}{% endfor %}` |
|
| 48 |
+
| 38 | UINT32 | 1 | general.quantization_version | 2 |
|
| 49 |
+
| 39 | UINT32 | 1 | general.file_type | 13 |
|
| 50 |
+
| 40 | STRING | 1 | quantize.imatrix.file | `./imatrix/imatrix-Watt-Tool-8B-small.dat` |
|
| 51 |
+
| 41 | STRING | 1 | quantize.imatrix.dataset | `../../datasets/imatrix/calibration_eur_small.txt` |
|
| 52 |
+
| 42 | INT32 | 1 | quantize.imatrix.entries_count | 225 |
|
| 53 |
+
| 43 | INT32 | 1 | quantize.imatrix.chunks_count | 962 |
|
| 54 |
+
|
| 55 |
+
## Tensors Overview ~8B Elements
|
| 56 |
+
|
| 57 |
+
Total number of elements in all tensors: 8030261312 Elements
|
| 58 |
+
|
| 59 |
+
- [Watt-Tool-8B-Q3\_K\_L.gguf - GGUF Internal File Dump](#watt-tool-8b-q3_k_lgguf---gguf-internal-file-dump)
|
| 60 |
+
- [Key Value Metadata Store](#key-value-metadata-store)
|
| 61 |
+
- [Tensors Overview ~8B Elements](#tensors-overview-8b-elements)
|
| 62 |
+
- [Tensor Data Offset](#tensor-data-offset)
|
| 63 |
+
- [Base Tensor Group : ~1B Elements](#base-tensor-group--1b-elements)
|
| 64 |
+
- [Block 0 Tensor Group : ~218M Elements](#block-0-tensor-group--218m-elements)
|
| 65 |
+
- [Block 1 Tensor Group : ~218M Elements](#block-1-tensor-group--218m-elements)
|
| 66 |
+
- [Block 2 Tensor Group : ~218M Elements](#block-2-tensor-group--218m-elements)
|
| 67 |
+
- [Block 3 Tensor Group : ~218M Elements](#block-3-tensor-group--218m-elements)
|
| 68 |
+
- [Block 4 Tensor Group : ~218M Elements](#block-4-tensor-group--218m-elements)
|
| 69 |
+
- [Block 5 Tensor Group : ~218M Elements](#block-5-tensor-group--218m-elements)
|
| 70 |
+
- [Block 6 Tensor Group : ~218M Elements](#block-6-tensor-group--218m-elements)
|
| 71 |
+
- [Block 7 Tensor Group : ~218M Elements](#block-7-tensor-group--218m-elements)
|
| 72 |
+
- [Block 8 Tensor Group : ~218M Elements](#block-8-tensor-group--218m-elements)
|
| 73 |
+
- [Block 9 Tensor Group : ~218M Elements](#block-9-tensor-group--218m-elements)
|
| 74 |
+
- [Block 10 Tensor Group : ~218M Elements](#block-10-tensor-group--218m-elements)
|
| 75 |
+
- [Block 11 Tensor Group : ~218M Elements](#block-11-tensor-group--218m-elements)
|
| 76 |
+
- [Block 12 Tensor Group : ~218M Elements](#block-12-tensor-group--218m-elements)
|
| 77 |
+
- [Block 13 Tensor Group : ~218M Elements](#block-13-tensor-group--218m-elements)
|
| 78 |
+
- [Block 14 Tensor Group : ~218M Elements](#block-14-tensor-group--218m-elements)
|
| 79 |
+
- [Block 15 Tensor Group : ~218M Elements](#block-15-tensor-group--218m-elements)
|
| 80 |
+
- [Block 16 Tensor Group : ~218M Elements](#block-16-tensor-group--218m-elements)
|
| 81 |
+
- [Block 17 Tensor Group : ~218M Elements](#block-17-tensor-group--218m-elements)
|
| 82 |
+
- [Block 18 Tensor Group : ~218M Elements](#block-18-tensor-group--218m-elements)
|
| 83 |
+
- [Block 19 Tensor Group : ~218M Elements](#block-19-tensor-group--218m-elements)
|
| 84 |
+
- [Block 20 Tensor Group : ~218M Elements](#block-20-tensor-group--218m-elements)
|
| 85 |
+
- [Block 21 Tensor Group : ~218M Elements](#block-21-tensor-group--218m-elements)
|
| 86 |
+
- [Block 22 Tensor Group : ~218M Elements](#block-22-tensor-group--218m-elements)
|
| 87 |
+
- [Block 23 Tensor Group : ~218M Elements](#block-23-tensor-group--218m-elements)
|
| 88 |
+
- [Block 24 Tensor Group : ~218M Elements](#block-24-tensor-group--218m-elements)
|
| 89 |
+
- [Block 25 Tensor Group : ~218M Elements](#block-25-tensor-group--218m-elements)
|
| 90 |
+
- [Block 26 Tensor Group : ~218M Elements](#block-26-tensor-group--218m-elements)
|
| 91 |
+
- [Block 27 Tensor Group : ~218M Elements](#block-27-tensor-group--218m-elements)
|
| 92 |
+
- [Block 28 Tensor Group : ~218M Elements](#block-28-tensor-group--218m-elements)
|
| 93 |
+
- [Block 29 Tensor Group : ~218M Elements](#block-29-tensor-group--218m-elements)
|
| 94 |
+
- [Block 30 Tensor Group : ~218M Elements](#block-30-tensor-group--218m-elements)
|
| 95 |
+
- [Block 31 Tensor Group : ~218M Elements](#block-31-tensor-group--218m-elements)
|
| 96 |
+
|
| 97 |
+
### Tensor Data Offset
|
| 98 |
+
|
| 99 |
+
This table contains the offset and data segment relative to start of file
|
| 100 |
+
|
| 101 |
+
| T_ID | Tensor Layer Name | Data Offset (B) | Data Size (B) |
|
| 102 |
+
|-----:|:--------------------------|-----------------:|-----------------:|
|
| 103 |
+
| 0 | output.weight | 0x779620 | 0xd746000 |
|
| 104 |
+
| 1 | output_norm.weight | 0xdebf620 | 0x4000 |
|
| 105 |
+
| 2 | rope_freqs.weight | 0xdec3620 | 0x100 |
|
| 106 |
+
| 3 | token_embd.weight | 0xdec3720 | 0xd746000 |
|
| 107 |
+
| 4 | blk.0.attn_k.weight | 0x1b609720 | 0x150000 |
|
| 108 |
+
| 5 | blk.0.attn_norm.weight | 0x1b759720 | 0x4000 |
|
| 109 |
+
| 6 | blk.0.attn_output.weight | 0x1b75d720 | 0xb00000 |
|
| 110 |
+
| 7 | blk.0.attn_q.weight | 0x1c25d720 | 0x540000 |
|
| 111 |
+
| 8 | blk.0.attn_v.weight | 0x1c79d720 | 0x240000 |
|
| 112 |
+
| 9 | blk.0.ffn_down.weight | 0x1c9dd720 | 0x1f80000 |
|
| 113 |
+
| 10 | blk.0.ffn_gate.weight | 0x1e95d720 | 0x1260000 |
|
| 114 |
+
| 11 | blk.0.ffn_norm.weight | 0x1fbbd720 | 0x4000 |
|
| 115 |
+
| 12 | blk.0.ffn_up.weight | 0x1fbc1720 | 0x1260000 |
|
| 116 |
+
| 13 | blk.1.attn_k.weight | 0x20e21720 | 0x150000 |
|
| 117 |
+
| 14 | blk.1.attn_norm.weight | 0x20f71720 | 0x4000 |
|
| 118 |
+
| 15 | blk.1.attn_output.weight | 0x20f75720 | 0xb00000 |
|
| 119 |
+
| 16 | blk.1.attn_q.weight | 0x21a75720 | 0x540000 |
|
| 120 |
+
| 17 | blk.1.attn_v.weight | 0x21fb5720 | 0x240000 |
|
| 121 |
+
| 18 | blk.1.ffn_down.weight | 0x221f5720 | 0x2680000 |
|
| 122 |
+
| 19 | blk.1.ffn_gate.weight | 0x24875720 | 0x1260000 |
|
| 123 |
+
| 20 | blk.1.ffn_norm.weight | 0x25ad5720 | 0x4000 |
|
| 124 |
+
| 21 | blk.1.ffn_up.weight | 0x25ad9720 | 0x1260000 |
|
| 125 |
+
| 22 | blk.2.attn_k.weight | 0x26d39720 | 0x150000 |
|
| 126 |
+
| 23 | blk.2.attn_norm.weight | 0x26e89720 | 0x4000 |
|
| 127 |
+
| 24 | blk.2.attn_output.weight | 0x26e8d720 | 0xb00000 |
|
| 128 |
+
| 25 | blk.2.attn_q.weight | 0x2798d720 | 0x540000 |
|
| 129 |
+
| 26 | blk.2.attn_v.weight | 0x27ecd720 | 0x240000 |
|
| 130 |
+
| 27 | blk.2.ffn_down.weight | 0x2810d720 | 0x1f80000 |
|
| 131 |
+
| 28 | blk.2.ffn_gate.weight | 0x2a08d720 | 0x1260000 |
|
| 132 |
+
| 29 | blk.2.ffn_norm.weight | 0x2b2ed720 | 0x4000 |
|
| 133 |
+
| 30 | blk.2.ffn_up.weight | 0x2b2f1720 | 0x1260000 |
|
| 134 |
+
| 31 | blk.3.attn_k.weight | 0x2c551720 | 0x150000 |
|
| 135 |
+
| 32 | blk.3.attn_norm.weight | 0x2c6a1720 | 0x4000 |
|
| 136 |
+
| 33 | blk.3.attn_output.weight | 0x2c6a5720 | 0xb00000 |
|
| 137 |
+
| 34 | blk.3.attn_q.weight | 0x2d1a5720 | 0x540000 |
|
| 138 |
+
| 35 | blk.3.attn_v.weight | 0x2d6e5720 | 0x240000 |
|
| 139 |
+
| 36 | blk.3.ffn_down.weight | 0x2d925720 | 0x1f80000 |
|
| 140 |
+
| 37 | blk.3.ffn_gate.weight | 0x2f8a5720 | 0x1260000 |
|
| 141 |
+
| 38 | blk.3.ffn_norm.weight | 0x30b05720 | 0x4000 |
|
| 142 |
+
| 39 | blk.3.ffn_up.weight | 0x30b09720 | 0x1260000 |
|
| 143 |
+
| 40 | blk.4.attn_k.weight | 0x31d69720 | 0x150000 |
|
| 144 |
+
| 41 | blk.4.attn_norm.weight | 0x31eb9720 | 0x4000 |
|
| 145 |
+
| 42 | blk.4.attn_output.weight | 0x31ebd720 | 0xb00000 |
|
| 146 |
+
| 43 | blk.4.attn_q.weight | 0x329bd720 | 0x540000 |
|
| 147 |
+
| 44 | blk.4.attn_v.weight | 0x32efd720 | 0x240000 |
|
| 148 |
+
| 45 | blk.4.ffn_down.weight | 0x3313d720 | 0x1f80000 |
|
| 149 |
+
| 46 | blk.4.ffn_gate.weight | 0x350bd720 | 0x1260000 |
|
| 150 |
+
| 47 | blk.4.ffn_norm.weight | 0x3631d720 | 0x4000 |
|
| 151 |
+
| 48 | blk.4.ffn_up.weight | 0x36321720 | 0x1260000 |
|
| 152 |
+
| 49 | blk.5.attn_k.weight | 0x37581720 | 0x150000 |
|
| 153 |
+
| 50 | blk.5.attn_norm.weight | 0x376d1720 | 0x4000 |
|
| 154 |
+
| 51 | blk.5.attn_output.weight | 0x376d5720 | 0xb00000 |
|
| 155 |
+
| 52 | blk.5.attn_q.weight | 0x381d5720 | 0x540000 |
|
| 156 |
+
| 53 | blk.5.attn_v.weight | 0x38715720 | 0x240000 |
|
| 157 |
+
| 54 | blk.5.ffn_down.weight | 0x38955720 | 0x1f80000 |
|
| 158 |
+
| 55 | blk.5.ffn_gate.weight | 0x3a8d5720 | 0x1260000 |
|
| 159 |
+
| 56 | blk.5.ffn_norm.weight | 0x3bb35720 | 0x4000 |
|
| 160 |
+
| 57 | blk.5.ffn_up.weight | 0x3bb39720 | 0x1260000 |
|
| 161 |
+
| 58 | blk.6.attn_k.weight | 0x3cd99720 | 0x150000 |
|
| 162 |
+
| 59 | blk.6.attn_norm.weight | 0x3cee9720 | 0x4000 |
|
| 163 |
+
| 60 | blk.6.attn_output.weight | 0x3ceed720 | 0xb00000 |
|
| 164 |
+
| 61 | blk.6.attn_q.weight | 0x3d9ed720 | 0x540000 |
|
| 165 |
+
| 62 | blk.6.attn_v.weight | 0x3df2d720 | 0x240000 |
|
| 166 |
+
| 63 | blk.6.ffn_down.weight | 0x3e16d720 | 0x1f80000 |
|
| 167 |
+
| 64 | blk.6.ffn_gate.weight | 0x400ed720 | 0x1260000 |
|
| 168 |
+
| 65 | blk.6.ffn_norm.weight | 0x4134d720 | 0x4000 |
|
| 169 |
+
| 66 | blk.6.ffn_up.weight | 0x41351720 | 0x1260000 |
|
| 170 |
+
| 67 | blk.7.attn_k.weight | 0x425b1720 | 0x150000 |
|
| 171 |
+
| 68 | blk.7.attn_norm.weight | 0x42701720 | 0x4000 |
|
| 172 |
+
| 69 | blk.7.attn_output.weight | 0x42705720 | 0xb00000 |
|
| 173 |
+
| 70 | blk.7.attn_q.weight | 0x43205720 | 0x540000 |
|
| 174 |
+
| 71 | blk.7.attn_v.weight | 0x43745720 | 0x240000 |
|
| 175 |
+
| 72 | blk.7.ffn_down.weight | 0x43985720 | 0x1f80000 |
|
| 176 |
+
| 73 | blk.7.ffn_gate.weight | 0x45905720 | 0x1260000 |
|
| 177 |
+
| 74 | blk.7.ffn_norm.weight | 0x46b65720 | 0x4000 |
|
| 178 |
+
| 75 | blk.7.ffn_up.weight | 0x46b69720 | 0x1260000 |
|
| 179 |
+
| 76 | blk.8.attn_k.weight | 0x47dc9720 | 0x150000 |
|
| 180 |
+
| 77 | blk.8.attn_norm.weight | 0x47f19720 | 0x4000 |
|
| 181 |
+
| 78 | blk.8.attn_output.weight | 0x47f1d720 | 0xb00000 |
|
| 182 |
+
| 79 | blk.8.attn_q.weight | 0x48a1d720 | 0x540000 |
|
| 183 |
+
| 80 | blk.8.attn_v.weight | 0x48f5d720 | 0x240000 |
|
| 184 |
+
| 81 | blk.8.ffn_down.weight | 0x4919d720 | 0x1f80000 |
|
| 185 |
+
| 82 | blk.8.ffn_gate.weight | 0x4b11d720 | 0x1260000 |
|
| 186 |
+
| 83 | blk.8.ffn_norm.weight | 0x4c37d720 | 0x4000 |
|
| 187 |
+
| 84 | blk.8.ffn_up.weight | 0x4c381720 | 0x1260000 |
|
| 188 |
+
| 85 | blk.9.attn_k.weight | 0x4d5e1720 | 0x150000 |
|
| 189 |
+
| 86 | blk.9.attn_norm.weight | 0x4d731720 | 0x4000 |
|
| 190 |
+
| 87 | blk.9.attn_output.weight | 0x4d735720 | 0xb00000 |
|
| 191 |
+
| 88 | blk.9.attn_q.weight | 0x4e235720 | 0x540000 |
|
| 192 |
+
| 89 | blk.9.attn_v.weight | 0x4e775720 | 0x240000 |
|
| 193 |
+
| 90 | blk.9.ffn_down.weight | 0x4e9b5720 | 0x1f80000 |
|
| 194 |
+
| 91 | blk.9.ffn_gate.weight | 0x50935720 | 0x1260000 |
|
| 195 |
+
| 92 | blk.9.ffn_norm.weight | 0x51b95720 | 0x4000 |
|
| 196 |
+
| 93 | blk.9.ffn_up.weight | 0x51b99720 | 0x1260000 |
|
| 197 |
+
| 94 | blk.10.attn_k.weight | 0x52df9720 | 0x150000 |
|
| 198 |
+
| 95 | blk.10.attn_norm.weight | 0x52f49720 | 0x4000 |
|
| 199 |
+
| 96 | blk.10.attn_output.weight | 0x52f4d720 | 0xb00000 |
|
| 200 |
+
| 97 | blk.10.attn_q.weight | 0x53a4d720 | 0x540000 |
|
| 201 |
+
| 98 | blk.10.attn_v.weight | 0x53f8d720 | 0x240000 |
|
| 202 |
+
| 99 | blk.10.ffn_down.weight | 0x541cd720 | 0x1f80000 |
|
| 203 |
+
| 100 | blk.10.ffn_gate.weight | 0x5614d720 | 0x1260000 |
|
| 204 |
+
| 101 | blk.10.ffn_norm.weight | 0x573ad720 | 0x4000 |
|
| 205 |
+
| 102 | blk.10.ffn_up.weight | 0x573b1720 | 0x1260000 |
|
| 206 |
+
| 103 | blk.11.attn_k.weight | 0x58611720 | 0x150000 |
|
| 207 |
+
| 104 | blk.11.attn_norm.weight | 0x58761720 | 0x4000 |
|
| 208 |
+
| 105 | blk.11.attn_output.weight | 0x58765720 | 0xb00000 |
|
| 209 |
+
| 106 | blk.11.attn_q.weight | 0x59265720 | 0x540000 |
|
| 210 |
+
| 107 | blk.11.attn_v.weight | 0x597a5720 | 0x240000 |
|
| 211 |
+
| 108 | blk.11.ffn_down.weight | 0x599e5720 | 0x1f80000 |
|
| 212 |
+
| 109 | blk.11.ffn_gate.weight | 0x5b965720 | 0x1260000 |
|
| 213 |
+
| 110 | blk.11.ffn_norm.weight | 0x5cbc5720 | 0x4000 |
|
| 214 |
+
| 111 | blk.11.ffn_up.weight | 0x5cbc9720 | 0x1260000 |
|
| 215 |
+
| 112 | blk.12.attn_k.weight | 0x5de29720 | 0x150000 |
|
| 216 |
+
| 113 | blk.12.attn_norm.weight | 0x5df79720 | 0x4000 |
|
| 217 |
+
| 114 | blk.12.attn_output.weight | 0x5df7d720 | 0xb00000 |
|
| 218 |
+
| 115 | blk.12.attn_q.weight | 0x5ea7d720 | 0x540000 |
|
| 219 |
+
| 116 | blk.12.attn_v.weight | 0x5efbd720 | 0x240000 |
|
| 220 |
+
| 117 | blk.12.ffn_down.weight | 0x5f1fd720 | 0x1f80000 |
|
| 221 |
+
| 118 | blk.12.ffn_gate.weight | 0x6117d720 | 0x1260000 |
|
| 222 |
+
| 119 | blk.12.ffn_norm.weight | 0x623dd720 | 0x4000 |
|
| 223 |
+
| 120 | blk.12.ffn_up.weight | 0x623e1720 | 0x1260000 |
|
| 224 |
+
| 121 | blk.13.attn_k.weight | 0x63641720 | 0x1b8000 |
|
| 225 |
+
| 122 | blk.13.attn_norm.weight | 0x637f9720 | 0x4000 |
|
| 226 |
+
| 123 | blk.13.attn_output.weight | 0x637fd720 | 0xb00000 |
|
| 227 |
+
| 124 | blk.13.attn_q.weight | 0x642fd720 | 0x6e0000 |
|
| 228 |
+
| 125 | blk.13.attn_v.weight | 0x649dd720 | 0x240000 |
|
| 229 |
+
| 126 | blk.13.ffn_down.weight | 0x64c1d720 | 0x1f80000 |
|
| 230 |
+
| 127 | blk.13.ffn_gate.weight | 0x66b9d720 | 0x1260000 |
|
| 231 |
+
| 128 | blk.13.ffn_norm.weight | 0x67dfd720 | 0x4000 |
|
| 232 |
+
| 129 | blk.13.ffn_up.weight | 0x67e01720 | 0x1260000 |
|
| 233 |
+
| 130 | blk.14.attn_k.weight | 0x69061720 | 0x1b8000 |
|
| 234 |
+
| 131 | blk.14.attn_norm.weight | 0x69219720 | 0x4000 |
|
| 235 |
+
| 132 | blk.14.attn_output.weight | 0x6921d720 | 0xb00000 |
|
| 236 |
+
| 133 | blk.14.attn_q.weight | 0x69d1d720 | 0x6e0000 |
|
| 237 |
+
| 134 | blk.14.attn_v.weight | 0x6a3fd720 | 0x240000 |
|
| 238 |
+
| 135 | blk.14.ffn_down.weight | 0x6a63d720 | 0x1f80000 |
|
| 239 |
+
| 136 | blk.14.ffn_gate.weight | 0x6c5bd720 | 0x1260000 |
|
| 240 |
+
| 137 | blk.14.ffn_norm.weight | 0x6d81d720 | 0x4000 |
|
| 241 |
+
| 138 | blk.14.ffn_up.weight | 0x6d821720 | 0x1260000 |
|
| 242 |
+
| 139 | blk.15.attn_k.weight | 0x6ea81720 | 0x150000 |
|
| 243 |
+
| 140 | blk.15.attn_norm.weight | 0x6ebd1720 | 0x4000 |
|
| 244 |
+
| 141 | blk.15.attn_output.weight | 0x6ebd5720 | 0xb00000 |
|
| 245 |
+
| 142 | blk.15.attn_q.weight | 0x6f6d5720 | 0x540000 |
|
| 246 |
+
| 143 | blk.15.attn_v.weight | 0x6fc15720 | 0x240000 |
|
| 247 |
+
| 144 | blk.15.ffn_down.weight | 0x6fe55720 | 0x1f80000 |
|
| 248 |
+
| 145 | blk.15.ffn_gate.weight | 0x71dd5720 | 0x1260000 |
|
| 249 |
+
| 146 | blk.15.ffn_norm.weight | 0x73035720 | 0x4000 |
|
| 250 |
+
| 147 | blk.15.ffn_up.weight | 0x73039720 | 0x1260000 |
|
| 251 |
+
| 148 | blk.16.attn_k.weight | 0x74299720 | 0x1b8000 |
|
| 252 |
+
| 149 | blk.16.attn_norm.weight | 0x74451720 | 0x4000 |
|
| 253 |
+
| 150 | blk.16.attn_output.weight | 0x74455720 | 0xb00000 |
|
| 254 |
+
| 151 | blk.16.attn_q.weight | 0x74f55720 | 0x6e0000 |
|
| 255 |
+
| 152 | blk.16.attn_v.weight | 0x75635720 | 0x240000 |
|
| 256 |
+
| 153 | blk.16.ffn_down.weight | 0x75875720 | 0x1f80000 |
|
| 257 |
+
| 154 | blk.16.ffn_gate.weight | 0x777f5720 | 0x1810000 |
|
| 258 |
+
| 155 | blk.16.ffn_norm.weight | 0x79005720 | 0x4000 |
|
| 259 |
+
| 156 | blk.16.ffn_up.weight | 0x79009720 | 0x1810000 |
|
| 260 |
+
| 157 | blk.17.attn_k.weight | 0x7a819720 | 0x150000 |
|
| 261 |
+
| 158 | blk.17.attn_norm.weight | 0x7a969720 | 0x4000 |
|
| 262 |
+
| 159 | blk.17.attn_output.weight | 0x7a96d720 | 0xb00000 |
|
| 263 |
+
| 160 | blk.17.attn_q.weight | 0x7b46d720 | 0x540000 |
|
| 264 |
+
| 161 | blk.17.attn_v.weight | 0x7b9ad720 | 0x240000 |
|
| 265 |
+
| 162 | blk.17.ffn_down.weight | 0x7bbed720 | 0x2680000 |
|
| 266 |
+
| 163 | blk.17.ffn_gate.weight | 0x7e26d720 | 0x1810000 |
|
| 267 |
+
| 164 | blk.17.ffn_norm.weight | 0x7fa7d720 | 0x4000 |
|
| 268 |
+
| 165 | blk.17.ffn_up.weight | 0x7fa81720 | 0x1810000 |
|
| 269 |
+
| 166 | blk.18.attn_k.weight | 0x81291720 | 0x1b8000 |
|
| 270 |
+
| 167 | blk.18.attn_norm.weight | 0x81449720 | 0x4000 |
|
| 271 |
+
| 168 | blk.18.attn_output.weight | 0x8144d720 | 0xb00000 |
|
| 272 |
+
| 169 | blk.18.attn_q.weight | 0x81f4d720 | 0x6e0000 |
|
| 273 |
+
| 170 | blk.18.attn_v.weight | 0x8262d720 | 0x240000 |
|
| 274 |
+
| 171 | blk.18.ffn_down.weight | 0x8286d720 | 0x2680000 |
|
| 275 |
+
| 172 | blk.18.ffn_gate.weight | 0x84eed720 | 0x1810000 |
|
| 276 |
+
| 173 | blk.18.ffn_norm.weight | 0x866fd720 | 0x4000 |
|
| 277 |
+
| 174 | blk.18.ffn_up.weight | 0x86701720 | 0x1810000 |
|
| 278 |
+
| 175 | blk.19.attn_k.weight | 0x87f11720 | 0x1b8000 |
|
| 279 |
+
| 176 | blk.19.attn_norm.weight | 0x880c9720 | 0x4000 |
|
| 280 |
+
| 177 | blk.19.attn_output.weight | 0x880cd720 | 0xb00000 |
|
| 281 |
+
| 178 | blk.19.attn_q.weight | 0x88bcd720 | 0x6e0000 |
|
| 282 |
+
| 179 | blk.19.attn_v.weight | 0x892ad720 | 0x240000 |
|
| 283 |
+
| 180 | blk.19.ffn_down.weight | 0x894ed720 | 0x2680000 |
|
| 284 |
+
| 181 | blk.19.ffn_gate.weight | 0x8bb6d720 | 0x1810000 |
|
| 285 |
+
| 182 | blk.19.ffn_norm.weight | 0x8d37d720 | 0x4000 |
|
| 286 |
+
| 183 | blk.19.ffn_up.weight | 0x8d381720 | 0x1810000 |
|
| 287 |
+
| 184 | blk.20.attn_k.weight | 0x8eb91720 | 0x1b8000 |
|
| 288 |
+
| 185 | blk.20.attn_norm.weight | 0x8ed49720 | 0x4000 |
|
| 289 |
+
| 186 | blk.20.attn_output.weight | 0x8ed4d720 | 0xb00000 |
|
| 290 |
+
| 187 | blk.20.attn_q.weight | 0x8f84d720 | 0x6e0000 |
|
| 291 |
+
| 188 | blk.20.attn_v.weight | 0x8ff2d720 | 0x240000 |
|
| 292 |
+
| 189 | blk.20.ffn_down.weight | 0x9016d720 | 0x2680000 |
|
| 293 |
+
| 190 | blk.20.ffn_gate.weight | 0x927ed720 | 0x1810000 |
|
| 294 |
+
| 191 | blk.20.ffn_norm.weight | 0x93ffd720 | 0x4000 |
|
| 295 |
+
| 192 | blk.20.ffn_up.weight | 0x94001720 | 0x1810000 |
|
| 296 |
+
| 193 | blk.21.attn_k.weight | 0x95811720 | 0x1b8000 |
|
| 297 |
+
| 194 | blk.21.attn_norm.weight | 0x959c9720 | 0x4000 |
|
| 298 |
+
| 195 | blk.21.attn_output.weight | 0x959cd720 | 0xb00000 |
|
| 299 |
+
| 196 | blk.21.attn_q.weight | 0x964cd720 | 0x6e0000 |
|
| 300 |
+
| 197 | blk.21.attn_v.weight | 0x96bad720 | 0x240000 |
|
| 301 |
+
| 198 | blk.21.ffn_down.weight | 0x96ded720 | 0x2680000 |
|
| 302 |
+
| 199 | blk.21.ffn_gate.weight | 0x9946d720 | 0x1810000 |
|
| 303 |
+
| 200 | blk.21.ffn_norm.weight | 0x9ac7d720 | 0x4000 |
|
| 304 |
+
| 201 | blk.21.ffn_up.weight | 0x9ac81720 | 0x1810000 |
|
| 305 |
+
| 202 | blk.22.attn_k.weight | 0x9c491720 | 0x1b8000 |
|
| 306 |
+
| 203 | blk.22.attn_norm.weight | 0x9c649720 | 0x4000 |
|
| 307 |
+
| 204 | blk.22.attn_output.weight | 0x9c64d720 | 0xb00000 |
|
| 308 |
+
| 205 | blk.22.attn_q.weight | 0x9d14d720 | 0x6e0000 |
|
| 309 |
+
| 206 | blk.22.attn_v.weight | 0x9d82d720 | 0x240000 |
|
| 310 |
+
| 207 | blk.22.ffn_down.weight | 0x9da6d720 | 0x2680000 |
|
| 311 |
+
| 208 | blk.22.ffn_gate.weight | 0xa00ed720 | 0x1810000 |
|
| 312 |
+
| 209 | blk.22.ffn_norm.weight | 0xa18fd720 | 0x4000 |
|
| 313 |
+
| 210 | blk.22.ffn_up.weight | 0xa1901720 | 0x1810000 |
|
| 314 |
+
| 211 | blk.23.attn_k.weight | 0xa3111720 | 0x1b8000 |
|
| 315 |
+
| 212 | blk.23.attn_norm.weight | 0xa32c9720 | 0x4000 |
|
| 316 |
+
| 213 | blk.23.attn_output.weight | 0xa32cd720 | 0xb00000 |
|
| 317 |
+
| 214 | blk.23.attn_q.weight | 0xa3dcd720 | 0x6e0000 |
|
| 318 |
+
| 215 | blk.23.attn_v.weight | 0xa44ad720 | 0x240000 |
|
| 319 |
+
| 216 | blk.23.ffn_down.weight | 0xa46ed720 | 0x2680000 |
|
| 320 |
+
| 217 | blk.23.ffn_gate.weight | 0xa6d6d720 | 0x1810000 |
|
| 321 |
+
| 218 | blk.23.ffn_norm.weight | 0xa857d720 | 0x4000 |
|
| 322 |
+
| 219 | blk.23.ffn_up.weight | 0xa8581720 | 0x1810000 |
|
| 323 |
+
| 220 | blk.24.attn_k.weight | 0xa9d91720 | 0x1b8000 |
|
| 324 |
+
| 221 | blk.24.attn_norm.weight | 0xa9f49720 | 0x4000 |
|
| 325 |
+
| 222 | blk.24.attn_output.weight | 0xa9f4d720 | 0xb00000 |
|
| 326 |
+
| 223 | blk.24.attn_q.weight | 0xaaa4d720 | 0x6e0000 |
|
| 327 |
+
| 224 | blk.24.attn_v.weight | 0xab12d720 | 0x240000 |
|
| 328 |
+
| 225 | blk.24.ffn_down.weight | 0xab36d720 | 0x2680000 |
|
| 329 |
+
| 226 | blk.24.ffn_gate.weight | 0xad9ed720 | 0x1810000 |
|
| 330 |
+
| 227 | blk.24.ffn_norm.weight | 0xaf1fd720 | 0x4000 |
|
| 331 |
+
| 228 | blk.24.ffn_up.weight | 0xaf201720 | 0x1810000 |
|
| 332 |
+
| 229 | blk.25.attn_k.weight | 0xb0a11720 | 0x1b8000 |
|
| 333 |
+
| 230 | blk.25.attn_norm.weight | 0xb0bc9720 | 0x4000 |
|
| 334 |
+
| 231 | blk.25.attn_output.weight | 0xb0bcd720 | 0xb00000 |
|
| 335 |
+
| 232 | blk.25.attn_q.weight | 0xb16cd720 | 0x6e0000 |
|
| 336 |
+
| 233 | blk.25.attn_v.weight | 0xb1dad720 | 0x240000 |
|
| 337 |
+
| 234 | blk.25.ffn_down.weight | 0xb1fed720 | 0x2680000 |
|
| 338 |
+
| 235 | blk.25.ffn_gate.weight | 0xb466d720 | 0x1810000 |
|
| 339 |
+
| 236 | blk.25.ffn_norm.weight | 0xb5e7d720 | 0x4000 |
|
| 340 |
+
| 237 | blk.25.ffn_up.weight | 0xb5e81720 | 0x1810000 |
|
| 341 |
+
| 238 | blk.26.attn_k.weight | 0xb7691720 | 0x1b8000 |
|
| 342 |
+
| 239 | blk.26.attn_norm.weight | 0xb7849720 | 0x4000 |
|
| 343 |
+
| 240 | blk.26.attn_output.weight | 0xb784d720 | 0xb00000 |
|
| 344 |
+
| 241 | blk.26.attn_q.weight | 0xb834d720 | 0x6e0000 |
|
| 345 |
+
| 242 | blk.26.attn_v.weight | 0xb8a2d720 | 0x240000 |
|
| 346 |
+
| 243 | blk.26.ffn_down.weight | 0xb8c6d720 | 0x2680000 |
|
| 347 |
+
| 244 | blk.26.ffn_gate.weight | 0xbb2ed720 | 0x1810000 |
|
| 348 |
+
| 245 | blk.26.ffn_norm.weight | 0xbcafd720 | 0x4000 |
|
| 349 |
+
| 246 | blk.26.ffn_up.weight | 0xbcb01720 | 0x1810000 |
|
| 350 |
+
| 247 | blk.27.attn_k.weight | 0xbe311720 | 0x1b8000 |
|
| 351 |
+
| 248 | blk.27.attn_norm.weight | 0xbe4c9720 | 0x4000 |
|
| 352 |
+
| 249 | blk.27.attn_output.weight | 0xbe4cd720 | 0xb00000 |
|
| 353 |
+
| 250 | blk.27.attn_q.weight | 0xbefcd720 | 0x6e0000 |
|
| 354 |
+
| 251 | blk.27.attn_v.weight | 0xbf6ad720 | 0x240000 |
|
| 355 |
+
| 252 | blk.27.ffn_down.weight | 0xbf8ed720 | 0x2680000 |
|
| 356 |
+
| 253 | blk.27.ffn_gate.weight | 0xc1f6d720 | 0x1810000 |
|
| 357 |
+
| 254 | blk.27.ffn_norm.weight | 0xc377d720 | 0x4000 |
|
| 358 |
+
| 255 | blk.27.ffn_up.weight | 0xc3781720 | 0x1810000 |
|
| 359 |
+
| 256 | blk.28.attn_k.weight | 0xc4f91720 | 0x1b8000 |
|
| 360 |
+
| 257 | blk.28.attn_norm.weight | 0xc5149720 | 0x4000 |
|
| 361 |
+
| 258 | blk.28.attn_output.weight | 0xc514d720 | 0xb00000 |
|
| 362 |
+
| 259 | blk.28.attn_q.weight | 0xc5c4d720 | 0x6e0000 |
|
| 363 |
+
| 260 | blk.28.attn_v.weight | 0xc632d720 | 0x240000 |
|
| 364 |
+
| 261 | blk.28.ffn_down.weight | 0xc656d720 | 0x2680000 |
|
| 365 |
+
| 262 | blk.28.ffn_gate.weight | 0xc8bed720 | 0x1810000 |
|
| 366 |
+
| 263 | blk.28.ffn_norm.weight | 0xca3fd720 | 0x4000 |
|
| 367 |
+
| 264 | blk.28.ffn_up.weight | 0xca401720 | 0x1810000 |
|
| 368 |
+
| 265 | blk.29.attn_k.weight | 0xcbc11720 | 0x1b8000 |
|
| 369 |
+
| 266 | blk.29.attn_norm.weight | 0xcbdc9720 | 0x4000 |
|
| 370 |
+
| 267 | blk.29.attn_output.weight | 0xcbdcd720 | 0xb00000 |
|
| 371 |
+
| 268 | blk.29.attn_q.weight | 0xcc8cd720 | 0x6e0000 |
|
| 372 |
+
| 269 | blk.29.attn_v.weight | 0xccfad720 | 0x240000 |
|
| 373 |
+
| 270 | blk.29.ffn_down.weight | 0xcd1ed720 | 0x2680000 |
|
| 374 |
+
| 271 | blk.29.ffn_gate.weight | 0xcf86d720 | 0x1810000 |
|
| 375 |
+
| 272 | blk.29.ffn_norm.weight | 0xd107d720 | 0x4000 |
|
| 376 |
+
| 273 | blk.29.ffn_up.weight | 0xd1081720 | 0x1810000 |
|
| 377 |
+
| 274 | blk.30.attn_k.weight | 0xd2891720 | 0x1b8000 |
|
| 378 |
+
| 275 | blk.30.attn_norm.weight | 0xd2a49720 | 0x4000 |
|
| 379 |
+
| 276 | blk.30.attn_output.weight | 0xd2a4d720 | 0xb00000 |
|
| 380 |
+
| 277 | blk.30.attn_q.weight | 0xd354d720 | 0x6e0000 |
|
| 381 |
+
| 278 | blk.30.attn_v.weight | 0xd3c2d720 | 0x240000 |
|
| 382 |
+
| 279 | blk.30.ffn_down.weight | 0xd3e6d720 | 0x2680000 |
|
| 383 |
+
| 280 | blk.30.ffn_gate.weight | 0xd64ed720 | 0x1810000 |
|
| 384 |
+
| 281 | blk.30.ffn_norm.weight | 0xd7cfd720 | 0x4000 |
|
| 385 |
+
| 282 | blk.30.ffn_up.weight | 0xd7d01720 | 0x1810000 |
|
| 386 |
+
| 283 | blk.31.attn_k.weight | 0xd9511720 | 0x150000 |
|
| 387 |
+
| 284 | blk.31.attn_norm.weight | 0xd9661720 | 0x4000 |
|
| 388 |
+
| 285 | blk.31.attn_output.weight | 0xd9665720 | 0xb00000 |
|
| 389 |
+
| 286 | blk.31.attn_q.weight | 0xda165720 | 0x540000 |
|
| 390 |
+
| 287 | blk.31.attn_v.weight | 0xda6a5720 | 0x240000 |
|
| 391 |
+
| 288 | blk.31.ffn_down.weight | 0xda8e5720 | 0x2680000 |
|
| 392 |
+
| 289 | blk.31.ffn_gate.weight | 0xdcf65720 | 0x1810000 |
|
| 393 |
+
| 290 | blk.31.ffn_norm.weight | 0xde775720 | 0x4000 |
|
| 394 |
+
| 291 | blk.31.ffn_up.weight | 0xde779720 | 0x1810000 |
|
| 395 |
+
|
| 396 |
+
### <a name="base">Base Tensor Group : ~1B Elements</a>
|
| 397 |
+
|
| 398 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 399 |
+
|-----:|:-------------------|:---------------------------------|:------------------|:----------------------|:-----|
|
| 400 |
+
| 0 | output.weight | Output (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | Q3_K |
|
| 401 |
+
| 1 | output_norm.weight | Output Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 402 |
+
| 2 | rope_freqs.weight | Rope_Freqs (W) | ( 64) 64 | 64 x 1 x 1 x 1 | F32 |
|
| 403 |
+
| 3 | token_embd.weight | Token Embedding (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | Q3_K |
|
| 404 |
+
|
| 405 |
+
- Total elements in base: ( ~1B) 1050677312
|
| 406 |
+
- Percentage of total elements: 13.08%
|
| 407 |
+
|
| 408 |
+
|
| 409 |
+
### <a name="blk_0">Block 0 Tensor Group : ~218M Elements</a>
|
| 410 |
+
|
| 411 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 412 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 413 |
+
| 4 | blk.0.attn_k.weight | Block 0 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 414 |
+
| 5 | blk.0.attn_norm.weight | Block 0 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 415 |
+
| 6 | blk.0.attn_output.weight | Block 0 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 416 |
+
| 7 | blk.0.attn_q.weight | Block 0 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 417 |
+
| 8 | blk.0.attn_v.weight | Block 0 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 418 |
+
| 9 | blk.0.ffn_down.weight | Block 0 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 419 |
+
| 10 | blk.0.ffn_gate.weight | Block 0 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 420 |
+
| 11 | blk.0.ffn_norm.weight | Block 0 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 421 |
+
| 12 | blk.0.ffn_up.weight | Block 0 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 422 |
+
|
| 423 |
+
- Total elements in blk.0: (~218M) 218112000
|
| 424 |
+
- Percentage of total elements: 2.72%
|
| 425 |
+
|
| 426 |
+
|
| 427 |
+
### <a name="blk_1">Block 1 Tensor Group : ~218M Elements</a>
|
| 428 |
+
|
| 429 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 430 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 431 |
+
| 13 | blk.1.attn_k.weight | Block 1 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 432 |
+
| 14 | blk.1.attn_norm.weight | Block 1 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 433 |
+
| 15 | blk.1.attn_output.weight | Block 1 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 434 |
+
| 16 | blk.1.attn_q.weight | Block 1 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 435 |
+
| 17 | blk.1.attn_v.weight | Block 1 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 436 |
+
| 18 | blk.1.ffn_down.weight | Block 1 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 437 |
+
| 19 | blk.1.ffn_gate.weight | Block 1 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 438 |
+
| 20 | blk.1.ffn_norm.weight | Block 1 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 439 |
+
| 21 | blk.1.ffn_up.weight | Block 1 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 440 |
+
|
| 441 |
+
- Total elements in blk.1: (~218M) 218112000
|
| 442 |
+
- Percentage of total elements: 2.72%
|
| 443 |
+
|
| 444 |
+
|
| 445 |
+
### <a name="blk_2">Block 2 Tensor Group : ~218M Elements</a>
|
| 446 |
+
|
| 447 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 448 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 449 |
+
| 22 | blk.2.attn_k.weight | Block 2 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 450 |
+
| 23 | blk.2.attn_norm.weight | Block 2 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 451 |
+
| 24 | blk.2.attn_output.weight | Block 2 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 452 |
+
| 25 | blk.2.attn_q.weight | Block 2 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 453 |
+
| 26 | blk.2.attn_v.weight | Block 2 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 454 |
+
| 27 | blk.2.ffn_down.weight | Block 2 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 455 |
+
| 28 | blk.2.ffn_gate.weight | Block 2 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 456 |
+
| 29 | blk.2.ffn_norm.weight | Block 2 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 457 |
+
| 30 | blk.2.ffn_up.weight | Block 2 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 458 |
+
|
| 459 |
+
- Total elements in blk.2: (~218M) 218112000
|
| 460 |
+
- Percentage of total elements: 2.72%
|
| 461 |
+
|
| 462 |
+
|
| 463 |
+
### <a name="blk_3">Block 3 Tensor Group : ~218M Elements</a>
|
| 464 |
+
|
| 465 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 466 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 467 |
+
| 31 | blk.3.attn_k.weight | Block 3 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 468 |
+
| 32 | blk.3.attn_norm.weight | Block 3 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 469 |
+
| 33 | blk.3.attn_output.weight | Block 3 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 470 |
+
| 34 | blk.3.attn_q.weight | Block 3 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 471 |
+
| 35 | blk.3.attn_v.weight | Block 3 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 472 |
+
| 36 | blk.3.ffn_down.weight | Block 3 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 473 |
+
| 37 | blk.3.ffn_gate.weight | Block 3 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 474 |
+
| 38 | blk.3.ffn_norm.weight | Block 3 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 475 |
+
| 39 | blk.3.ffn_up.weight | Block 3 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 476 |
+
|
| 477 |
+
- Total elements in blk.3: (~218M) 218112000
|
| 478 |
+
- Percentage of total elements: 2.72%
|
| 479 |
+
|
| 480 |
+
|
| 481 |
+
### <a name="blk_4">Block 4 Tensor Group : ~218M Elements</a>
|
| 482 |
+
|
| 483 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 484 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 485 |
+
| 40 | blk.4.attn_k.weight | Block 4 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 486 |
+
| 41 | blk.4.attn_norm.weight | Block 4 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 487 |
+
| 42 | blk.4.attn_output.weight | Block 4 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 488 |
+
| 43 | blk.4.attn_q.weight | Block 4 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 489 |
+
| 44 | blk.4.attn_v.weight | Block 4 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 490 |
+
| 45 | blk.4.ffn_down.weight | Block 4 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 491 |
+
| 46 | blk.4.ffn_gate.weight | Block 4 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 492 |
+
| 47 | blk.4.ffn_norm.weight | Block 4 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 493 |
+
| 48 | blk.4.ffn_up.weight | Block 4 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 494 |
+
|
| 495 |
+
- Total elements in blk.4: (~218M) 218112000
|
| 496 |
+
- Percentage of total elements: 2.72%
|
| 497 |
+
|
| 498 |
+
|
| 499 |
+
### <a name="blk_5">Block 5 Tensor Group : ~218M Elements</a>
|
| 500 |
+
|
| 501 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 502 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 503 |
+
| 49 | blk.5.attn_k.weight | Block 5 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 504 |
+
| 50 | blk.5.attn_norm.weight | Block 5 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 505 |
+
| 51 | blk.5.attn_output.weight | Block 5 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 506 |
+
| 52 | blk.5.attn_q.weight | Block 5 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 507 |
+
| 53 | blk.5.attn_v.weight | Block 5 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 508 |
+
| 54 | blk.5.ffn_down.weight | Block 5 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 509 |
+
| 55 | blk.5.ffn_gate.weight | Block 5 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 510 |
+
| 56 | blk.5.ffn_norm.weight | Block 5 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 511 |
+
| 57 | blk.5.ffn_up.weight | Block 5 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 512 |
+
|
| 513 |
+
- Total elements in blk.5: (~218M) 218112000
|
| 514 |
+
- Percentage of total elements: 2.72%
|
| 515 |
+
|
| 516 |
+
|
| 517 |
+
### <a name="blk_6">Block 6 Tensor Group : ~218M Elements</a>
|
| 518 |
+
|
| 519 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 520 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 521 |
+
| 58 | blk.6.attn_k.weight | Block 6 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 522 |
+
| 59 | blk.6.attn_norm.weight | Block 6 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 523 |
+
| 60 | blk.6.attn_output.weight | Block 6 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 524 |
+
| 61 | blk.6.attn_q.weight | Block 6 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 525 |
+
| 62 | blk.6.attn_v.weight | Block 6 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 526 |
+
| 63 | blk.6.ffn_down.weight | Block 6 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 527 |
+
| 64 | blk.6.ffn_gate.weight | Block 6 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 528 |
+
| 65 | blk.6.ffn_norm.weight | Block 6 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 529 |
+
| 66 | blk.6.ffn_up.weight | Block 6 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 530 |
+
|
| 531 |
+
- Total elements in blk.6: (~218M) 218112000
|
| 532 |
+
- Percentage of total elements: 2.72%
|
| 533 |
+
|
| 534 |
+
|
| 535 |
+
### <a name="blk_7">Block 7 Tensor Group : ~218M Elements</a>
|
| 536 |
+
|
| 537 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 538 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 539 |
+
| 67 | blk.7.attn_k.weight | Block 7 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 540 |
+
| 68 | blk.7.attn_norm.weight | Block 7 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 541 |
+
| 69 | blk.7.attn_output.weight | Block 7 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 542 |
+
| 70 | blk.7.attn_q.weight | Block 7 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 543 |
+
| 71 | blk.7.attn_v.weight | Block 7 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 544 |
+
| 72 | blk.7.ffn_down.weight | Block 7 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 545 |
+
| 73 | blk.7.ffn_gate.weight | Block 7 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 546 |
+
| 74 | blk.7.ffn_norm.weight | Block 7 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 547 |
+
| 75 | blk.7.ffn_up.weight | Block 7 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 548 |
+
|
| 549 |
+
- Total elements in blk.7: (~218M) 218112000
|
| 550 |
+
- Percentage of total elements: 2.72%
|
| 551 |
+
|
| 552 |
+
|
| 553 |
+
### <a name="blk_8">Block 8 Tensor Group : ~218M Elements</a>
|
| 554 |
+
|
| 555 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 556 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 557 |
+
| 76 | blk.8.attn_k.weight | Block 8 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 558 |
+
| 77 | blk.8.attn_norm.weight | Block 8 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 559 |
+
| 78 | blk.8.attn_output.weight | Block 8 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 560 |
+
| 79 | blk.8.attn_q.weight | Block 8 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 561 |
+
| 80 | blk.8.attn_v.weight | Block 8 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 562 |
+
| 81 | blk.8.ffn_down.weight | Block 8 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 563 |
+
| 82 | blk.8.ffn_gate.weight | Block 8 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 564 |
+
| 83 | blk.8.ffn_norm.weight | Block 8 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 565 |
+
| 84 | blk.8.ffn_up.weight | Block 8 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 566 |
+
|
| 567 |
+
- Total elements in blk.8: (~218M) 218112000
|
| 568 |
+
- Percentage of total elements: 2.72%
|
| 569 |
+
|
| 570 |
+
|
| 571 |
+
### <a name="blk_9">Block 9 Tensor Group : ~218M Elements</a>
|
| 572 |
+
|
| 573 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 574 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 575 |
+
| 85 | blk.9.attn_k.weight | Block 9 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 576 |
+
| 86 | blk.9.attn_norm.weight | Block 9 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 577 |
+
| 87 | blk.9.attn_output.weight | Block 9 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 578 |
+
| 88 | blk.9.attn_q.weight | Block 9 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 579 |
+
| 89 | blk.9.attn_v.weight | Block 9 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 580 |
+
| 90 | blk.9.ffn_down.weight | Block 9 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 581 |
+
| 91 | blk.9.ffn_gate.weight | Block 9 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 582 |
+
| 92 | blk.9.ffn_norm.weight | Block 9 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 583 |
+
| 93 | blk.9.ffn_up.weight | Block 9 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 584 |
+
|
| 585 |
+
- Total elements in blk.9: (~218M) 218112000
|
| 586 |
+
- Percentage of total elements: 2.72%
|
| 587 |
+
|
| 588 |
+
|
| 589 |
+
### <a name="blk_10">Block 10 Tensor Group : ~218M Elements</a>
|
| 590 |
+
|
| 591 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 592 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 593 |
+
| 94 | blk.10.attn_k.weight | Block 10 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 594 |
+
| 95 | blk.10.attn_norm.weight | Block 10 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 595 |
+
| 96 | blk.10.attn_output.weight | Block 10 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 596 |
+
| 97 | blk.10.attn_q.weight | Block 10 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 597 |
+
| 98 | blk.10.attn_v.weight | Block 10 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 598 |
+
| 99 | blk.10.ffn_down.weight | Block 10 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 599 |
+
| 100 | blk.10.ffn_gate.weight | Block 10 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 600 |
+
| 101 | blk.10.ffn_norm.weight | Block 10 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 601 |
+
| 102 | blk.10.ffn_up.weight | Block 10 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 602 |
+
|
| 603 |
+
- Total elements in blk.10: (~218M) 218112000
|
| 604 |
+
- Percentage of total elements: 2.72%
|
| 605 |
+
|
| 606 |
+
|
| 607 |
+
### <a name="blk_11">Block 11 Tensor Group : ~218M Elements</a>
|
| 608 |
+
|
| 609 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 610 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 611 |
+
| 103 | blk.11.attn_k.weight | Block 11 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 612 |
+
| 104 | blk.11.attn_norm.weight | Block 11 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 613 |
+
| 105 | blk.11.attn_output.weight | Block 11 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 614 |
+
| 106 | blk.11.attn_q.weight | Block 11 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 615 |
+
| 107 | blk.11.attn_v.weight | Block 11 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 616 |
+
| 108 | blk.11.ffn_down.weight | Block 11 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 617 |
+
| 109 | blk.11.ffn_gate.weight | Block 11 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 618 |
+
| 110 | blk.11.ffn_norm.weight | Block 11 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 619 |
+
| 111 | blk.11.ffn_up.weight | Block 11 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 620 |
+
|
| 621 |
+
- Total elements in blk.11: (~218M) 218112000
|
| 622 |
+
- Percentage of total elements: 2.72%
|
| 623 |
+
|
| 624 |
+
|
| 625 |
+
### <a name="blk_12">Block 12 Tensor Group : ~218M Elements</a>
|
| 626 |
+
|
| 627 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 628 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 629 |
+
| 112 | blk.12.attn_k.weight | Block 12 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 630 |
+
| 113 | blk.12.attn_norm.weight | Block 12 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 631 |
+
| 114 | blk.12.attn_output.weight | Block 12 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 632 |
+
| 115 | blk.12.attn_q.weight | Block 12 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 633 |
+
| 116 | blk.12.attn_v.weight | Block 12 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 634 |
+
| 117 | blk.12.ffn_down.weight | Block 12 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 635 |
+
| 118 | blk.12.ffn_gate.weight | Block 12 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 636 |
+
| 119 | blk.12.ffn_norm.weight | Block 12 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 637 |
+
| 120 | blk.12.ffn_up.weight | Block 12 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 638 |
+
|
| 639 |
+
- Total elements in blk.12: (~218M) 218112000
|
| 640 |
+
- Percentage of total elements: 2.72%
|
| 641 |
+
|
| 642 |
+
|
| 643 |
+
### <a name="blk_13">Block 13 Tensor Group : ~218M Elements</a>
|
| 644 |
+
|
| 645 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 646 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 647 |
+
| 121 | blk.13.attn_k.weight | Block 13 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 648 |
+
| 122 | blk.13.attn_norm.weight | Block 13 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 649 |
+
| 123 | blk.13.attn_output.weight | Block 13 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 650 |
+
| 124 | blk.13.attn_q.weight | Block 13 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 651 |
+
| 125 | blk.13.attn_v.weight | Block 13 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 652 |
+
| 126 | blk.13.ffn_down.weight | Block 13 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 653 |
+
| 127 | blk.13.ffn_gate.weight | Block 13 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 654 |
+
| 128 | blk.13.ffn_norm.weight | Block 13 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 655 |
+
| 129 | blk.13.ffn_up.weight | Block 13 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 656 |
+
|
| 657 |
+
- Total elements in blk.13: (~218M) 218112000
|
| 658 |
+
- Percentage of total elements: 2.72%
|
| 659 |
+
|
| 660 |
+
|
| 661 |
+
### <a name="blk_14">Block 14 Tensor Group : ~218M Elements</a>
|
| 662 |
+
|
| 663 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 664 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 665 |
+
| 130 | blk.14.attn_k.weight | Block 14 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 666 |
+
| 131 | blk.14.attn_norm.weight | Block 14 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 667 |
+
| 132 | blk.14.attn_output.weight | Block 14 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 668 |
+
| 133 | blk.14.attn_q.weight | Block 14 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 669 |
+
| 134 | blk.14.attn_v.weight | Block 14 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 670 |
+
| 135 | blk.14.ffn_down.weight | Block 14 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 671 |
+
| 136 | blk.14.ffn_gate.weight | Block 14 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 672 |
+
| 137 | blk.14.ffn_norm.weight | Block 14 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 673 |
+
| 138 | blk.14.ffn_up.weight | Block 14 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 674 |
+
|
| 675 |
+
- Total elements in blk.14: (~218M) 218112000
|
| 676 |
+
- Percentage of total elements: 2.72%
|
| 677 |
+
|
| 678 |
+
|
| 679 |
+
### <a name="blk_15">Block 15 Tensor Group : ~218M Elements</a>
|
| 680 |
+
|
| 681 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 682 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 683 |
+
| 139 | blk.15.attn_k.weight | Block 15 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 684 |
+
| 140 | blk.15.attn_norm.weight | Block 15 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 685 |
+
| 141 | blk.15.attn_output.weight | Block 15 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 686 |
+
| 142 | blk.15.attn_q.weight | Block 15 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 687 |
+
| 143 | blk.15.attn_v.weight | Block 15 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 688 |
+
| 144 | blk.15.ffn_down.weight | Block 15 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 689 |
+
| 145 | blk.15.ffn_gate.weight | Block 15 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 690 |
+
| 146 | blk.15.ffn_norm.weight | Block 15 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 691 |
+
| 147 | blk.15.ffn_up.weight | Block 15 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 692 |
+
|
| 693 |
+
- Total elements in blk.15: (~218M) 218112000
|
| 694 |
+
- Percentage of total elements: 2.72%
|
| 695 |
+
|
| 696 |
+
|
| 697 |
+
### <a name="blk_16">Block 16 Tensor Group : ~218M Elements</a>
|
| 698 |
+
|
| 699 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 700 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 701 |
+
| 148 | blk.16.attn_k.weight | Block 16 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 702 |
+
| 149 | blk.16.attn_norm.weight | Block 16 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 703 |
+
| 150 | blk.16.attn_output.weight | Block 16 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 704 |
+
| 151 | blk.16.attn_q.weight | Block 16 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 705 |
+
| 152 | blk.16.attn_v.weight | Block 16 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 706 |
+
| 153 | blk.16.ffn_down.weight | Block 16 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 707 |
+
| 154 | blk.16.ffn_gate.weight | Block 16 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 708 |
+
| 155 | blk.16.ffn_norm.weight | Block 16 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 709 |
+
| 156 | blk.16.ffn_up.weight | Block 16 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 710 |
+
|
| 711 |
+
- Total elements in blk.16: (~218M) 218112000
|
| 712 |
+
- Percentage of total elements: 2.72%
|
| 713 |
+
|
| 714 |
+
|
| 715 |
+
### <a name="blk_17">Block 17 Tensor Group : ~218M Elements</a>
|
| 716 |
+
|
| 717 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 718 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 719 |
+
| 157 | blk.17.attn_k.weight | Block 17 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 720 |
+
| 158 | blk.17.attn_norm.weight | Block 17 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 721 |
+
| 159 | blk.17.attn_output.weight | Block 17 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 722 |
+
| 160 | blk.17.attn_q.weight | Block 17 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 723 |
+
| 161 | blk.17.attn_v.weight | Block 17 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 724 |
+
| 162 | blk.17.ffn_down.weight | Block 17 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 725 |
+
| 163 | blk.17.ffn_gate.weight | Block 17 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 726 |
+
| 164 | blk.17.ffn_norm.weight | Block 17 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 727 |
+
| 165 | blk.17.ffn_up.weight | Block 17 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 728 |
+
|
| 729 |
+
- Total elements in blk.17: (~218M) 218112000
|
| 730 |
+
- Percentage of total elements: 2.72%
|
| 731 |
+
|
| 732 |
+
|
| 733 |
+
### <a name="blk_18">Block 18 Tensor Group : ~218M Elements</a>
|
| 734 |
+
|
| 735 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 736 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 737 |
+
| 166 | blk.18.attn_k.weight | Block 18 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 738 |
+
| 167 | blk.18.attn_norm.weight | Block 18 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 739 |
+
| 168 | blk.18.attn_output.weight | Block 18 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 740 |
+
| 169 | blk.18.attn_q.weight | Block 18 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 741 |
+
| 170 | blk.18.attn_v.weight | Block 18 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 742 |
+
| 171 | blk.18.ffn_down.weight | Block 18 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 743 |
+
| 172 | blk.18.ffn_gate.weight | Block 18 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 744 |
+
| 173 | blk.18.ffn_norm.weight | Block 18 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 745 |
+
| 174 | blk.18.ffn_up.weight | Block 18 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 746 |
+
|
| 747 |
+
- Total elements in blk.18: (~218M) 218112000
|
| 748 |
+
- Percentage of total elements: 2.72%
|
| 749 |
+
|
| 750 |
+
|
| 751 |
+
### <a name="blk_19">Block 19 Tensor Group : ~218M Elements</a>
|
| 752 |
+
|
| 753 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 754 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 755 |
+
| 175 | blk.19.attn_k.weight | Block 19 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 756 |
+
| 176 | blk.19.attn_norm.weight | Block 19 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 757 |
+
| 177 | blk.19.attn_output.weight | Block 19 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 758 |
+
| 178 | blk.19.attn_q.weight | Block 19 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 759 |
+
| 179 | blk.19.attn_v.weight | Block 19 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 760 |
+
| 180 | blk.19.ffn_down.weight | Block 19 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 761 |
+
| 181 | blk.19.ffn_gate.weight | Block 19 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 762 |
+
| 182 | blk.19.ffn_norm.weight | Block 19 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 763 |
+
| 183 | blk.19.ffn_up.weight | Block 19 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 764 |
+
|
| 765 |
+
- Total elements in blk.19: (~218M) 218112000
|
| 766 |
+
- Percentage of total elements: 2.72%
|
| 767 |
+
|
| 768 |
+
|
| 769 |
+
### <a name="blk_20">Block 20 Tensor Group : ~218M Elements</a>
|
| 770 |
+
|
| 771 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 772 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 773 |
+
| 184 | blk.20.attn_k.weight | Block 20 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 774 |
+
| 185 | blk.20.attn_norm.weight | Block 20 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 775 |
+
| 186 | blk.20.attn_output.weight | Block 20 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 776 |
+
| 187 | blk.20.attn_q.weight | Block 20 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 777 |
+
| 188 | blk.20.attn_v.weight | Block 20 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 778 |
+
| 189 | blk.20.ffn_down.weight | Block 20 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 779 |
+
| 190 | blk.20.ffn_gate.weight | Block 20 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 780 |
+
| 191 | blk.20.ffn_norm.weight | Block 20 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 781 |
+
| 192 | blk.20.ffn_up.weight | Block 20 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 782 |
+
|
| 783 |
+
- Total elements in blk.20: (~218M) 218112000
|
| 784 |
+
- Percentage of total elements: 2.72%
|
| 785 |
+
|
| 786 |
+
|
| 787 |
+
### <a name="blk_21">Block 21 Tensor Group : ~218M Elements</a>
|
| 788 |
+
|
| 789 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 790 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 791 |
+
| 193 | blk.21.attn_k.weight | Block 21 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 792 |
+
| 194 | blk.21.attn_norm.weight | Block 21 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 793 |
+
| 195 | blk.21.attn_output.weight | Block 21 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 794 |
+
| 196 | blk.21.attn_q.weight | Block 21 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 795 |
+
| 197 | blk.21.attn_v.weight | Block 21 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 796 |
+
| 198 | blk.21.ffn_down.weight | Block 21 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 797 |
+
| 199 | blk.21.ffn_gate.weight | Block 21 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 798 |
+
| 200 | blk.21.ffn_norm.weight | Block 21 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 799 |
+
| 201 | blk.21.ffn_up.weight | Block 21 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 800 |
+
|
| 801 |
+
- Total elements in blk.21: (~218M) 218112000
|
| 802 |
+
- Percentage of total elements: 2.72%
|
| 803 |
+
|
| 804 |
+
|
| 805 |
+
### <a name="blk_22">Block 22 Tensor Group : ~218M Elements</a>
|
| 806 |
+
|
| 807 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 808 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 809 |
+
| 202 | blk.22.attn_k.weight | Block 22 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 810 |
+
| 203 | blk.22.attn_norm.weight | Block 22 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 811 |
+
| 204 | blk.22.attn_output.weight | Block 22 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 812 |
+
| 205 | blk.22.attn_q.weight | Block 22 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 813 |
+
| 206 | blk.22.attn_v.weight | Block 22 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 814 |
+
| 207 | blk.22.ffn_down.weight | Block 22 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 815 |
+
| 208 | blk.22.ffn_gate.weight | Block 22 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 816 |
+
| 209 | blk.22.ffn_norm.weight | Block 22 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 817 |
+
| 210 | blk.22.ffn_up.weight | Block 22 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 818 |
+
|
| 819 |
+
- Total elements in blk.22: (~218M) 218112000
|
| 820 |
+
- Percentage of total elements: 2.72%
|
| 821 |
+
|
| 822 |
+
|
| 823 |
+
### <a name="blk_23">Block 23 Tensor Group : ~218M Elements</a>
|
| 824 |
+
|
| 825 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 826 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 827 |
+
| 211 | blk.23.attn_k.weight | Block 23 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 828 |
+
| 212 | blk.23.attn_norm.weight | Block 23 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 829 |
+
| 213 | blk.23.attn_output.weight | Block 23 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 830 |
+
| 214 | blk.23.attn_q.weight | Block 23 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 831 |
+
| 215 | blk.23.attn_v.weight | Block 23 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 832 |
+
| 216 | blk.23.ffn_down.weight | Block 23 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 833 |
+
| 217 | blk.23.ffn_gate.weight | Block 23 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 834 |
+
| 218 | blk.23.ffn_norm.weight | Block 23 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 835 |
+
| 219 | blk.23.ffn_up.weight | Block 23 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 836 |
+
|
| 837 |
+
- Total elements in blk.23: (~218M) 218112000
|
| 838 |
+
- Percentage of total elements: 2.72%
|
| 839 |
+
|
| 840 |
+
|
| 841 |
+
### <a name="blk_24">Block 24 Tensor Group : ~218M Elements</a>
|
| 842 |
+
|
| 843 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 844 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 845 |
+
| 220 | blk.24.attn_k.weight | Block 24 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 846 |
+
| 221 | blk.24.attn_norm.weight | Block 24 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 847 |
+
| 222 | blk.24.attn_output.weight | Block 24 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 848 |
+
| 223 | blk.24.attn_q.weight | Block 24 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 849 |
+
| 224 | blk.24.attn_v.weight | Block 24 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 850 |
+
| 225 | blk.24.ffn_down.weight | Block 24 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 851 |
+
| 226 | blk.24.ffn_gate.weight | Block 24 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 852 |
+
| 227 | blk.24.ffn_norm.weight | Block 24 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 853 |
+
| 228 | blk.24.ffn_up.weight | Block 24 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 854 |
+
|
| 855 |
+
- Total elements in blk.24: (~218M) 218112000
|
| 856 |
+
- Percentage of total elements: 2.72%
|
| 857 |
+
|
| 858 |
+
|
| 859 |
+
### <a name="blk_25">Block 25 Tensor Group : ~218M Elements</a>
|
| 860 |
+
|
| 861 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 862 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 863 |
+
| 229 | blk.25.attn_k.weight | Block 25 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 864 |
+
| 230 | blk.25.attn_norm.weight | Block 25 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 865 |
+
| 231 | blk.25.attn_output.weight | Block 25 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 866 |
+
| 232 | blk.25.attn_q.weight | Block 25 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 867 |
+
| 233 | blk.25.attn_v.weight | Block 25 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 868 |
+
| 234 | blk.25.ffn_down.weight | Block 25 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 869 |
+
| 235 | blk.25.ffn_gate.weight | Block 25 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 870 |
+
| 236 | blk.25.ffn_norm.weight | Block 25 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 871 |
+
| 237 | blk.25.ffn_up.weight | Block 25 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 872 |
+
|
| 873 |
+
- Total elements in blk.25: (~218M) 218112000
|
| 874 |
+
- Percentage of total elements: 2.72%
|
| 875 |
+
|
| 876 |
+
|
| 877 |
+
### <a name="blk_26">Block 26 Tensor Group : ~218M Elements</a>
|
| 878 |
+
|
| 879 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 880 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 881 |
+
| 238 | blk.26.attn_k.weight | Block 26 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 882 |
+
| 239 | blk.26.attn_norm.weight | Block 26 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 883 |
+
| 240 | blk.26.attn_output.weight | Block 26 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 884 |
+
| 241 | blk.26.attn_q.weight | Block 26 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 885 |
+
| 242 | blk.26.attn_v.weight | Block 26 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 886 |
+
| 243 | blk.26.ffn_down.weight | Block 26 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 887 |
+
| 244 | blk.26.ffn_gate.weight | Block 26 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 888 |
+
| 245 | blk.26.ffn_norm.weight | Block 26 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 889 |
+
| 246 | blk.26.ffn_up.weight | Block 26 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 890 |
+
|
| 891 |
+
- Total elements in blk.26: (~218M) 218112000
|
| 892 |
+
- Percentage of total elements: 2.72%
|
| 893 |
+
|
| 894 |
+
|
| 895 |
+
### <a name="blk_27">Block 27 Tensor Group : ~218M Elements</a>
|
| 896 |
+
|
| 897 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 898 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 899 |
+
| 247 | blk.27.attn_k.weight | Block 27 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 900 |
+
| 248 | blk.27.attn_norm.weight | Block 27 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 901 |
+
| 249 | blk.27.attn_output.weight | Block 27 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 902 |
+
| 250 | blk.27.attn_q.weight | Block 27 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 903 |
+
| 251 | blk.27.attn_v.weight | Block 27 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 904 |
+
| 252 | blk.27.ffn_down.weight | Block 27 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 905 |
+
| 253 | blk.27.ffn_gate.weight | Block 27 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 906 |
+
| 254 | blk.27.ffn_norm.weight | Block 27 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 907 |
+
| 255 | blk.27.ffn_up.weight | Block 27 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 908 |
+
|
| 909 |
+
- Total elements in blk.27: (~218M) 218112000
|
| 910 |
+
- Percentage of total elements: 2.72%
|
| 911 |
+
|
| 912 |
+
|
| 913 |
+
### <a name="blk_28">Block 28 Tensor Group : ~218M Elements</a>
|
| 914 |
+
|
| 915 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 916 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 917 |
+
| 256 | blk.28.attn_k.weight | Block 28 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 918 |
+
| 257 | blk.28.attn_norm.weight | Block 28 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 919 |
+
| 258 | blk.28.attn_output.weight | Block 28 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 920 |
+
| 259 | blk.28.attn_q.weight | Block 28 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 921 |
+
| 260 | blk.28.attn_v.weight | Block 28 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 922 |
+
| 261 | blk.28.ffn_down.weight | Block 28 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 923 |
+
| 262 | blk.28.ffn_gate.weight | Block 28 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 924 |
+
| 263 | blk.28.ffn_norm.weight | Block 28 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 925 |
+
| 264 | blk.28.ffn_up.weight | Block 28 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 926 |
+
|
| 927 |
+
- Total elements in blk.28: (~218M) 218112000
|
| 928 |
+
- Percentage of total elements: 2.72%
|
| 929 |
+
|
| 930 |
+
|
| 931 |
+
### <a name="blk_29">Block 29 Tensor Group : ~218M Elements</a>
|
| 932 |
+
|
| 933 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 934 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 935 |
+
| 265 | blk.29.attn_k.weight | Block 29 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 936 |
+
| 266 | blk.29.attn_norm.weight | Block 29 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 937 |
+
| 267 | blk.29.attn_output.weight | Block 29 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 938 |
+
| 268 | blk.29.attn_q.weight | Block 29 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 939 |
+
| 269 | blk.29.attn_v.weight | Block 29 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 940 |
+
| 270 | blk.29.ffn_down.weight | Block 29 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 941 |
+
| 271 | blk.29.ffn_gate.weight | Block 29 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 942 |
+
| 272 | blk.29.ffn_norm.weight | Block 29 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 943 |
+
| 273 | blk.29.ffn_up.weight | Block 29 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 944 |
+
|
| 945 |
+
- Total elements in blk.29: (~218M) 218112000
|
| 946 |
+
- Percentage of total elements: 2.72%
|
| 947 |
+
|
| 948 |
+
|
| 949 |
+
### <a name="blk_30">Block 30 Tensor Group : ~218M Elements</a>
|
| 950 |
+
|
| 951 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 952 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 953 |
+
| 274 | blk.30.attn_k.weight | Block 30 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 954 |
+
| 275 | blk.30.attn_norm.weight | Block 30 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 955 |
+
| 276 | blk.30.attn_output.weight | Block 30 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 956 |
+
| 277 | blk.30.attn_q.weight | Block 30 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 957 |
+
| 278 | blk.30.attn_v.weight | Block 30 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 958 |
+
| 279 | blk.30.ffn_down.weight | Block 30 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 959 |
+
| 280 | blk.30.ffn_gate.weight | Block 30 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 960 |
+
| 281 | blk.30.ffn_norm.weight | Block 30 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 961 |
+
| 282 | blk.30.ffn_up.weight | Block 30 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 962 |
+
|
| 963 |
+
- Total elements in blk.30: (~218M) 218112000
|
| 964 |
+
- Percentage of total elements: 2.72%
|
| 965 |
+
|
| 966 |
+
|
| 967 |
+
### <a name="blk_31">Block 31 Tensor Group : ~218M Elements</a>
|
| 968 |
+
|
| 969 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 970 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 971 |
+
| 283 | blk.31.attn_k.weight | Block 31 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 972 |
+
| 284 | blk.31.attn_norm.weight | Block 31 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 973 |
+
| 285 | blk.31.attn_output.weight | Block 31 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 974 |
+
| 286 | blk.31.attn_q.weight | Block 31 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 975 |
+
| 287 | blk.31.attn_v.weight | Block 31 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 976 |
+
| 288 | blk.31.ffn_down.weight | Block 31 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 977 |
+
| 289 | blk.31.ffn_gate.weight | Block 31 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 978 |
+
| 290 | blk.31.ffn_norm.weight | Block 31 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 979 |
+
| 291 | blk.31.ffn_up.weight | Block 31 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 980 |
+
|
| 981 |
+
- Total elements in blk.31: (~218M) 218112000
|
| 982 |
+
- Percentage of total elements: 2.72%
|
scores/Watt-Tool-8B-Q3_K_M.md
ADDED
|
@@ -0,0 +1,982 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Watt-Tool-8B-Q3_K_M.gguf - GGUF Internal File Dump
|
| 2 |
+
|
| 3 |
+
- Endian: LITTLE endian
|
| 4 |
+
|
| 5 |
+
## Key Value Metadata Store
|
| 6 |
+
|
| 7 |
+
There are 43 key-value pairs in this file
|
| 8 |
+
|
| 9 |
+
| POS | TYPE | Count | Key | Value |
|
| 10 |
+
|----:|:---------|-------:|:---------------------------------------|:--------------------------------------------------------------------|
|
| 11 |
+
| 1 | UINT32 | 1 | GGUF.version | 3 |
|
| 12 |
+
| 2 | UINT64 | 1 | GGUF.tensor_count | 292 |
|
| 13 |
+
| 3 | UINT64 | 1 | GGUF.kv_count | 40 |
|
| 14 |
+
| 4 | STRING | 1 | general.architecture | `llama` |
|
| 15 |
+
| 5 | STRING | 1 | general.type | `model` |
|
| 16 |
+
| 6 | STRING | 1 | general.name | `Watt Tool 8B GGUF` |
|
| 17 |
+
| 7 | STRING | 1 | general.finetune | `GGUF` |
|
| 18 |
+
| 8 | STRING | 1 | general.basename | `Watt-Tool` |
|
| 19 |
+
| 9 | STRING | 1 | general.size_label | `8B` |
|
| 20 |
+
| 10 | STRING | 1 | general.license | `apache-2.0` |
|
| 21 |
+
| 11 | UINT32 | 1 | general.base_model.count | 1 |
|
| 22 |
+
| 12 | STRING | 1 | general.base_model.0.name | `Llama 3.1 8B Instruct` |
|
| 23 |
+
| 13 | STRING | 1 | general.base_model.0.organization | `Meta Llama` |
|
| 24 |
+
| 14 | STRING | 1 | general.base_model.0.repo_url | `https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct` |
|
| 25 |
+
| 15 | [STRING] | 4 | general.tags | [ `function-calling`, `tool-use`, `llama`, `bfcl` ] |
|
| 26 |
+
| 16 | [STRING] | 1 | general.languages | [ `en` ] |
|
| 27 |
+
| 17 | UINT32 | 1 | llama.block_count | 32 |
|
| 28 |
+
| 18 | UINT32 | 1 | llama.context_length | 131072 |
|
| 29 |
+
| 19 | UINT32 | 1 | llama.embedding_length | 4096 |
|
| 30 |
+
| 20 | UINT32 | 1 | llama.feed_forward_length | 14336 |
|
| 31 |
+
| 21 | UINT32 | 1 | llama.attention.head_count | 32 |
|
| 32 |
+
| 22 | UINT32 | 1 | llama.attention.head_count_kv | 8 |
|
| 33 |
+
| 23 | FLOAT32 | 1 | llama.rope.freq_base | 500000.0 |
|
| 34 |
+
| 24 | FLOAT32 | 1 | llama.attention.layer_norm_rms_epsilon | 1e-05 |
|
| 35 |
+
| 25 | UINT32 | 1 | llama.attention.key_length | 128 |
|
| 36 |
+
| 26 | UINT32 | 1 | llama.attention.value_length | 128 |
|
| 37 |
+
| 27 | UINT32 | 1 | llama.vocab_size | 128256 |
|
| 38 |
+
| 28 | UINT32 | 1 | llama.rope.dimension_count | 128 |
|
| 39 |
+
| 29 | STRING | 1 | tokenizer.ggml.model | `gpt2` |
|
| 40 |
+
| 30 | STRING | 1 | tokenizer.ggml.pre | `llama-bpe` |
|
| 41 |
+
| 31 | [STRING] | 128256 | tokenizer.ggml.tokens | [ `!`, `"`, `#`, `$`, `%`, ... ] |
|
| 42 |
+
| 32 | [INT32] | 128256 | tokenizer.ggml.token_type | [ 1, 1, 1, 1, 1, 1, 1, ... ] |
|
| 43 |
+
| 33 | [STRING] | 280147 | tokenizer.ggml.merges | [ `Ġ Ġ`, `Ġ ĠĠĠ`, `ĠĠ ĠĠ`, `ĠĠĠ Ġ`, `i n`, ... ] |
|
| 44 |
+
| 34 | UINT32 | 1 | tokenizer.ggml.bos_token_id | 128000 |
|
| 45 |
+
| 35 | UINT32 | 1 | tokenizer.ggml.eos_token_id | 128009 |
|
| 46 |
+
| 36 | UINT32 | 1 | tokenizer.ggml.padding_token_id | 128009 |
|
| 47 |
+
| 37 | STRING | 1 | tokenizer.chat_template | `{{ '<|begin_of_text|>' }}{% if`...`d|>' }}{% endif %}{% endfor %}` |
|
| 48 |
+
| 38 | UINT32 | 1 | general.quantization_version | 2 |
|
| 49 |
+
| 39 | UINT32 | 1 | general.file_type | 12 |
|
| 50 |
+
| 40 | STRING | 1 | quantize.imatrix.file | `./imatrix/imatrix-Watt-Tool-8B-small.dat` |
|
| 51 |
+
| 41 | STRING | 1 | quantize.imatrix.dataset | `../../datasets/imatrix/calibration_eur_small.txt` |
|
| 52 |
+
| 42 | INT32 | 1 | quantize.imatrix.entries_count | 225 |
|
| 53 |
+
| 43 | INT32 | 1 | quantize.imatrix.chunks_count | 962 |
|
| 54 |
+
|
| 55 |
+
## Tensors Overview ~8B Elements
|
| 56 |
+
|
| 57 |
+
Total number of elements in all tensors: 8030261312 Elements
|
| 58 |
+
|
| 59 |
+
- [Watt-Tool-8B-Q3\_K\_M.gguf - GGUF Internal File Dump](#watt-tool-8b-q3_k_mgguf---gguf-internal-file-dump)
|
| 60 |
+
- [Key Value Metadata Store](#key-value-metadata-store)
|
| 61 |
+
- [Tensors Overview ~8B Elements](#tensors-overview-8b-elements)
|
| 62 |
+
- [Tensor Data Offset](#tensor-data-offset)
|
| 63 |
+
- [Base Tensor Group : ~1B Elements](#base-tensor-group--1b-elements)
|
| 64 |
+
- [Block 0 Tensor Group : ~218M Elements](#block-0-tensor-group--218m-elements)
|
| 65 |
+
- [Block 1 Tensor Group : ~218M Elements](#block-1-tensor-group--218m-elements)
|
| 66 |
+
- [Block 2 Tensor Group : ~218M Elements](#block-2-tensor-group--218m-elements)
|
| 67 |
+
- [Block 3 Tensor Group : ~218M Elements](#block-3-tensor-group--218m-elements)
|
| 68 |
+
- [Block 4 Tensor Group : ~218M Elements](#block-4-tensor-group--218m-elements)
|
| 69 |
+
- [Block 5 Tensor Group : ~218M Elements](#block-5-tensor-group--218m-elements)
|
| 70 |
+
- [Block 6 Tensor Group : ~218M Elements](#block-6-tensor-group--218m-elements)
|
| 71 |
+
- [Block 7 Tensor Group : ~218M Elements](#block-7-tensor-group--218m-elements)
|
| 72 |
+
- [Block 8 Tensor Group : ~218M Elements](#block-8-tensor-group--218m-elements)
|
| 73 |
+
- [Block 9 Tensor Group : ~218M Elements](#block-9-tensor-group--218m-elements)
|
| 74 |
+
- [Block 10 Tensor Group : ~218M Elements](#block-10-tensor-group--218m-elements)
|
| 75 |
+
- [Block 11 Tensor Group : ~218M Elements](#block-11-tensor-group--218m-elements)
|
| 76 |
+
- [Block 12 Tensor Group : ~218M Elements](#block-12-tensor-group--218m-elements)
|
| 77 |
+
- [Block 13 Tensor Group : ~218M Elements](#block-13-tensor-group--218m-elements)
|
| 78 |
+
- [Block 14 Tensor Group : ~218M Elements](#block-14-tensor-group--218m-elements)
|
| 79 |
+
- [Block 15 Tensor Group : ~218M Elements](#block-15-tensor-group--218m-elements)
|
| 80 |
+
- [Block 16 Tensor Group : ~218M Elements](#block-16-tensor-group--218m-elements)
|
| 81 |
+
- [Block 17 Tensor Group : ~218M Elements](#block-17-tensor-group--218m-elements)
|
| 82 |
+
- [Block 18 Tensor Group : ~218M Elements](#block-18-tensor-group--218m-elements)
|
| 83 |
+
- [Block 19 Tensor Group : ~218M Elements](#block-19-tensor-group--218m-elements)
|
| 84 |
+
- [Block 20 Tensor Group : ~218M Elements](#block-20-tensor-group--218m-elements)
|
| 85 |
+
- [Block 21 Tensor Group : ~218M Elements](#block-21-tensor-group--218m-elements)
|
| 86 |
+
- [Block 22 Tensor Group : ~218M Elements](#block-22-tensor-group--218m-elements)
|
| 87 |
+
- [Block 23 Tensor Group : ~218M Elements](#block-23-tensor-group--218m-elements)
|
| 88 |
+
- [Block 24 Tensor Group : ~218M Elements](#block-24-tensor-group--218m-elements)
|
| 89 |
+
- [Block 25 Tensor Group : ~218M Elements](#block-25-tensor-group--218m-elements)
|
| 90 |
+
- [Block 26 Tensor Group : ~218M Elements](#block-26-tensor-group--218m-elements)
|
| 91 |
+
- [Block 27 Tensor Group : ~218M Elements](#block-27-tensor-group--218m-elements)
|
| 92 |
+
- [Block 28 Tensor Group : ~218M Elements](#block-28-tensor-group--218m-elements)
|
| 93 |
+
- [Block 29 Tensor Group : ~218M Elements](#block-29-tensor-group--218m-elements)
|
| 94 |
+
- [Block 30 Tensor Group : ~218M Elements](#block-30-tensor-group--218m-elements)
|
| 95 |
+
- [Block 31 Tensor Group : ~218M Elements](#block-31-tensor-group--218m-elements)
|
| 96 |
+
|
| 97 |
+
### Tensor Data Offset
|
| 98 |
+
|
| 99 |
+
This table contains the offset and data segment relative to start of file
|
| 100 |
+
|
| 101 |
+
| T_ID | Tensor Layer Name | Data Offset (B) | Data Size (B) |
|
| 102 |
+
|-----:|:--------------------------|-----------------:|-----------------:|
|
| 103 |
+
| 0 | output.weight | 0x779620 | 0xd746000 |
|
| 104 |
+
| 1 | output_norm.weight | 0xdebf620 | 0x4000 |
|
| 105 |
+
| 2 | rope_freqs.weight | 0xdec3620 | 0x100 |
|
| 106 |
+
| 3 | token_embd.weight | 0xdec3720 | 0xd746000 |
|
| 107 |
+
| 4 | blk.0.attn_k.weight | 0x1b609720 | 0x150000 |
|
| 108 |
+
| 5 | blk.0.attn_norm.weight | 0x1b759720 | 0x4000 |
|
| 109 |
+
| 6 | blk.0.attn_output.weight | 0x1b75d720 | 0x900000 |
|
| 110 |
+
| 7 | blk.0.attn_q.weight | 0x1c05d720 | 0x540000 |
|
| 111 |
+
| 8 | blk.0.attn_v.weight | 0x1c59d720 | 0x1b8000 |
|
| 112 |
+
| 9 | blk.0.ffn_down.weight | 0x1c755720 | 0x1f80000 |
|
| 113 |
+
| 10 | blk.0.ffn_gate.weight | 0x1e6d5720 | 0x1260000 |
|
| 114 |
+
| 11 | blk.0.ffn_norm.weight | 0x1f935720 | 0x4000 |
|
| 115 |
+
| 12 | blk.0.ffn_up.weight | 0x1f939720 | 0x1260000 |
|
| 116 |
+
| 13 | blk.1.attn_k.weight | 0x20b99720 | 0x150000 |
|
| 117 |
+
| 14 | blk.1.attn_norm.weight | 0x20ce9720 | 0x4000 |
|
| 118 |
+
| 15 | blk.1.attn_output.weight | 0x20ced720 | 0x900000 |
|
| 119 |
+
| 16 | blk.1.attn_q.weight | 0x215ed720 | 0x540000 |
|
| 120 |
+
| 17 | blk.1.attn_v.weight | 0x21b2d720 | 0x1b8000 |
|
| 121 |
+
| 18 | blk.1.ffn_down.weight | 0x21ce5720 | 0x1f80000 |
|
| 122 |
+
| 19 | blk.1.ffn_gate.weight | 0x23c65720 | 0x1260000 |
|
| 123 |
+
| 20 | blk.1.ffn_norm.weight | 0x24ec5720 | 0x4000 |
|
| 124 |
+
| 21 | blk.1.ffn_up.weight | 0x24ec9720 | 0x1260000 |
|
| 125 |
+
| 22 | blk.2.attn_k.weight | 0x26129720 | 0x150000 |
|
| 126 |
+
| 23 | blk.2.attn_norm.weight | 0x26279720 | 0x4000 |
|
| 127 |
+
| 24 | blk.2.attn_output.weight | 0x2627d720 | 0x900000 |
|
| 128 |
+
| 25 | blk.2.attn_q.weight | 0x26b7d720 | 0x540000 |
|
| 129 |
+
| 26 | blk.2.attn_v.weight | 0x270bd720 | 0x1b8000 |
|
| 130 |
+
| 27 | blk.2.ffn_down.weight | 0x27275720 | 0x1f80000 |
|
| 131 |
+
| 28 | blk.2.ffn_gate.weight | 0x291f5720 | 0x1260000 |
|
| 132 |
+
| 29 | blk.2.ffn_norm.weight | 0x2a455720 | 0x4000 |
|
| 133 |
+
| 30 | blk.2.ffn_up.weight | 0x2a459720 | 0x1260000 |
|
| 134 |
+
| 31 | blk.3.attn_k.weight | 0x2b6b9720 | 0x150000 |
|
| 135 |
+
| 32 | blk.3.attn_norm.weight | 0x2b809720 | 0x4000 |
|
| 136 |
+
| 33 | blk.3.attn_output.weight | 0x2b80d720 | 0x900000 |
|
| 137 |
+
| 34 | blk.3.attn_q.weight | 0x2c10d720 | 0x540000 |
|
| 138 |
+
| 35 | blk.3.attn_v.weight | 0x2c64d720 | 0x1b8000 |
|
| 139 |
+
| 36 | blk.3.ffn_down.weight | 0x2c805720 | 0x1f80000 |
|
| 140 |
+
| 37 | blk.3.ffn_gate.weight | 0x2e785720 | 0x1260000 |
|
| 141 |
+
| 38 | blk.3.ffn_norm.weight | 0x2f9e5720 | 0x4000 |
|
| 142 |
+
| 39 | blk.3.ffn_up.weight | 0x2f9e9720 | 0x1260000 |
|
| 143 |
+
| 40 | blk.4.attn_k.weight | 0x30c49720 | 0x150000 |
|
| 144 |
+
| 41 | blk.4.attn_norm.weight | 0x30d99720 | 0x4000 |
|
| 145 |
+
| 42 | blk.4.attn_output.weight | 0x30d9d720 | 0x900000 |
|
| 146 |
+
| 43 | blk.4.attn_q.weight | 0x3169d720 | 0x540000 |
|
| 147 |
+
| 44 | blk.4.attn_v.weight | 0x31bdd720 | 0x1b8000 |
|
| 148 |
+
| 45 | blk.4.ffn_down.weight | 0x31d95720 | 0x1f80000 |
|
| 149 |
+
| 46 | blk.4.ffn_gate.weight | 0x33d15720 | 0x1260000 |
|
| 150 |
+
| 47 | blk.4.ffn_norm.weight | 0x34f75720 | 0x4000 |
|
| 151 |
+
| 48 | blk.4.ffn_up.weight | 0x34f79720 | 0x1260000 |
|
| 152 |
+
| 49 | blk.5.attn_k.weight | 0x361d9720 | 0x150000 |
|
| 153 |
+
| 50 | blk.5.attn_norm.weight | 0x36329720 | 0x4000 |
|
| 154 |
+
| 51 | blk.5.attn_output.weight | 0x3632d720 | 0x900000 |
|
| 155 |
+
| 52 | blk.5.attn_q.weight | 0x36c2d720 | 0x540000 |
|
| 156 |
+
| 53 | blk.5.attn_v.weight | 0x3716d720 | 0x1b8000 |
|
| 157 |
+
| 54 | blk.5.ffn_down.weight | 0x37325720 | 0x1f80000 |
|
| 158 |
+
| 55 | blk.5.ffn_gate.weight | 0x392a5720 | 0x1260000 |
|
| 159 |
+
| 56 | blk.5.ffn_norm.weight | 0x3a505720 | 0x4000 |
|
| 160 |
+
| 57 | blk.5.ffn_up.weight | 0x3a509720 | 0x1260000 |
|
| 161 |
+
| 58 | blk.6.attn_k.weight | 0x3b769720 | 0x150000 |
|
| 162 |
+
| 59 | blk.6.attn_norm.weight | 0x3b8b9720 | 0x4000 |
|
| 163 |
+
| 60 | blk.6.attn_output.weight | 0x3b8bd720 | 0x900000 |
|
| 164 |
+
| 61 | blk.6.attn_q.weight | 0x3c1bd720 | 0x540000 |
|
| 165 |
+
| 62 | blk.6.attn_v.weight | 0x3c6fd720 | 0x1b8000 |
|
| 166 |
+
| 63 | blk.6.ffn_down.weight | 0x3c8b5720 | 0x1f80000 |
|
| 167 |
+
| 64 | blk.6.ffn_gate.weight | 0x3e835720 | 0x1260000 |
|
| 168 |
+
| 65 | blk.6.ffn_norm.weight | 0x3fa95720 | 0x4000 |
|
| 169 |
+
| 66 | blk.6.ffn_up.weight | 0x3fa99720 | 0x1260000 |
|
| 170 |
+
| 67 | blk.7.attn_k.weight | 0x40cf9720 | 0x150000 |
|
| 171 |
+
| 68 | blk.7.attn_norm.weight | 0x40e49720 | 0x4000 |
|
| 172 |
+
| 69 | blk.7.attn_output.weight | 0x40e4d720 | 0x900000 |
|
| 173 |
+
| 70 | blk.7.attn_q.weight | 0x4174d720 | 0x540000 |
|
| 174 |
+
| 71 | blk.7.attn_v.weight | 0x41c8d720 | 0x1b8000 |
|
| 175 |
+
| 72 | blk.7.ffn_down.weight | 0x41e45720 | 0x1f80000 |
|
| 176 |
+
| 73 | blk.7.ffn_gate.weight | 0x43dc5720 | 0x1260000 |
|
| 177 |
+
| 74 | blk.7.ffn_norm.weight | 0x45025720 | 0x4000 |
|
| 178 |
+
| 75 | blk.7.ffn_up.weight | 0x45029720 | 0x1260000 |
|
| 179 |
+
| 76 | blk.8.attn_k.weight | 0x46289720 | 0x150000 |
|
| 180 |
+
| 77 | blk.8.attn_norm.weight | 0x463d9720 | 0x4000 |
|
| 181 |
+
| 78 | blk.8.attn_output.weight | 0x463dd720 | 0x900000 |
|
| 182 |
+
| 79 | blk.8.attn_q.weight | 0x46cdd720 | 0x540000 |
|
| 183 |
+
| 80 | blk.8.attn_v.weight | 0x4721d720 | 0x1b8000 |
|
| 184 |
+
| 81 | blk.8.ffn_down.weight | 0x473d5720 | 0x1f80000 |
|
| 185 |
+
| 82 | blk.8.ffn_gate.weight | 0x49355720 | 0x1260000 |
|
| 186 |
+
| 83 | blk.8.ffn_norm.weight | 0x4a5b5720 | 0x4000 |
|
| 187 |
+
| 84 | blk.8.ffn_up.weight | 0x4a5b9720 | 0x1260000 |
|
| 188 |
+
| 85 | blk.9.attn_k.weight | 0x4b819720 | 0x150000 |
|
| 189 |
+
| 86 | blk.9.attn_norm.weight | 0x4b969720 | 0x4000 |
|
| 190 |
+
| 87 | blk.9.attn_output.weight | 0x4b96d720 | 0x900000 |
|
| 191 |
+
| 88 | blk.9.attn_q.weight | 0x4c26d720 | 0x540000 |
|
| 192 |
+
| 89 | blk.9.attn_v.weight | 0x4c7ad720 | 0x1b8000 |
|
| 193 |
+
| 90 | blk.9.ffn_down.weight | 0x4c965720 | 0x1f80000 |
|
| 194 |
+
| 91 | blk.9.ffn_gate.weight | 0x4e8e5720 | 0x1260000 |
|
| 195 |
+
| 92 | blk.9.ffn_norm.weight | 0x4fb45720 | 0x4000 |
|
| 196 |
+
| 93 | blk.9.ffn_up.weight | 0x4fb49720 | 0x1260000 |
|
| 197 |
+
| 94 | blk.10.attn_k.weight | 0x50da9720 | 0x150000 |
|
| 198 |
+
| 95 | blk.10.attn_norm.weight | 0x50ef9720 | 0x4000 |
|
| 199 |
+
| 96 | blk.10.attn_output.weight | 0x50efd720 | 0x900000 |
|
| 200 |
+
| 97 | blk.10.attn_q.weight | 0x517fd720 | 0x540000 |
|
| 201 |
+
| 98 | blk.10.attn_v.weight | 0x51d3d720 | 0x1b8000 |
|
| 202 |
+
| 99 | blk.10.ffn_down.weight | 0x51ef5720 | 0x1f80000 |
|
| 203 |
+
| 100 | blk.10.ffn_gate.weight | 0x53e75720 | 0x1260000 |
|
| 204 |
+
| 101 | blk.10.ffn_norm.weight | 0x550d5720 | 0x4000 |
|
| 205 |
+
| 102 | blk.10.ffn_up.weight | 0x550d9720 | 0x1260000 |
|
| 206 |
+
| 103 | blk.11.attn_k.weight | 0x56339720 | 0x150000 |
|
| 207 |
+
| 104 | blk.11.attn_norm.weight | 0x56489720 | 0x4000 |
|
| 208 |
+
| 105 | blk.11.attn_output.weight | 0x5648d720 | 0x900000 |
|
| 209 |
+
| 106 | blk.11.attn_q.weight | 0x56d8d720 | 0x540000 |
|
| 210 |
+
| 107 | blk.11.attn_v.weight | 0x572cd720 | 0x1b8000 |
|
| 211 |
+
| 108 | blk.11.ffn_down.weight | 0x57485720 | 0x1f80000 |
|
| 212 |
+
| 109 | blk.11.ffn_gate.weight | 0x59405720 | 0x1260000 |
|
| 213 |
+
| 110 | blk.11.ffn_norm.weight | 0x5a665720 | 0x4000 |
|
| 214 |
+
| 111 | blk.11.ffn_up.weight | 0x5a669720 | 0x1260000 |
|
| 215 |
+
| 112 | blk.12.attn_k.weight | 0x5b8c9720 | 0x150000 |
|
| 216 |
+
| 113 | blk.12.attn_norm.weight | 0x5ba19720 | 0x4000 |
|
| 217 |
+
| 114 | blk.12.attn_output.weight | 0x5ba1d720 | 0x900000 |
|
| 218 |
+
| 115 | blk.12.attn_q.weight | 0x5c31d720 | 0x540000 |
|
| 219 |
+
| 116 | blk.12.attn_v.weight | 0x5c85d720 | 0x1b8000 |
|
| 220 |
+
| 117 | blk.12.ffn_down.weight | 0x5ca15720 | 0x1f80000 |
|
| 221 |
+
| 118 | blk.12.ffn_gate.weight | 0x5e995720 | 0x1260000 |
|
| 222 |
+
| 119 | blk.12.ffn_norm.weight | 0x5fbf5720 | 0x4000 |
|
| 223 |
+
| 120 | blk.12.ffn_up.weight | 0x5fbf9720 | 0x1260000 |
|
| 224 |
+
| 121 | blk.13.attn_k.weight | 0x60e59720 | 0x1b8000 |
|
| 225 |
+
| 122 | blk.13.attn_norm.weight | 0x61011720 | 0x4000 |
|
| 226 |
+
| 123 | blk.13.attn_output.weight | 0x61015720 | 0x900000 |
|
| 227 |
+
| 124 | blk.13.attn_q.weight | 0x61915720 | 0x6e0000 |
|
| 228 |
+
| 125 | blk.13.attn_v.weight | 0x61ff5720 | 0x240000 |
|
| 229 |
+
| 126 | blk.13.ffn_down.weight | 0x62235720 | 0x1f80000 |
|
| 230 |
+
| 127 | blk.13.ffn_gate.weight | 0x641b5720 | 0x1260000 |
|
| 231 |
+
| 128 | blk.13.ffn_norm.weight | 0x65415720 | 0x4000 |
|
| 232 |
+
| 129 | blk.13.ffn_up.weight | 0x65419720 | 0x1260000 |
|
| 233 |
+
| 130 | blk.14.attn_k.weight | 0x66679720 | 0x1b8000 |
|
| 234 |
+
| 131 | blk.14.attn_norm.weight | 0x66831720 | 0x4000 |
|
| 235 |
+
| 132 | blk.14.attn_output.weight | 0x66835720 | 0x900000 |
|
| 236 |
+
| 133 | blk.14.attn_q.weight | 0x67135720 | 0x6e0000 |
|
| 237 |
+
| 134 | blk.14.attn_v.weight | 0x67815720 | 0x240000 |
|
| 238 |
+
| 135 | blk.14.ffn_down.weight | 0x67a55720 | 0x1f80000 |
|
| 239 |
+
| 136 | blk.14.ffn_gate.weight | 0x699d5720 | 0x1260000 |
|
| 240 |
+
| 137 | blk.14.ffn_norm.weight | 0x6ac35720 | 0x4000 |
|
| 241 |
+
| 138 | blk.14.ffn_up.weight | 0x6ac39720 | 0x1260000 |
|
| 242 |
+
| 139 | blk.15.attn_k.weight | 0x6be99720 | 0x150000 |
|
| 243 |
+
| 140 | blk.15.attn_norm.weight | 0x6bfe9720 | 0x4000 |
|
| 244 |
+
| 141 | blk.15.attn_output.weight | 0x6bfed720 | 0x900000 |
|
| 245 |
+
| 142 | blk.15.attn_q.weight | 0x6c8ed720 | 0x540000 |
|
| 246 |
+
| 143 | blk.15.attn_v.weight | 0x6ce2d720 | 0x1b8000 |
|
| 247 |
+
| 144 | blk.15.ffn_down.weight | 0x6cfe5720 | 0x1f80000 |
|
| 248 |
+
| 145 | blk.15.ffn_gate.weight | 0x6ef65720 | 0x1260000 |
|
| 249 |
+
| 146 | blk.15.ffn_norm.weight | 0x701c5720 | 0x4000 |
|
| 250 |
+
| 147 | blk.15.ffn_up.weight | 0x701c9720 | 0x1260000 |
|
| 251 |
+
| 148 | blk.16.attn_k.weight | 0x71429720 | 0x1b8000 |
|
| 252 |
+
| 149 | blk.16.attn_norm.weight | 0x715e1720 | 0x4000 |
|
| 253 |
+
| 150 | blk.16.attn_output.weight | 0x715e5720 | 0x900000 |
|
| 254 |
+
| 151 | blk.16.attn_q.weight | 0x71ee5720 | 0x6e0000 |
|
| 255 |
+
| 152 | blk.16.attn_v.weight | 0x725c5720 | 0x240000 |
|
| 256 |
+
| 153 | blk.16.ffn_down.weight | 0x72805720 | 0x1f80000 |
|
| 257 |
+
| 154 | blk.16.ffn_gate.weight | 0x74785720 | 0x1810000 |
|
| 258 |
+
| 155 | blk.16.ffn_norm.weight | 0x75f95720 | 0x4000 |
|
| 259 |
+
| 156 | blk.16.ffn_up.weight | 0x75f99720 | 0x1810000 |
|
| 260 |
+
| 157 | blk.17.attn_k.weight | 0x777a9720 | 0x150000 |
|
| 261 |
+
| 158 | blk.17.attn_norm.weight | 0x778f9720 | 0x4000 |
|
| 262 |
+
| 159 | blk.17.attn_output.weight | 0x778fd720 | 0x900000 |
|
| 263 |
+
| 160 | blk.17.attn_q.weight | 0x781fd720 | 0x540000 |
|
| 264 |
+
| 161 | blk.17.attn_v.weight | 0x7873d720 | 0x1b8000 |
|
| 265 |
+
| 162 | blk.17.ffn_down.weight | 0x788f5720 | 0x1f80000 |
|
| 266 |
+
| 163 | blk.17.ffn_gate.weight | 0x7a875720 | 0x1810000 |
|
| 267 |
+
| 164 | blk.17.ffn_norm.weight | 0x7c085720 | 0x4000 |
|
| 268 |
+
| 165 | blk.17.ffn_up.weight | 0x7c089720 | 0x1810000 |
|
| 269 |
+
| 166 | blk.18.attn_k.weight | 0x7d899720 | 0x1b8000 |
|
| 270 |
+
| 167 | blk.18.attn_norm.weight | 0x7da51720 | 0x4000 |
|
| 271 |
+
| 168 | blk.18.attn_output.weight | 0x7da55720 | 0x900000 |
|
| 272 |
+
| 169 | blk.18.attn_q.weight | 0x7e355720 | 0x6e0000 |
|
| 273 |
+
| 170 | blk.18.attn_v.weight | 0x7ea35720 | 0x240000 |
|
| 274 |
+
| 171 | blk.18.ffn_down.weight | 0x7ec75720 | 0x1f80000 |
|
| 275 |
+
| 172 | blk.18.ffn_gate.weight | 0x80bf5720 | 0x1810000 |
|
| 276 |
+
| 173 | blk.18.ffn_norm.weight | 0x82405720 | 0x4000 |
|
| 277 |
+
| 174 | blk.18.ffn_up.weight | 0x82409720 | 0x1810000 |
|
| 278 |
+
| 175 | blk.19.attn_k.weight | 0x83c19720 | 0x1b8000 |
|
| 279 |
+
| 176 | blk.19.attn_norm.weight | 0x83dd1720 | 0x4000 |
|
| 280 |
+
| 177 | blk.19.attn_output.weight | 0x83dd5720 | 0x900000 |
|
| 281 |
+
| 178 | blk.19.attn_q.weight | 0x846d5720 | 0x6e0000 |
|
| 282 |
+
| 179 | blk.19.attn_v.weight | 0x84db5720 | 0x240000 |
|
| 283 |
+
| 180 | blk.19.ffn_down.weight | 0x84ff5720 | 0x1f80000 |
|
| 284 |
+
| 181 | blk.19.ffn_gate.weight | 0x86f75720 | 0x1810000 |
|
| 285 |
+
| 182 | blk.19.ffn_norm.weight | 0x88785720 | 0x4000 |
|
| 286 |
+
| 183 | blk.19.ffn_up.weight | 0x88789720 | 0x1810000 |
|
| 287 |
+
| 184 | blk.20.attn_k.weight | 0x89f99720 | 0x1b8000 |
|
| 288 |
+
| 185 | blk.20.attn_norm.weight | 0x8a151720 | 0x4000 |
|
| 289 |
+
| 186 | blk.20.attn_output.weight | 0x8a155720 | 0x900000 |
|
| 290 |
+
| 187 | blk.20.attn_q.weight | 0x8aa55720 | 0x6e0000 |
|
| 291 |
+
| 188 | blk.20.attn_v.weight | 0x8b135720 | 0x240000 |
|
| 292 |
+
| 189 | blk.20.ffn_down.weight | 0x8b375720 | 0x1f80000 |
|
| 293 |
+
| 190 | blk.20.ffn_gate.weight | 0x8d2f5720 | 0x1810000 |
|
| 294 |
+
| 191 | blk.20.ffn_norm.weight | 0x8eb05720 | 0x4000 |
|
| 295 |
+
| 192 | blk.20.ffn_up.weight | 0x8eb09720 | 0x1810000 |
|
| 296 |
+
| 193 | blk.21.attn_k.weight | 0x90319720 | 0x1b8000 |
|
| 297 |
+
| 194 | blk.21.attn_norm.weight | 0x904d1720 | 0x4000 |
|
| 298 |
+
| 195 | blk.21.attn_output.weight | 0x904d5720 | 0x900000 |
|
| 299 |
+
| 196 | blk.21.attn_q.weight | 0x90dd5720 | 0x6e0000 |
|
| 300 |
+
| 197 | blk.21.attn_v.weight | 0x914b5720 | 0x240000 |
|
| 301 |
+
| 198 | blk.21.ffn_down.weight | 0x916f5720 | 0x1f80000 |
|
| 302 |
+
| 199 | blk.21.ffn_gate.weight | 0x93675720 | 0x1810000 |
|
| 303 |
+
| 200 | blk.21.ffn_norm.weight | 0x94e85720 | 0x4000 |
|
| 304 |
+
| 201 | blk.21.ffn_up.weight | 0x94e89720 | 0x1810000 |
|
| 305 |
+
| 202 | blk.22.attn_k.weight | 0x96699720 | 0x1b8000 |
|
| 306 |
+
| 203 | blk.22.attn_norm.weight | 0x96851720 | 0x4000 |
|
| 307 |
+
| 204 | blk.22.attn_output.weight | 0x96855720 | 0x900000 |
|
| 308 |
+
| 205 | blk.22.attn_q.weight | 0x97155720 | 0x6e0000 |
|
| 309 |
+
| 206 | blk.22.attn_v.weight | 0x97835720 | 0x240000 |
|
| 310 |
+
| 207 | blk.22.ffn_down.weight | 0x97a75720 | 0x1f80000 |
|
| 311 |
+
| 208 | blk.22.ffn_gate.weight | 0x999f5720 | 0x1810000 |
|
| 312 |
+
| 209 | blk.22.ffn_norm.weight | 0x9b205720 | 0x4000 |
|
| 313 |
+
| 210 | blk.22.ffn_up.weight | 0x9b209720 | 0x1810000 |
|
| 314 |
+
| 211 | blk.23.attn_k.weight | 0x9ca19720 | 0x1b8000 |
|
| 315 |
+
| 212 | blk.23.attn_norm.weight | 0x9cbd1720 | 0x4000 |
|
| 316 |
+
| 213 | blk.23.attn_output.weight | 0x9cbd5720 | 0x900000 |
|
| 317 |
+
| 214 | blk.23.attn_q.weight | 0x9d4d5720 | 0x6e0000 |
|
| 318 |
+
| 215 | blk.23.attn_v.weight | 0x9dbb5720 | 0x240000 |
|
| 319 |
+
| 216 | blk.23.ffn_down.weight | 0x9ddf5720 | 0x1f80000 |
|
| 320 |
+
| 217 | blk.23.ffn_gate.weight | 0x9fd75720 | 0x1810000 |
|
| 321 |
+
| 218 | blk.23.ffn_norm.weight | 0xa1585720 | 0x4000 |
|
| 322 |
+
| 219 | blk.23.ffn_up.weight | 0xa1589720 | 0x1810000 |
|
| 323 |
+
| 220 | blk.24.attn_k.weight | 0xa2d99720 | 0x1b8000 |
|
| 324 |
+
| 221 | blk.24.attn_norm.weight | 0xa2f51720 | 0x4000 |
|
| 325 |
+
| 222 | blk.24.attn_output.weight | 0xa2f55720 | 0x900000 |
|
| 326 |
+
| 223 | blk.24.attn_q.weight | 0xa3855720 | 0x6e0000 |
|
| 327 |
+
| 224 | blk.24.attn_v.weight | 0xa3f35720 | 0x240000 |
|
| 328 |
+
| 225 | blk.24.ffn_down.weight | 0xa4175720 | 0x1f80000 |
|
| 329 |
+
| 226 | blk.24.ffn_gate.weight | 0xa60f5720 | 0x1810000 |
|
| 330 |
+
| 227 | blk.24.ffn_norm.weight | 0xa7905720 | 0x4000 |
|
| 331 |
+
| 228 | blk.24.ffn_up.weight | 0xa7909720 | 0x1810000 |
|
| 332 |
+
| 229 | blk.25.attn_k.weight | 0xa9119720 | 0x1b8000 |
|
| 333 |
+
| 230 | blk.25.attn_norm.weight | 0xa92d1720 | 0x4000 |
|
| 334 |
+
| 231 | blk.25.attn_output.weight | 0xa92d5720 | 0x900000 |
|
| 335 |
+
| 232 | blk.25.attn_q.weight | 0xa9bd5720 | 0x6e0000 |
|
| 336 |
+
| 233 | blk.25.attn_v.weight | 0xaa2b5720 | 0x240000 |
|
| 337 |
+
| 234 | blk.25.ffn_down.weight | 0xaa4f5720 | 0x1f80000 |
|
| 338 |
+
| 235 | blk.25.ffn_gate.weight | 0xac475720 | 0x1810000 |
|
| 339 |
+
| 236 | blk.25.ffn_norm.weight | 0xadc85720 | 0x4000 |
|
| 340 |
+
| 237 | blk.25.ffn_up.weight | 0xadc89720 | 0x1810000 |
|
| 341 |
+
| 238 | blk.26.attn_k.weight | 0xaf499720 | 0x1b8000 |
|
| 342 |
+
| 239 | blk.26.attn_norm.weight | 0xaf651720 | 0x4000 |
|
| 343 |
+
| 240 | blk.26.attn_output.weight | 0xaf655720 | 0x900000 |
|
| 344 |
+
| 241 | blk.26.attn_q.weight | 0xaff55720 | 0x6e0000 |
|
| 345 |
+
| 242 | blk.26.attn_v.weight | 0xb0635720 | 0x240000 |
|
| 346 |
+
| 243 | blk.26.ffn_down.weight | 0xb0875720 | 0x1f80000 |
|
| 347 |
+
| 244 | blk.26.ffn_gate.weight | 0xb27f5720 | 0x1810000 |
|
| 348 |
+
| 245 | blk.26.ffn_norm.weight | 0xb4005720 | 0x4000 |
|
| 349 |
+
| 246 | blk.26.ffn_up.weight | 0xb4009720 | 0x1810000 |
|
| 350 |
+
| 247 | blk.27.attn_k.weight | 0xb5819720 | 0x1b8000 |
|
| 351 |
+
| 248 | blk.27.attn_norm.weight | 0xb59d1720 | 0x4000 |
|
| 352 |
+
| 249 | blk.27.attn_output.weight | 0xb59d5720 | 0x900000 |
|
| 353 |
+
| 250 | blk.27.attn_q.weight | 0xb62d5720 | 0x6e0000 |
|
| 354 |
+
| 251 | blk.27.attn_v.weight | 0xb69b5720 | 0x240000 |
|
| 355 |
+
| 252 | blk.27.ffn_down.weight | 0xb6bf5720 | 0x1f80000 |
|
| 356 |
+
| 253 | blk.27.ffn_gate.weight | 0xb8b75720 | 0x1810000 |
|
| 357 |
+
| 254 | blk.27.ffn_norm.weight | 0xba385720 | 0x4000 |
|
| 358 |
+
| 255 | blk.27.ffn_up.weight | 0xba389720 | 0x1810000 |
|
| 359 |
+
| 256 | blk.28.attn_k.weight | 0xbbb99720 | 0x1b8000 |
|
| 360 |
+
| 257 | blk.28.attn_norm.weight | 0xbbd51720 | 0x4000 |
|
| 361 |
+
| 258 | blk.28.attn_output.weight | 0xbbd55720 | 0x900000 |
|
| 362 |
+
| 259 | blk.28.attn_q.weight | 0xbc655720 | 0x6e0000 |
|
| 363 |
+
| 260 | blk.28.attn_v.weight | 0xbcd35720 | 0x240000 |
|
| 364 |
+
| 261 | blk.28.ffn_down.weight | 0xbcf75720 | 0x1f80000 |
|
| 365 |
+
| 262 | blk.28.ffn_gate.weight | 0xbeef5720 | 0x1810000 |
|
| 366 |
+
| 263 | blk.28.ffn_norm.weight | 0xc0705720 | 0x4000 |
|
| 367 |
+
| 264 | blk.28.ffn_up.weight | 0xc0709720 | 0x1810000 |
|
| 368 |
+
| 265 | blk.29.attn_k.weight | 0xc1f19720 | 0x1b8000 |
|
| 369 |
+
| 266 | blk.29.attn_norm.weight | 0xc20d1720 | 0x4000 |
|
| 370 |
+
| 267 | blk.29.attn_output.weight | 0xc20d5720 | 0x900000 |
|
| 371 |
+
| 268 | blk.29.attn_q.weight | 0xc29d5720 | 0x6e0000 |
|
| 372 |
+
| 269 | blk.29.attn_v.weight | 0xc30b5720 | 0x240000 |
|
| 373 |
+
| 270 | blk.29.ffn_down.weight | 0xc32f5720 | 0x1f80000 |
|
| 374 |
+
| 271 | blk.29.ffn_gate.weight | 0xc5275720 | 0x1810000 |
|
| 375 |
+
| 272 | blk.29.ffn_norm.weight | 0xc6a85720 | 0x4000 |
|
| 376 |
+
| 273 | blk.29.ffn_up.weight | 0xc6a89720 | 0x1810000 |
|
| 377 |
+
| 274 | blk.30.attn_k.weight | 0xc8299720 | 0x1b8000 |
|
| 378 |
+
| 275 | blk.30.attn_norm.weight | 0xc8451720 | 0x4000 |
|
| 379 |
+
| 276 | blk.30.attn_output.weight | 0xc8455720 | 0x900000 |
|
| 380 |
+
| 277 | blk.30.attn_q.weight | 0xc8d55720 | 0x6e0000 |
|
| 381 |
+
| 278 | blk.30.attn_v.weight | 0xc9435720 | 0x240000 |
|
| 382 |
+
| 279 | blk.30.ffn_down.weight | 0xc9675720 | 0x1f80000 |
|
| 383 |
+
| 280 | blk.30.ffn_gate.weight | 0xcb5f5720 | 0x1810000 |
|
| 384 |
+
| 281 | blk.30.ffn_norm.weight | 0xcce05720 | 0x4000 |
|
| 385 |
+
| 282 | blk.30.ffn_up.weight | 0xcce09720 | 0x1810000 |
|
| 386 |
+
| 283 | blk.31.attn_k.weight | 0xce619720 | 0x150000 |
|
| 387 |
+
| 284 | blk.31.attn_norm.weight | 0xce769720 | 0x4000 |
|
| 388 |
+
| 285 | blk.31.attn_output.weight | 0xce76d720 | 0x900000 |
|
| 389 |
+
| 286 | blk.31.attn_q.weight | 0xcf06d720 | 0x540000 |
|
| 390 |
+
| 287 | blk.31.attn_v.weight | 0xcf5ad720 | 0x1b8000 |
|
| 391 |
+
| 288 | blk.31.ffn_down.weight | 0xcf765720 | 0x1f80000 |
|
| 392 |
+
| 289 | blk.31.ffn_gate.weight | 0xd16e5720 | 0x1810000 |
|
| 393 |
+
| 290 | blk.31.ffn_norm.weight | 0xd2ef5720 | 0x4000 |
|
| 394 |
+
| 291 | blk.31.ffn_up.weight | 0xd2ef9720 | 0x1810000 |
|
| 395 |
+
|
| 396 |
+
### <a name="base">Base Tensor Group : ~1B Elements</a>
|
| 397 |
+
|
| 398 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 399 |
+
|-----:|:-------------------|:---------------------------------|:------------------|:----------------------|:-----|
|
| 400 |
+
| 0 | output.weight | Output (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | Q3_K |
|
| 401 |
+
| 1 | output_norm.weight | Output Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 402 |
+
| 2 | rope_freqs.weight | Rope_Freqs (W) | ( 64) 64 | 64 x 1 x 1 x 1 | F32 |
|
| 403 |
+
| 3 | token_embd.weight | Token Embedding (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | Q3_K |
|
| 404 |
+
|
| 405 |
+
- Total elements in base: ( ~1B) 1050677312
|
| 406 |
+
- Percentage of total elements: 13.08%
|
| 407 |
+
|
| 408 |
+
|
| 409 |
+
### <a name="blk_0">Block 0 Tensor Group : ~218M Elements</a>
|
| 410 |
+
|
| 411 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 412 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 413 |
+
| 4 | blk.0.attn_k.weight | Block 0 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 414 |
+
| 5 | blk.0.attn_norm.weight | Block 0 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 415 |
+
| 6 | blk.0.attn_output.weight | Block 0 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 416 |
+
| 7 | blk.0.attn_q.weight | Block 0 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 417 |
+
| 8 | blk.0.attn_v.weight | Block 0 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 418 |
+
| 9 | blk.0.ffn_down.weight | Block 0 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 419 |
+
| 10 | blk.0.ffn_gate.weight | Block 0 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 420 |
+
| 11 | blk.0.ffn_norm.weight | Block 0 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 421 |
+
| 12 | blk.0.ffn_up.weight | Block 0 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 422 |
+
|
| 423 |
+
- Total elements in blk.0: (~218M) 218112000
|
| 424 |
+
- Percentage of total elements: 2.72%
|
| 425 |
+
|
| 426 |
+
|
| 427 |
+
### <a name="blk_1">Block 1 Tensor Group : ~218M Elements</a>
|
| 428 |
+
|
| 429 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 430 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 431 |
+
| 13 | blk.1.attn_k.weight | Block 1 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 432 |
+
| 14 | blk.1.attn_norm.weight | Block 1 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 433 |
+
| 15 | blk.1.attn_output.weight | Block 1 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 434 |
+
| 16 | blk.1.attn_q.weight | Block 1 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 435 |
+
| 17 | blk.1.attn_v.weight | Block 1 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 436 |
+
| 18 | blk.1.ffn_down.weight | Block 1 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 437 |
+
| 19 | blk.1.ffn_gate.weight | Block 1 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 438 |
+
| 20 | blk.1.ffn_norm.weight | Block 1 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 439 |
+
| 21 | blk.1.ffn_up.weight | Block 1 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 440 |
+
|
| 441 |
+
- Total elements in blk.1: (~218M) 218112000
|
| 442 |
+
- Percentage of total elements: 2.72%
|
| 443 |
+
|
| 444 |
+
|
| 445 |
+
### <a name="blk_2">Block 2 Tensor Group : ~218M Elements</a>
|
| 446 |
+
|
| 447 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 448 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 449 |
+
| 22 | blk.2.attn_k.weight | Block 2 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 450 |
+
| 23 | blk.2.attn_norm.weight | Block 2 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 451 |
+
| 24 | blk.2.attn_output.weight | Block 2 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 452 |
+
| 25 | blk.2.attn_q.weight | Block 2 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 453 |
+
| 26 | blk.2.attn_v.weight | Block 2 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 454 |
+
| 27 | blk.2.ffn_down.weight | Block 2 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 455 |
+
| 28 | blk.2.ffn_gate.weight | Block 2 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 456 |
+
| 29 | blk.2.ffn_norm.weight | Block 2 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 457 |
+
| 30 | blk.2.ffn_up.weight | Block 2 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 458 |
+
|
| 459 |
+
- Total elements in blk.2: (~218M) 218112000
|
| 460 |
+
- Percentage of total elements: 2.72%
|
| 461 |
+
|
| 462 |
+
|
| 463 |
+
### <a name="blk_3">Block 3 Tensor Group : ~218M Elements</a>
|
| 464 |
+
|
| 465 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 466 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 467 |
+
| 31 | blk.3.attn_k.weight | Block 3 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 468 |
+
| 32 | blk.3.attn_norm.weight | Block 3 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 469 |
+
| 33 | blk.3.attn_output.weight | Block 3 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 470 |
+
| 34 | blk.3.attn_q.weight | Block 3 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 471 |
+
| 35 | blk.3.attn_v.weight | Block 3 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 472 |
+
| 36 | blk.3.ffn_down.weight | Block 3 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 473 |
+
| 37 | blk.3.ffn_gate.weight | Block 3 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 474 |
+
| 38 | blk.3.ffn_norm.weight | Block 3 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 475 |
+
| 39 | blk.3.ffn_up.weight | Block 3 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 476 |
+
|
| 477 |
+
- Total elements in blk.3: (~218M) 218112000
|
| 478 |
+
- Percentage of total elements: 2.72%
|
| 479 |
+
|
| 480 |
+
|
| 481 |
+
### <a name="blk_4">Block 4 Tensor Group : ~218M Elements</a>
|
| 482 |
+
|
| 483 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 484 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 485 |
+
| 40 | blk.4.attn_k.weight | Block 4 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 486 |
+
| 41 | blk.4.attn_norm.weight | Block 4 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 487 |
+
| 42 | blk.4.attn_output.weight | Block 4 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 488 |
+
| 43 | blk.4.attn_q.weight | Block 4 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 489 |
+
| 44 | blk.4.attn_v.weight | Block 4 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 490 |
+
| 45 | blk.4.ffn_down.weight | Block 4 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 491 |
+
| 46 | blk.4.ffn_gate.weight | Block 4 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 492 |
+
| 47 | blk.4.ffn_norm.weight | Block 4 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 493 |
+
| 48 | blk.4.ffn_up.weight | Block 4 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 494 |
+
|
| 495 |
+
- Total elements in blk.4: (~218M) 218112000
|
| 496 |
+
- Percentage of total elements: 2.72%
|
| 497 |
+
|
| 498 |
+
|
| 499 |
+
### <a name="blk_5">Block 5 Tensor Group : ~218M Elements</a>
|
| 500 |
+
|
| 501 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 502 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 503 |
+
| 49 | blk.5.attn_k.weight | Block 5 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 504 |
+
| 50 | blk.5.attn_norm.weight | Block 5 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 505 |
+
| 51 | blk.5.attn_output.weight | Block 5 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 506 |
+
| 52 | blk.5.attn_q.weight | Block 5 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 507 |
+
| 53 | blk.5.attn_v.weight | Block 5 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 508 |
+
| 54 | blk.5.ffn_down.weight | Block 5 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 509 |
+
| 55 | blk.5.ffn_gate.weight | Block 5 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 510 |
+
| 56 | blk.5.ffn_norm.weight | Block 5 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 511 |
+
| 57 | blk.5.ffn_up.weight | Block 5 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 512 |
+
|
| 513 |
+
- Total elements in blk.5: (~218M) 218112000
|
| 514 |
+
- Percentage of total elements: 2.72%
|
| 515 |
+
|
| 516 |
+
|
| 517 |
+
### <a name="blk_6">Block 6 Tensor Group : ~218M Elements</a>
|
| 518 |
+
|
| 519 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 520 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 521 |
+
| 58 | blk.6.attn_k.weight | Block 6 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 522 |
+
| 59 | blk.6.attn_norm.weight | Block 6 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 523 |
+
| 60 | blk.6.attn_output.weight | Block 6 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 524 |
+
| 61 | blk.6.attn_q.weight | Block 6 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 525 |
+
| 62 | blk.6.attn_v.weight | Block 6 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 526 |
+
| 63 | blk.6.ffn_down.weight | Block 6 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 527 |
+
| 64 | blk.6.ffn_gate.weight | Block 6 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 528 |
+
| 65 | blk.6.ffn_norm.weight | Block 6 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 529 |
+
| 66 | blk.6.ffn_up.weight | Block 6 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 530 |
+
|
| 531 |
+
- Total elements in blk.6: (~218M) 218112000
|
| 532 |
+
- Percentage of total elements: 2.72%
|
| 533 |
+
|
| 534 |
+
|
| 535 |
+
### <a name="blk_7">Block 7 Tensor Group : ~218M Elements</a>
|
| 536 |
+
|
| 537 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 538 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 539 |
+
| 67 | blk.7.attn_k.weight | Block 7 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 540 |
+
| 68 | blk.7.attn_norm.weight | Block 7 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 541 |
+
| 69 | blk.7.attn_output.weight | Block 7 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 542 |
+
| 70 | blk.7.attn_q.weight | Block 7 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 543 |
+
| 71 | blk.7.attn_v.weight | Block 7 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 544 |
+
| 72 | blk.7.ffn_down.weight | Block 7 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 545 |
+
| 73 | blk.7.ffn_gate.weight | Block 7 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 546 |
+
| 74 | blk.7.ffn_norm.weight | Block 7 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 547 |
+
| 75 | blk.7.ffn_up.weight | Block 7 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 548 |
+
|
| 549 |
+
- Total elements in blk.7: (~218M) 218112000
|
| 550 |
+
- Percentage of total elements: 2.72%
|
| 551 |
+
|
| 552 |
+
|
| 553 |
+
### <a name="blk_8">Block 8 Tensor Group : ~218M Elements</a>
|
| 554 |
+
|
| 555 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 556 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 557 |
+
| 76 | blk.8.attn_k.weight | Block 8 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 558 |
+
| 77 | blk.8.attn_norm.weight | Block 8 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 559 |
+
| 78 | blk.8.attn_output.weight | Block 8 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 560 |
+
| 79 | blk.8.attn_q.weight | Block 8 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 561 |
+
| 80 | blk.8.attn_v.weight | Block 8 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 562 |
+
| 81 | blk.8.ffn_down.weight | Block 8 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 563 |
+
| 82 | blk.8.ffn_gate.weight | Block 8 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 564 |
+
| 83 | blk.8.ffn_norm.weight | Block 8 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 565 |
+
| 84 | blk.8.ffn_up.weight | Block 8 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 566 |
+
|
| 567 |
+
- Total elements in blk.8: (~218M) 218112000
|
| 568 |
+
- Percentage of total elements: 2.72%
|
| 569 |
+
|
| 570 |
+
|
| 571 |
+
### <a name="blk_9">Block 9 Tensor Group : ~218M Elements</a>
|
| 572 |
+
|
| 573 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 574 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 575 |
+
| 85 | blk.9.attn_k.weight | Block 9 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 576 |
+
| 86 | blk.9.attn_norm.weight | Block 9 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 577 |
+
| 87 | blk.9.attn_output.weight | Block 9 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 578 |
+
| 88 | blk.9.attn_q.weight | Block 9 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 579 |
+
| 89 | blk.9.attn_v.weight | Block 9 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 580 |
+
| 90 | blk.9.ffn_down.weight | Block 9 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 581 |
+
| 91 | blk.9.ffn_gate.weight | Block 9 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 582 |
+
| 92 | blk.9.ffn_norm.weight | Block 9 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 583 |
+
| 93 | blk.9.ffn_up.weight | Block 9 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 584 |
+
|
| 585 |
+
- Total elements in blk.9: (~218M) 218112000
|
| 586 |
+
- Percentage of total elements: 2.72%
|
| 587 |
+
|
| 588 |
+
|
| 589 |
+
### <a name="blk_10">Block 10 Tensor Group : ~218M Elements</a>
|
| 590 |
+
|
| 591 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 592 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 593 |
+
| 94 | blk.10.attn_k.weight | Block 10 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 594 |
+
| 95 | blk.10.attn_norm.weight | Block 10 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 595 |
+
| 96 | blk.10.attn_output.weight | Block 10 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 596 |
+
| 97 | blk.10.attn_q.weight | Block 10 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 597 |
+
| 98 | blk.10.attn_v.weight | Block 10 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 598 |
+
| 99 | blk.10.ffn_down.weight | Block 10 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 599 |
+
| 100 | blk.10.ffn_gate.weight | Block 10 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 600 |
+
| 101 | blk.10.ffn_norm.weight | Block 10 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 601 |
+
| 102 | blk.10.ffn_up.weight | Block 10 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 602 |
+
|
| 603 |
+
- Total elements in blk.10: (~218M) 218112000
|
| 604 |
+
- Percentage of total elements: 2.72%
|
| 605 |
+
|
| 606 |
+
|
| 607 |
+
### <a name="blk_11">Block 11 Tensor Group : ~218M Elements</a>
|
| 608 |
+
|
| 609 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 610 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 611 |
+
| 103 | blk.11.attn_k.weight | Block 11 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 612 |
+
| 104 | blk.11.attn_norm.weight | Block 11 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 613 |
+
| 105 | blk.11.attn_output.weight | Block 11 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 614 |
+
| 106 | blk.11.attn_q.weight | Block 11 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 615 |
+
| 107 | blk.11.attn_v.weight | Block 11 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 616 |
+
| 108 | blk.11.ffn_down.weight | Block 11 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 617 |
+
| 109 | blk.11.ffn_gate.weight | Block 11 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 618 |
+
| 110 | blk.11.ffn_norm.weight | Block 11 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 619 |
+
| 111 | blk.11.ffn_up.weight | Block 11 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 620 |
+
|
| 621 |
+
- Total elements in blk.11: (~218M) 218112000
|
| 622 |
+
- Percentage of total elements: 2.72%
|
| 623 |
+
|
| 624 |
+
|
| 625 |
+
### <a name="blk_12">Block 12 Tensor Group : ~218M Elements</a>
|
| 626 |
+
|
| 627 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 628 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 629 |
+
| 112 | blk.12.attn_k.weight | Block 12 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 630 |
+
| 113 | blk.12.attn_norm.weight | Block 12 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 631 |
+
| 114 | blk.12.attn_output.weight | Block 12 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 632 |
+
| 115 | blk.12.attn_q.weight | Block 12 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 633 |
+
| 116 | blk.12.attn_v.weight | Block 12 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 634 |
+
| 117 | blk.12.ffn_down.weight | Block 12 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 635 |
+
| 118 | blk.12.ffn_gate.weight | Block 12 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 636 |
+
| 119 | blk.12.ffn_norm.weight | Block 12 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 637 |
+
| 120 | blk.12.ffn_up.weight | Block 12 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 638 |
+
|
| 639 |
+
- Total elements in blk.12: (~218M) 218112000
|
| 640 |
+
- Percentage of total elements: 2.72%
|
| 641 |
+
|
| 642 |
+
|
| 643 |
+
### <a name="blk_13">Block 13 Tensor Group : ~218M Elements</a>
|
| 644 |
+
|
| 645 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 646 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 647 |
+
| 121 | blk.13.attn_k.weight | Block 13 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 648 |
+
| 122 | blk.13.attn_norm.weight | Block 13 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 649 |
+
| 123 | blk.13.attn_output.weight | Block 13 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 650 |
+
| 124 | blk.13.attn_q.weight | Block 13 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 651 |
+
| 125 | blk.13.attn_v.weight | Block 13 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 652 |
+
| 126 | blk.13.ffn_down.weight | Block 13 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 653 |
+
| 127 | blk.13.ffn_gate.weight | Block 13 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 654 |
+
| 128 | blk.13.ffn_norm.weight | Block 13 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 655 |
+
| 129 | blk.13.ffn_up.weight | Block 13 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 656 |
+
|
| 657 |
+
- Total elements in blk.13: (~218M) 218112000
|
| 658 |
+
- Percentage of total elements: 2.72%
|
| 659 |
+
|
| 660 |
+
|
| 661 |
+
### <a name="blk_14">Block 14 Tensor Group : ~218M Elements</a>
|
| 662 |
+
|
| 663 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 664 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 665 |
+
| 130 | blk.14.attn_k.weight | Block 14 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 666 |
+
| 131 | blk.14.attn_norm.weight | Block 14 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 667 |
+
| 132 | blk.14.attn_output.weight | Block 14 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 668 |
+
| 133 | blk.14.attn_q.weight | Block 14 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 669 |
+
| 134 | blk.14.attn_v.weight | Block 14 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 670 |
+
| 135 | blk.14.ffn_down.weight | Block 14 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 671 |
+
| 136 | blk.14.ffn_gate.weight | Block 14 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 672 |
+
| 137 | blk.14.ffn_norm.weight | Block 14 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 673 |
+
| 138 | blk.14.ffn_up.weight | Block 14 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 674 |
+
|
| 675 |
+
- Total elements in blk.14: (~218M) 218112000
|
| 676 |
+
- Percentage of total elements: 2.72%
|
| 677 |
+
|
| 678 |
+
|
| 679 |
+
### <a name="blk_15">Block 15 Tensor Group : ~218M Elements</a>
|
| 680 |
+
|
| 681 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 682 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 683 |
+
| 139 | blk.15.attn_k.weight | Block 15 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 684 |
+
| 140 | blk.15.attn_norm.weight | Block 15 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 685 |
+
| 141 | blk.15.attn_output.weight | Block 15 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 686 |
+
| 142 | blk.15.attn_q.weight | Block 15 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 687 |
+
| 143 | blk.15.attn_v.weight | Block 15 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 688 |
+
| 144 | blk.15.ffn_down.weight | Block 15 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 689 |
+
| 145 | blk.15.ffn_gate.weight | Block 15 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 690 |
+
| 146 | blk.15.ffn_norm.weight | Block 15 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 691 |
+
| 147 | blk.15.ffn_up.weight | Block 15 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 692 |
+
|
| 693 |
+
- Total elements in blk.15: (~218M) 218112000
|
| 694 |
+
- Percentage of total elements: 2.72%
|
| 695 |
+
|
| 696 |
+
|
| 697 |
+
### <a name="blk_16">Block 16 Tensor Group : ~218M Elements</a>
|
| 698 |
+
|
| 699 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 700 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 701 |
+
| 148 | blk.16.attn_k.weight | Block 16 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 702 |
+
| 149 | blk.16.attn_norm.weight | Block 16 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 703 |
+
| 150 | blk.16.attn_output.weight | Block 16 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 704 |
+
| 151 | blk.16.attn_q.weight | Block 16 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 705 |
+
| 152 | blk.16.attn_v.weight | Block 16 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 706 |
+
| 153 | blk.16.ffn_down.weight | Block 16 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 707 |
+
| 154 | blk.16.ffn_gate.weight | Block 16 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 708 |
+
| 155 | blk.16.ffn_norm.weight | Block 16 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 709 |
+
| 156 | blk.16.ffn_up.weight | Block 16 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 710 |
+
|
| 711 |
+
- Total elements in blk.16: (~218M) 218112000
|
| 712 |
+
- Percentage of total elements: 2.72%
|
| 713 |
+
|
| 714 |
+
|
| 715 |
+
### <a name="blk_17">Block 17 Tensor Group : ~218M Elements</a>
|
| 716 |
+
|
| 717 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 718 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 719 |
+
| 157 | blk.17.attn_k.weight | Block 17 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 720 |
+
| 158 | blk.17.attn_norm.weight | Block 17 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 721 |
+
| 159 | blk.17.attn_output.weight | Block 17 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 722 |
+
| 160 | blk.17.attn_q.weight | Block 17 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 723 |
+
| 161 | blk.17.attn_v.weight | Block 17 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 724 |
+
| 162 | blk.17.ffn_down.weight | Block 17 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 725 |
+
| 163 | blk.17.ffn_gate.weight | Block 17 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 726 |
+
| 164 | blk.17.ffn_norm.weight | Block 17 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 727 |
+
| 165 | blk.17.ffn_up.weight | Block 17 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 728 |
+
|
| 729 |
+
- Total elements in blk.17: (~218M) 218112000
|
| 730 |
+
- Percentage of total elements: 2.72%
|
| 731 |
+
|
| 732 |
+
|
| 733 |
+
### <a name="blk_18">Block 18 Tensor Group : ~218M Elements</a>
|
| 734 |
+
|
| 735 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 736 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 737 |
+
| 166 | blk.18.attn_k.weight | Block 18 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 738 |
+
| 167 | blk.18.attn_norm.weight | Block 18 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 739 |
+
| 168 | blk.18.attn_output.weight | Block 18 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 740 |
+
| 169 | blk.18.attn_q.weight | Block 18 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 741 |
+
| 170 | blk.18.attn_v.weight | Block 18 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 742 |
+
| 171 | blk.18.ffn_down.weight | Block 18 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 743 |
+
| 172 | blk.18.ffn_gate.weight | Block 18 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 744 |
+
| 173 | blk.18.ffn_norm.weight | Block 18 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 745 |
+
| 174 | blk.18.ffn_up.weight | Block 18 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 746 |
+
|
| 747 |
+
- Total elements in blk.18: (~218M) 218112000
|
| 748 |
+
- Percentage of total elements: 2.72%
|
| 749 |
+
|
| 750 |
+
|
| 751 |
+
### <a name="blk_19">Block 19 Tensor Group : ~218M Elements</a>
|
| 752 |
+
|
| 753 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 754 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 755 |
+
| 175 | blk.19.attn_k.weight | Block 19 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 756 |
+
| 176 | blk.19.attn_norm.weight | Block 19 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 757 |
+
| 177 | blk.19.attn_output.weight | Block 19 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 758 |
+
| 178 | blk.19.attn_q.weight | Block 19 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 759 |
+
| 179 | blk.19.attn_v.weight | Block 19 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 760 |
+
| 180 | blk.19.ffn_down.weight | Block 19 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 761 |
+
| 181 | blk.19.ffn_gate.weight | Block 19 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 762 |
+
| 182 | blk.19.ffn_norm.weight | Block 19 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 763 |
+
| 183 | blk.19.ffn_up.weight | Block 19 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 764 |
+
|
| 765 |
+
- Total elements in blk.19: (~218M) 218112000
|
| 766 |
+
- Percentage of total elements: 2.72%
|
| 767 |
+
|
| 768 |
+
|
| 769 |
+
### <a name="blk_20">Block 20 Tensor Group : ~218M Elements</a>
|
| 770 |
+
|
| 771 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 772 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 773 |
+
| 184 | blk.20.attn_k.weight | Block 20 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 774 |
+
| 185 | blk.20.attn_norm.weight | Block 20 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 775 |
+
| 186 | blk.20.attn_output.weight | Block 20 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 776 |
+
| 187 | blk.20.attn_q.weight | Block 20 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 777 |
+
| 188 | blk.20.attn_v.weight | Block 20 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 778 |
+
| 189 | blk.20.ffn_down.weight | Block 20 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 779 |
+
| 190 | blk.20.ffn_gate.weight | Block 20 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 780 |
+
| 191 | blk.20.ffn_norm.weight | Block 20 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 781 |
+
| 192 | blk.20.ffn_up.weight | Block 20 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 782 |
+
|
| 783 |
+
- Total elements in blk.20: (~218M) 218112000
|
| 784 |
+
- Percentage of total elements: 2.72%
|
| 785 |
+
|
| 786 |
+
|
| 787 |
+
### <a name="blk_21">Block 21 Tensor Group : ~218M Elements</a>
|
| 788 |
+
|
| 789 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 790 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 791 |
+
| 193 | blk.21.attn_k.weight | Block 21 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 792 |
+
| 194 | blk.21.attn_norm.weight | Block 21 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 793 |
+
| 195 | blk.21.attn_output.weight | Block 21 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 794 |
+
| 196 | blk.21.attn_q.weight | Block 21 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 795 |
+
| 197 | blk.21.attn_v.weight | Block 21 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 796 |
+
| 198 | blk.21.ffn_down.weight | Block 21 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 797 |
+
| 199 | blk.21.ffn_gate.weight | Block 21 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 798 |
+
| 200 | blk.21.ffn_norm.weight | Block 21 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 799 |
+
| 201 | blk.21.ffn_up.weight | Block 21 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 800 |
+
|
| 801 |
+
- Total elements in blk.21: (~218M) 218112000
|
| 802 |
+
- Percentage of total elements: 2.72%
|
| 803 |
+
|
| 804 |
+
|
| 805 |
+
### <a name="blk_22">Block 22 Tensor Group : ~218M Elements</a>
|
| 806 |
+
|
| 807 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 808 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 809 |
+
| 202 | blk.22.attn_k.weight | Block 22 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 810 |
+
| 203 | blk.22.attn_norm.weight | Block 22 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 811 |
+
| 204 | blk.22.attn_output.weight | Block 22 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 812 |
+
| 205 | blk.22.attn_q.weight | Block 22 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 813 |
+
| 206 | blk.22.attn_v.weight | Block 22 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 814 |
+
| 207 | blk.22.ffn_down.weight | Block 22 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 815 |
+
| 208 | blk.22.ffn_gate.weight | Block 22 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 816 |
+
| 209 | blk.22.ffn_norm.weight | Block 22 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 817 |
+
| 210 | blk.22.ffn_up.weight | Block 22 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 818 |
+
|
| 819 |
+
- Total elements in blk.22: (~218M) 218112000
|
| 820 |
+
- Percentage of total elements: 2.72%
|
| 821 |
+
|
| 822 |
+
|
| 823 |
+
### <a name="blk_23">Block 23 Tensor Group : ~218M Elements</a>
|
| 824 |
+
|
| 825 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 826 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 827 |
+
| 211 | blk.23.attn_k.weight | Block 23 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 828 |
+
| 212 | blk.23.attn_norm.weight | Block 23 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 829 |
+
| 213 | blk.23.attn_output.weight | Block 23 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 830 |
+
| 214 | blk.23.attn_q.weight | Block 23 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 831 |
+
| 215 | blk.23.attn_v.weight | Block 23 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 832 |
+
| 216 | blk.23.ffn_down.weight | Block 23 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 833 |
+
| 217 | blk.23.ffn_gate.weight | Block 23 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 834 |
+
| 218 | blk.23.ffn_norm.weight | Block 23 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 835 |
+
| 219 | blk.23.ffn_up.weight | Block 23 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 836 |
+
|
| 837 |
+
- Total elements in blk.23: (~218M) 218112000
|
| 838 |
+
- Percentage of total elements: 2.72%
|
| 839 |
+
|
| 840 |
+
|
| 841 |
+
### <a name="blk_24">Block 24 Tensor Group : ~218M Elements</a>
|
| 842 |
+
|
| 843 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 844 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 845 |
+
| 220 | blk.24.attn_k.weight | Block 24 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 846 |
+
| 221 | blk.24.attn_norm.weight | Block 24 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 847 |
+
| 222 | blk.24.attn_output.weight | Block 24 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 848 |
+
| 223 | blk.24.attn_q.weight | Block 24 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 849 |
+
| 224 | blk.24.attn_v.weight | Block 24 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 850 |
+
| 225 | blk.24.ffn_down.weight | Block 24 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 851 |
+
| 226 | blk.24.ffn_gate.weight | Block 24 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 852 |
+
| 227 | blk.24.ffn_norm.weight | Block 24 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 853 |
+
| 228 | blk.24.ffn_up.weight | Block 24 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 854 |
+
|
| 855 |
+
- Total elements in blk.24: (~218M) 218112000
|
| 856 |
+
- Percentage of total elements: 2.72%
|
| 857 |
+
|
| 858 |
+
|
| 859 |
+
### <a name="blk_25">Block 25 Tensor Group : ~218M Elements</a>
|
| 860 |
+
|
| 861 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 862 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 863 |
+
| 229 | blk.25.attn_k.weight | Block 25 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 864 |
+
| 230 | blk.25.attn_norm.weight | Block 25 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 865 |
+
| 231 | blk.25.attn_output.weight | Block 25 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 866 |
+
| 232 | blk.25.attn_q.weight | Block 25 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 867 |
+
| 233 | blk.25.attn_v.weight | Block 25 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 868 |
+
| 234 | blk.25.ffn_down.weight | Block 25 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 869 |
+
| 235 | blk.25.ffn_gate.weight | Block 25 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 870 |
+
| 236 | blk.25.ffn_norm.weight | Block 25 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 871 |
+
| 237 | blk.25.ffn_up.weight | Block 25 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 872 |
+
|
| 873 |
+
- Total elements in blk.25: (~218M) 218112000
|
| 874 |
+
- Percentage of total elements: 2.72%
|
| 875 |
+
|
| 876 |
+
|
| 877 |
+
### <a name="blk_26">Block 26 Tensor Group : ~218M Elements</a>
|
| 878 |
+
|
| 879 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 880 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 881 |
+
| 238 | blk.26.attn_k.weight | Block 26 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 882 |
+
| 239 | blk.26.attn_norm.weight | Block 26 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 883 |
+
| 240 | blk.26.attn_output.weight | Block 26 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 884 |
+
| 241 | blk.26.attn_q.weight | Block 26 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 885 |
+
| 242 | blk.26.attn_v.weight | Block 26 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 886 |
+
| 243 | blk.26.ffn_down.weight | Block 26 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 887 |
+
| 244 | blk.26.ffn_gate.weight | Block 26 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 888 |
+
| 245 | blk.26.ffn_norm.weight | Block 26 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 889 |
+
| 246 | blk.26.ffn_up.weight | Block 26 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 890 |
+
|
| 891 |
+
- Total elements in blk.26: (~218M) 218112000
|
| 892 |
+
- Percentage of total elements: 2.72%
|
| 893 |
+
|
| 894 |
+
|
| 895 |
+
### <a name="blk_27">Block 27 Tensor Group : ~218M Elements</a>
|
| 896 |
+
|
| 897 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 898 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 899 |
+
| 247 | blk.27.attn_k.weight | Block 27 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 900 |
+
| 248 | blk.27.attn_norm.weight | Block 27 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 901 |
+
| 249 | blk.27.attn_output.weight | Block 27 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 902 |
+
| 250 | blk.27.attn_q.weight | Block 27 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 903 |
+
| 251 | blk.27.attn_v.weight | Block 27 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 904 |
+
| 252 | blk.27.ffn_down.weight | Block 27 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 905 |
+
| 253 | blk.27.ffn_gate.weight | Block 27 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 906 |
+
| 254 | blk.27.ffn_norm.weight | Block 27 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 907 |
+
| 255 | blk.27.ffn_up.weight | Block 27 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 908 |
+
|
| 909 |
+
- Total elements in blk.27: (~218M) 218112000
|
| 910 |
+
- Percentage of total elements: 2.72%
|
| 911 |
+
|
| 912 |
+
|
| 913 |
+
### <a name="blk_28">Block 28 Tensor Group : ~218M Elements</a>
|
| 914 |
+
|
| 915 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 916 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 917 |
+
| 256 | blk.28.attn_k.weight | Block 28 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 918 |
+
| 257 | blk.28.attn_norm.weight | Block 28 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 919 |
+
| 258 | blk.28.attn_output.weight | Block 28 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 920 |
+
| 259 | blk.28.attn_q.weight | Block 28 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 921 |
+
| 260 | blk.28.attn_v.weight | Block 28 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 922 |
+
| 261 | blk.28.ffn_down.weight | Block 28 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 923 |
+
| 262 | blk.28.ffn_gate.weight | Block 28 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 924 |
+
| 263 | blk.28.ffn_norm.weight | Block 28 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 925 |
+
| 264 | blk.28.ffn_up.weight | Block 28 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 926 |
+
|
| 927 |
+
- Total elements in blk.28: (~218M) 218112000
|
| 928 |
+
- Percentage of total elements: 2.72%
|
| 929 |
+
|
| 930 |
+
|
| 931 |
+
### <a name="blk_29">Block 29 Tensor Group : ~218M Elements</a>
|
| 932 |
+
|
| 933 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 934 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 935 |
+
| 265 | blk.29.attn_k.weight | Block 29 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 936 |
+
| 266 | blk.29.attn_norm.weight | Block 29 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 937 |
+
| 267 | blk.29.attn_output.weight | Block 29 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 938 |
+
| 268 | blk.29.attn_q.weight | Block 29 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 939 |
+
| 269 | blk.29.attn_v.weight | Block 29 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 940 |
+
| 270 | blk.29.ffn_down.weight | Block 29 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 941 |
+
| 271 | blk.29.ffn_gate.weight | Block 29 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 942 |
+
| 272 | blk.29.ffn_norm.weight | Block 29 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 943 |
+
| 273 | blk.29.ffn_up.weight | Block 29 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 944 |
+
|
| 945 |
+
- Total elements in blk.29: (~218M) 218112000
|
| 946 |
+
- Percentage of total elements: 2.72%
|
| 947 |
+
|
| 948 |
+
|
| 949 |
+
### <a name="blk_30">Block 30 Tensor Group : ~218M Elements</a>
|
| 950 |
+
|
| 951 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 952 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 953 |
+
| 274 | blk.30.attn_k.weight | Block 30 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 954 |
+
| 275 | blk.30.attn_norm.weight | Block 30 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 955 |
+
| 276 | blk.30.attn_output.weight | Block 30 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 956 |
+
| 277 | blk.30.attn_q.weight | Block 30 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 957 |
+
| 278 | blk.30.attn_v.weight | Block 30 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 958 |
+
| 279 | blk.30.ffn_down.weight | Block 30 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 959 |
+
| 280 | blk.30.ffn_gate.weight | Block 30 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 960 |
+
| 281 | blk.30.ffn_norm.weight | Block 30 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 961 |
+
| 282 | blk.30.ffn_up.weight | Block 30 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 962 |
+
|
| 963 |
+
- Total elements in blk.30: (~218M) 218112000
|
| 964 |
+
- Percentage of total elements: 2.72%
|
| 965 |
+
|
| 966 |
+
|
| 967 |
+
### <a name="blk_31">Block 31 Tensor Group : ~218M Elements</a>
|
| 968 |
+
|
| 969 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 970 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 971 |
+
| 283 | blk.31.attn_k.weight | Block 31 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 972 |
+
| 284 | blk.31.attn_norm.weight | Block 31 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 973 |
+
| 285 | blk.31.attn_output.weight | Block 31 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 974 |
+
| 286 | blk.31.attn_q.weight | Block 31 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 975 |
+
| 287 | blk.31.attn_v.weight | Block 31 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 976 |
+
| 288 | blk.31.ffn_down.weight | Block 31 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 977 |
+
| 289 | blk.31.ffn_gate.weight | Block 31 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 978 |
+
| 290 | blk.31.ffn_norm.weight | Block 31 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 979 |
+
| 291 | blk.31.ffn_up.weight | Block 31 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 980 |
+
|
| 981 |
+
- Total elements in blk.31: (~218M) 218112000
|
| 982 |
+
- Percentage of total elements: 2.72%
|
scores/Watt-Tool-8B-Q3_K_S.md
ADDED
|
@@ -0,0 +1,982 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Watt-Tool-8B-Q3_K_S.gguf - GGUF Internal File Dump
|
| 2 |
+
|
| 3 |
+
- Endian: LITTLE endian
|
| 4 |
+
|
| 5 |
+
## Key Value Metadata Store
|
| 6 |
+
|
| 7 |
+
There are 43 key-value pairs in this file
|
| 8 |
+
|
| 9 |
+
| POS | TYPE | Count | Key | Value |
|
| 10 |
+
|----:|:---------|-------:|:---------------------------------------|:--------------------------------------------------------------------|
|
| 11 |
+
| 1 | UINT32 | 1 | GGUF.version | 3 |
|
| 12 |
+
| 2 | UINT64 | 1 | GGUF.tensor_count | 292 |
|
| 13 |
+
| 3 | UINT64 | 1 | GGUF.kv_count | 40 |
|
| 14 |
+
| 4 | STRING | 1 | general.architecture | `llama` |
|
| 15 |
+
| 5 | STRING | 1 | general.type | `model` |
|
| 16 |
+
| 6 | STRING | 1 | general.name | `Watt Tool 8B GGUF` |
|
| 17 |
+
| 7 | STRING | 1 | general.finetune | `GGUF` |
|
| 18 |
+
| 8 | STRING | 1 | general.basename | `Watt-Tool` |
|
| 19 |
+
| 9 | STRING | 1 | general.size_label | `8B` |
|
| 20 |
+
| 10 | STRING | 1 | general.license | `apache-2.0` |
|
| 21 |
+
| 11 | UINT32 | 1 | general.base_model.count | 1 |
|
| 22 |
+
| 12 | STRING | 1 | general.base_model.0.name | `Llama 3.1 8B Instruct` |
|
| 23 |
+
| 13 | STRING | 1 | general.base_model.0.organization | `Meta Llama` |
|
| 24 |
+
| 14 | STRING | 1 | general.base_model.0.repo_url | `https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct` |
|
| 25 |
+
| 15 | [STRING] | 4 | general.tags | [ `function-calling`, `tool-use`, `llama`, `bfcl` ] |
|
| 26 |
+
| 16 | [STRING] | 1 | general.languages | [ `en` ] |
|
| 27 |
+
| 17 | UINT32 | 1 | llama.block_count | 32 |
|
| 28 |
+
| 18 | UINT32 | 1 | llama.context_length | 131072 |
|
| 29 |
+
| 19 | UINT32 | 1 | llama.embedding_length | 4096 |
|
| 30 |
+
| 20 | UINT32 | 1 | llama.feed_forward_length | 14336 |
|
| 31 |
+
| 21 | UINT32 | 1 | llama.attention.head_count | 32 |
|
| 32 |
+
| 22 | UINT32 | 1 | llama.attention.head_count_kv | 8 |
|
| 33 |
+
| 23 | FLOAT32 | 1 | llama.rope.freq_base | 500000.0 |
|
| 34 |
+
| 24 | FLOAT32 | 1 | llama.attention.layer_norm_rms_epsilon | 1e-05 |
|
| 35 |
+
| 25 | UINT32 | 1 | llama.attention.key_length | 128 |
|
| 36 |
+
| 26 | UINT32 | 1 | llama.attention.value_length | 128 |
|
| 37 |
+
| 27 | UINT32 | 1 | llama.vocab_size | 128256 |
|
| 38 |
+
| 28 | UINT32 | 1 | llama.rope.dimension_count | 128 |
|
| 39 |
+
| 29 | STRING | 1 | tokenizer.ggml.model | `gpt2` |
|
| 40 |
+
| 30 | STRING | 1 | tokenizer.ggml.pre | `llama-bpe` |
|
| 41 |
+
| 31 | [STRING] | 128256 | tokenizer.ggml.tokens | [ `!`, `"`, `#`, `$`, `%`, ... ] |
|
| 42 |
+
| 32 | [INT32] | 128256 | tokenizer.ggml.token_type | [ 1, 1, 1, 1, 1, 1, 1, ... ] |
|
| 43 |
+
| 33 | [STRING] | 280147 | tokenizer.ggml.merges | [ `Ġ Ġ`, `Ġ ĠĠĠ`, `ĠĠ ĠĠ`, `ĠĠĠ Ġ`, `i n`, ... ] |
|
| 44 |
+
| 34 | UINT32 | 1 | tokenizer.ggml.bos_token_id | 128000 |
|
| 45 |
+
| 35 | UINT32 | 1 | tokenizer.ggml.eos_token_id | 128009 |
|
| 46 |
+
| 36 | UINT32 | 1 | tokenizer.ggml.padding_token_id | 128009 |
|
| 47 |
+
| 37 | STRING | 1 | tokenizer.chat_template | `{{ '<|begin_of_text|>' }}{% if`...`d|>' }}{% endif %}{% endfor %}` |
|
| 48 |
+
| 38 | UINT32 | 1 | general.quantization_version | 2 |
|
| 49 |
+
| 39 | UINT32 | 1 | general.file_type | 11 |
|
| 50 |
+
| 40 | STRING | 1 | quantize.imatrix.file | `./imatrix/imatrix-Watt-Tool-8B-small.dat` |
|
| 51 |
+
| 41 | STRING | 1 | quantize.imatrix.dataset | `../../datasets/imatrix/calibration_eur_small.txt` |
|
| 52 |
+
| 42 | INT32 | 1 | quantize.imatrix.entries_count | 225 |
|
| 53 |
+
| 43 | INT32 | 1 | quantize.imatrix.chunks_count | 962 |
|
| 54 |
+
|
| 55 |
+
## Tensors Overview ~8B Elements
|
| 56 |
+
|
| 57 |
+
Total number of elements in all tensors: 8030261312 Elements
|
| 58 |
+
|
| 59 |
+
- [Watt-Tool-8B-Q3\_K\_S.gguf - GGUF Internal File Dump](#watt-tool-8b-q3_k_sgguf---gguf-internal-file-dump)
|
| 60 |
+
- [Key Value Metadata Store](#key-value-metadata-store)
|
| 61 |
+
- [Tensors Overview ~8B Elements](#tensors-overview-8b-elements)
|
| 62 |
+
- [Tensor Data Offset](#tensor-data-offset)
|
| 63 |
+
- [Base Tensor Group : ~1B Elements](#base-tensor-group--1b-elements)
|
| 64 |
+
- [Block 0 Tensor Group : ~218M Elements](#block-0-tensor-group--218m-elements)
|
| 65 |
+
- [Block 1 Tensor Group : ~218M Elements](#block-1-tensor-group--218m-elements)
|
| 66 |
+
- [Block 2 Tensor Group : ~218M Elements](#block-2-tensor-group--218m-elements)
|
| 67 |
+
- [Block 3 Tensor Group : ~218M Elements](#block-3-tensor-group--218m-elements)
|
| 68 |
+
- [Block 4 Tensor Group : ~218M Elements](#block-4-tensor-group--218m-elements)
|
| 69 |
+
- [Block 5 Tensor Group : ~218M Elements](#block-5-tensor-group--218m-elements)
|
| 70 |
+
- [Block 6 Tensor Group : ~218M Elements](#block-6-tensor-group--218m-elements)
|
| 71 |
+
- [Block 7 Tensor Group : ~218M Elements](#block-7-tensor-group--218m-elements)
|
| 72 |
+
- [Block 8 Tensor Group : ~218M Elements](#block-8-tensor-group--218m-elements)
|
| 73 |
+
- [Block 9 Tensor Group : ~218M Elements](#block-9-tensor-group--218m-elements)
|
| 74 |
+
- [Block 10 Tensor Group : ~218M Elements](#block-10-tensor-group--218m-elements)
|
| 75 |
+
- [Block 11 Tensor Group : ~218M Elements](#block-11-tensor-group--218m-elements)
|
| 76 |
+
- [Block 12 Tensor Group : ~218M Elements](#block-12-tensor-group--218m-elements)
|
| 77 |
+
- [Block 13 Tensor Group : ~218M Elements](#block-13-tensor-group--218m-elements)
|
| 78 |
+
- [Block 14 Tensor Group : ~218M Elements](#block-14-tensor-group--218m-elements)
|
| 79 |
+
- [Block 15 Tensor Group : ~218M Elements](#block-15-tensor-group--218m-elements)
|
| 80 |
+
- [Block 16 Tensor Group : ~218M Elements](#block-16-tensor-group--218m-elements)
|
| 81 |
+
- [Block 17 Tensor Group : ~218M Elements](#block-17-tensor-group--218m-elements)
|
| 82 |
+
- [Block 18 Tensor Group : ~218M Elements](#block-18-tensor-group--218m-elements)
|
| 83 |
+
- [Block 19 Tensor Group : ~218M Elements](#block-19-tensor-group--218m-elements)
|
| 84 |
+
- [Block 20 Tensor Group : ~218M Elements](#block-20-tensor-group--218m-elements)
|
| 85 |
+
- [Block 21 Tensor Group : ~218M Elements](#block-21-tensor-group--218m-elements)
|
| 86 |
+
- [Block 22 Tensor Group : ~218M Elements](#block-22-tensor-group--218m-elements)
|
| 87 |
+
- [Block 23 Tensor Group : ~218M Elements](#block-23-tensor-group--218m-elements)
|
| 88 |
+
- [Block 24 Tensor Group : ~218M Elements](#block-24-tensor-group--218m-elements)
|
| 89 |
+
- [Block 25 Tensor Group : ~218M Elements](#block-25-tensor-group--218m-elements)
|
| 90 |
+
- [Block 26 Tensor Group : ~218M Elements](#block-26-tensor-group--218m-elements)
|
| 91 |
+
- [Block 27 Tensor Group : ~218M Elements](#block-27-tensor-group--218m-elements)
|
| 92 |
+
- [Block 28 Tensor Group : ~218M Elements](#block-28-tensor-group--218m-elements)
|
| 93 |
+
- [Block 29 Tensor Group : ~218M Elements](#block-29-tensor-group--218m-elements)
|
| 94 |
+
- [Block 30 Tensor Group : ~218M Elements](#block-30-tensor-group--218m-elements)
|
| 95 |
+
- [Block 31 Tensor Group : ~218M Elements](#block-31-tensor-group--218m-elements)
|
| 96 |
+
|
| 97 |
+
### Tensor Data Offset
|
| 98 |
+
|
| 99 |
+
This table contains the offset and data segment relative to start of file
|
| 100 |
+
|
| 101 |
+
| T_ID | Tensor Layer Name | Data Offset (B) | Data Size (B) |
|
| 102 |
+
|-----:|:--------------------------|-----------------:|-----------------:|
|
| 103 |
+
| 0 | output.weight | 0x779620 | 0xd746000 |
|
| 104 |
+
| 1 | output_norm.weight | 0xdebf620 | 0x4000 |
|
| 105 |
+
| 2 | rope_freqs.weight | 0xdec3620 | 0x100 |
|
| 106 |
+
| 3 | token_embd.weight | 0xdec3720 | 0xa464000 |
|
| 107 |
+
| 4 | blk.0.attn_k.weight | 0x18327720 | 0x150000 |
|
| 108 |
+
| 5 | blk.0.attn_norm.weight | 0x18477720 | 0x4000 |
|
| 109 |
+
| 6 | blk.0.attn_output.weight | 0x1847b720 | 0x6e0000 |
|
| 110 |
+
| 7 | blk.0.attn_q.weight | 0x18b5b720 | 0x540000 |
|
| 111 |
+
| 8 | blk.0.attn_v.weight | 0x1909b720 | 0x1b8000 |
|
| 112 |
+
| 9 | blk.0.ffn_down.weight | 0x19253720 | 0x1810000 |
|
| 113 |
+
| 10 | blk.0.ffn_gate.weight | 0x1aa63720 | 0x1260000 |
|
| 114 |
+
| 11 | blk.0.ffn_norm.weight | 0x1bcc3720 | 0x4000 |
|
| 115 |
+
| 12 | blk.0.ffn_up.weight | 0x1bcc7720 | 0x1260000 |
|
| 116 |
+
| 13 | blk.1.attn_k.weight | 0x1cf27720 | 0x150000 |
|
| 117 |
+
| 14 | blk.1.attn_norm.weight | 0x1d077720 | 0x4000 |
|
| 118 |
+
| 15 | blk.1.attn_output.weight | 0x1d07b720 | 0x6e0000 |
|
| 119 |
+
| 16 | blk.1.attn_q.weight | 0x1d75b720 | 0x540000 |
|
| 120 |
+
| 17 | blk.1.attn_v.weight | 0x1dc9b720 | 0x1b8000 |
|
| 121 |
+
| 18 | blk.1.ffn_down.weight | 0x1de53720 | 0x1f80000 |
|
| 122 |
+
| 19 | blk.1.ffn_gate.weight | 0x1fdd3720 | 0x1260000 |
|
| 123 |
+
| 20 | blk.1.ffn_norm.weight | 0x21033720 | 0x4000 |
|
| 124 |
+
| 21 | blk.1.ffn_up.weight | 0x21037720 | 0x1260000 |
|
| 125 |
+
| 22 | blk.2.attn_k.weight | 0x22297720 | 0x150000 |
|
| 126 |
+
| 23 | blk.2.attn_norm.weight | 0x223e7720 | 0x4000 |
|
| 127 |
+
| 24 | blk.2.attn_output.weight | 0x223eb720 | 0x6e0000 |
|
| 128 |
+
| 25 | blk.2.attn_q.weight | 0x22acb720 | 0x540000 |
|
| 129 |
+
| 26 | blk.2.attn_v.weight | 0x2300b720 | 0x1b8000 |
|
| 130 |
+
| 27 | blk.2.ffn_down.weight | 0x231c3720 | 0x1810000 |
|
| 131 |
+
| 28 | blk.2.ffn_gate.weight | 0x249d3720 | 0x1260000 |
|
| 132 |
+
| 29 | blk.2.ffn_norm.weight | 0x25c33720 | 0x4000 |
|
| 133 |
+
| 30 | blk.2.ffn_up.weight | 0x25c37720 | 0x1260000 |
|
| 134 |
+
| 31 | blk.3.attn_k.weight | 0x26e97720 | 0x150000 |
|
| 135 |
+
| 32 | blk.3.attn_norm.weight | 0x26fe7720 | 0x4000 |
|
| 136 |
+
| 33 | blk.3.attn_output.weight | 0x26feb720 | 0x6e0000 |
|
| 137 |
+
| 34 | blk.3.attn_q.weight | 0x276cb720 | 0x540000 |
|
| 138 |
+
| 35 | blk.3.attn_v.weight | 0x27c0b720 | 0x1b8000 |
|
| 139 |
+
| 36 | blk.3.ffn_down.weight | 0x27dc3720 | 0x1810000 |
|
| 140 |
+
| 37 | blk.3.ffn_gate.weight | 0x295d3720 | 0x1260000 |
|
| 141 |
+
| 38 | blk.3.ffn_norm.weight | 0x2a833720 | 0x4000 |
|
| 142 |
+
| 39 | blk.3.ffn_up.weight | 0x2a837720 | 0x1260000 |
|
| 143 |
+
| 40 | blk.4.attn_k.weight | 0x2ba97720 | 0x150000 |
|
| 144 |
+
| 41 | blk.4.attn_norm.weight | 0x2bbe7720 | 0x4000 |
|
| 145 |
+
| 42 | blk.4.attn_output.weight | 0x2bbeb720 | 0x6e0000 |
|
| 146 |
+
| 43 | blk.4.attn_q.weight | 0x2c2cb720 | 0x540000 |
|
| 147 |
+
| 44 | blk.4.attn_v.weight | 0x2c80b720 | 0x1b8000 |
|
| 148 |
+
| 45 | blk.4.ffn_down.weight | 0x2c9c3720 | 0x1810000 |
|
| 149 |
+
| 46 | blk.4.ffn_gate.weight | 0x2e1d3720 | 0x1260000 |
|
| 150 |
+
| 47 | blk.4.ffn_norm.weight | 0x2f433720 | 0x4000 |
|
| 151 |
+
| 48 | blk.4.ffn_up.weight | 0x2f437720 | 0x1260000 |
|
| 152 |
+
| 49 | blk.5.attn_k.weight | 0x30697720 | 0x150000 |
|
| 153 |
+
| 50 | blk.5.attn_norm.weight | 0x307e7720 | 0x4000 |
|
| 154 |
+
| 51 | blk.5.attn_output.weight | 0x307eb720 | 0x6e0000 |
|
| 155 |
+
| 52 | blk.5.attn_q.weight | 0x30ecb720 | 0x540000 |
|
| 156 |
+
| 53 | blk.5.attn_v.weight | 0x3140b720 | 0x1b8000 |
|
| 157 |
+
| 54 | blk.5.ffn_down.weight | 0x315c3720 | 0x1810000 |
|
| 158 |
+
| 55 | blk.5.ffn_gate.weight | 0x32dd3720 | 0x1260000 |
|
| 159 |
+
| 56 | blk.5.ffn_norm.weight | 0x34033720 | 0x4000 |
|
| 160 |
+
| 57 | blk.5.ffn_up.weight | 0x34037720 | 0x1260000 |
|
| 161 |
+
| 58 | blk.6.attn_k.weight | 0x35297720 | 0x150000 |
|
| 162 |
+
| 59 | blk.6.attn_norm.weight | 0x353e7720 | 0x4000 |
|
| 163 |
+
| 60 | blk.6.attn_output.weight | 0x353eb720 | 0x6e0000 |
|
| 164 |
+
| 61 | blk.6.attn_q.weight | 0x35acb720 | 0x540000 |
|
| 165 |
+
| 62 | blk.6.attn_v.weight | 0x3600b720 | 0x1b8000 |
|
| 166 |
+
| 63 | blk.6.ffn_down.weight | 0x361c3720 | 0x1810000 |
|
| 167 |
+
| 64 | blk.6.ffn_gate.weight | 0x379d3720 | 0x1260000 |
|
| 168 |
+
| 65 | blk.6.ffn_norm.weight | 0x38c33720 | 0x4000 |
|
| 169 |
+
| 66 | blk.6.ffn_up.weight | 0x38c37720 | 0x1260000 |
|
| 170 |
+
| 67 | blk.7.attn_k.weight | 0x39e97720 | 0x150000 |
|
| 171 |
+
| 68 | blk.7.attn_norm.weight | 0x39fe7720 | 0x4000 |
|
| 172 |
+
| 69 | blk.7.attn_output.weight | 0x39feb720 | 0x6e0000 |
|
| 173 |
+
| 70 | blk.7.attn_q.weight | 0x3a6cb720 | 0x540000 |
|
| 174 |
+
| 71 | blk.7.attn_v.weight | 0x3ac0b720 | 0x1b8000 |
|
| 175 |
+
| 72 | blk.7.ffn_down.weight | 0x3adc3720 | 0x1810000 |
|
| 176 |
+
| 73 | blk.7.ffn_gate.weight | 0x3c5d3720 | 0x1260000 |
|
| 177 |
+
| 74 | blk.7.ffn_norm.weight | 0x3d833720 | 0x4000 |
|
| 178 |
+
| 75 | blk.7.ffn_up.weight | 0x3d837720 | 0x1260000 |
|
| 179 |
+
| 76 | blk.8.attn_k.weight | 0x3ea97720 | 0x150000 |
|
| 180 |
+
| 77 | blk.8.attn_norm.weight | 0x3ebe7720 | 0x4000 |
|
| 181 |
+
| 78 | blk.8.attn_output.weight | 0x3ebeb720 | 0x6e0000 |
|
| 182 |
+
| 79 | blk.8.attn_q.weight | 0x3f2cb720 | 0x540000 |
|
| 183 |
+
| 80 | blk.8.attn_v.weight | 0x3f80b720 | 0x1b8000 |
|
| 184 |
+
| 81 | blk.8.ffn_down.weight | 0x3f9c3720 | 0x1810000 |
|
| 185 |
+
| 82 | blk.8.ffn_gate.weight | 0x411d3720 | 0x1260000 |
|
| 186 |
+
| 83 | blk.8.ffn_norm.weight | 0x42433720 | 0x4000 |
|
| 187 |
+
| 84 | blk.8.ffn_up.weight | 0x42437720 | 0x1260000 |
|
| 188 |
+
| 85 | blk.9.attn_k.weight | 0x43697720 | 0x150000 |
|
| 189 |
+
| 86 | blk.9.attn_norm.weight | 0x437e7720 | 0x4000 |
|
| 190 |
+
| 87 | blk.9.attn_output.weight | 0x437eb720 | 0x6e0000 |
|
| 191 |
+
| 88 | blk.9.attn_q.weight | 0x43ecb720 | 0x540000 |
|
| 192 |
+
| 89 | blk.9.attn_v.weight | 0x4440b720 | 0x1b8000 |
|
| 193 |
+
| 90 | blk.9.ffn_down.weight | 0x445c3720 | 0x1810000 |
|
| 194 |
+
| 91 | blk.9.ffn_gate.weight | 0x45dd3720 | 0x1260000 |
|
| 195 |
+
| 92 | blk.9.ffn_norm.weight | 0x47033720 | 0x4000 |
|
| 196 |
+
| 93 | blk.9.ffn_up.weight | 0x47037720 | 0x1260000 |
|
| 197 |
+
| 94 | blk.10.attn_k.weight | 0x48297720 | 0x150000 |
|
| 198 |
+
| 95 | blk.10.attn_norm.weight | 0x483e7720 | 0x4000 |
|
| 199 |
+
| 96 | blk.10.attn_output.weight | 0x483eb720 | 0x6e0000 |
|
| 200 |
+
| 97 | blk.10.attn_q.weight | 0x48acb720 | 0x540000 |
|
| 201 |
+
| 98 | blk.10.attn_v.weight | 0x4900b720 | 0x1b8000 |
|
| 202 |
+
| 99 | blk.10.ffn_down.weight | 0x491c3720 | 0x1810000 |
|
| 203 |
+
| 100 | blk.10.ffn_gate.weight | 0x4a9d3720 | 0x1260000 |
|
| 204 |
+
| 101 | blk.10.ffn_norm.weight | 0x4bc33720 | 0x4000 |
|
| 205 |
+
| 102 | blk.10.ffn_up.weight | 0x4bc37720 | 0x1260000 |
|
| 206 |
+
| 103 | blk.11.attn_k.weight | 0x4ce97720 | 0x150000 |
|
| 207 |
+
| 104 | blk.11.attn_norm.weight | 0x4cfe7720 | 0x4000 |
|
| 208 |
+
| 105 | blk.11.attn_output.weight | 0x4cfeb720 | 0x6e0000 |
|
| 209 |
+
| 106 | blk.11.attn_q.weight | 0x4d6cb720 | 0x540000 |
|
| 210 |
+
| 107 | blk.11.attn_v.weight | 0x4dc0b720 | 0x1b8000 |
|
| 211 |
+
| 108 | blk.11.ffn_down.weight | 0x4ddc3720 | 0x1810000 |
|
| 212 |
+
| 109 | blk.11.ffn_gate.weight | 0x4f5d3720 | 0x1260000 |
|
| 213 |
+
| 110 | blk.11.ffn_norm.weight | 0x50833720 | 0x4000 |
|
| 214 |
+
| 111 | blk.11.ffn_up.weight | 0x50837720 | 0x1260000 |
|
| 215 |
+
| 112 | blk.12.attn_k.weight | 0x51a97720 | 0x150000 |
|
| 216 |
+
| 113 | blk.12.attn_norm.weight | 0x51be7720 | 0x4000 |
|
| 217 |
+
| 114 | blk.12.attn_output.weight | 0x51beb720 | 0x6e0000 |
|
| 218 |
+
| 115 | blk.12.attn_q.weight | 0x522cb720 | 0x540000 |
|
| 219 |
+
| 116 | blk.12.attn_v.weight | 0x5280b720 | 0x1b8000 |
|
| 220 |
+
| 117 | blk.12.ffn_down.weight | 0x529c3720 | 0x1810000 |
|
| 221 |
+
| 118 | blk.12.ffn_gate.weight | 0x541d3720 | 0x1260000 |
|
| 222 |
+
| 119 | blk.12.ffn_norm.weight | 0x55433720 | 0x4000 |
|
| 223 |
+
| 120 | blk.12.ffn_up.weight | 0x55437720 | 0x1260000 |
|
| 224 |
+
| 121 | blk.13.attn_k.weight | 0x56697720 | 0x1b8000 |
|
| 225 |
+
| 122 | blk.13.attn_norm.weight | 0x5684f720 | 0x4000 |
|
| 226 |
+
| 123 | blk.13.attn_output.weight | 0x56853720 | 0x6e0000 |
|
| 227 |
+
| 124 | blk.13.attn_q.weight | 0x56f33720 | 0x6e0000 |
|
| 228 |
+
| 125 | blk.13.attn_v.weight | 0x57613720 | 0x1b8000 |
|
| 229 |
+
| 126 | blk.13.ffn_down.weight | 0x577cb720 | 0x1810000 |
|
| 230 |
+
| 127 | blk.13.ffn_gate.weight | 0x58fdb720 | 0x1260000 |
|
| 231 |
+
| 128 | blk.13.ffn_norm.weight | 0x5a23b720 | 0x4000 |
|
| 232 |
+
| 129 | blk.13.ffn_up.weight | 0x5a23f720 | 0x1260000 |
|
| 233 |
+
| 130 | blk.14.attn_k.weight | 0x5b49f720 | 0x1b8000 |
|
| 234 |
+
| 131 | blk.14.attn_norm.weight | 0x5b657720 | 0x4000 |
|
| 235 |
+
| 132 | blk.14.attn_output.weight | 0x5b65b720 | 0x6e0000 |
|
| 236 |
+
| 133 | blk.14.attn_q.weight | 0x5bd3b720 | 0x6e0000 |
|
| 237 |
+
| 134 | blk.14.attn_v.weight | 0x5c41b720 | 0x1b8000 |
|
| 238 |
+
| 135 | blk.14.ffn_down.weight | 0x5c5d3720 | 0x1810000 |
|
| 239 |
+
| 136 | blk.14.ffn_gate.weight | 0x5dde3720 | 0x1260000 |
|
| 240 |
+
| 137 | blk.14.ffn_norm.weight | 0x5f043720 | 0x4000 |
|
| 241 |
+
| 138 | blk.14.ffn_up.weight | 0x5f047720 | 0x1260000 |
|
| 242 |
+
| 139 | blk.15.attn_k.weight | 0x602a7720 | 0x150000 |
|
| 243 |
+
| 140 | blk.15.attn_norm.weight | 0x603f7720 | 0x4000 |
|
| 244 |
+
| 141 | blk.15.attn_output.weight | 0x603fb720 | 0x6e0000 |
|
| 245 |
+
| 142 | blk.15.attn_q.weight | 0x60adb720 | 0x540000 |
|
| 246 |
+
| 143 | blk.15.attn_v.weight | 0x6101b720 | 0x1b8000 |
|
| 247 |
+
| 144 | blk.15.ffn_down.weight | 0x611d3720 | 0x1810000 |
|
| 248 |
+
| 145 | blk.15.ffn_gate.weight | 0x629e3720 | 0x1260000 |
|
| 249 |
+
| 146 | blk.15.ffn_norm.weight | 0x63c43720 | 0x4000 |
|
| 250 |
+
| 147 | blk.15.ffn_up.weight | 0x63c47720 | 0x1260000 |
|
| 251 |
+
| 148 | blk.16.attn_k.weight | 0x64ea7720 | 0x1b8000 |
|
| 252 |
+
| 149 | blk.16.attn_norm.weight | 0x6505f720 | 0x4000 |
|
| 253 |
+
| 150 | blk.16.attn_output.weight | 0x65063720 | 0x6e0000 |
|
| 254 |
+
| 151 | blk.16.attn_q.weight | 0x65743720 | 0x6e0000 |
|
| 255 |
+
| 152 | blk.16.attn_v.weight | 0x65e23720 | 0x1b8000 |
|
| 256 |
+
| 153 | blk.16.ffn_down.weight | 0x65fdb720 | 0x1810000 |
|
| 257 |
+
| 154 | blk.16.ffn_gate.weight | 0x677eb720 | 0x1810000 |
|
| 258 |
+
| 155 | blk.16.ffn_norm.weight | 0x68ffb720 | 0x4000 |
|
| 259 |
+
| 156 | blk.16.ffn_up.weight | 0x68fff720 | 0x1810000 |
|
| 260 |
+
| 157 | blk.17.attn_k.weight | 0x6a80f720 | 0x150000 |
|
| 261 |
+
| 158 | blk.17.attn_norm.weight | 0x6a95f720 | 0x4000 |
|
| 262 |
+
| 159 | blk.17.attn_output.weight | 0x6a963720 | 0x6e0000 |
|
| 263 |
+
| 160 | blk.17.attn_q.weight | 0x6b043720 | 0x540000 |
|
| 264 |
+
| 161 | blk.17.attn_v.weight | 0x6b583720 | 0x1b8000 |
|
| 265 |
+
| 162 | blk.17.ffn_down.weight | 0x6b73b720 | 0x1f80000 |
|
| 266 |
+
| 163 | blk.17.ffn_gate.weight | 0x6d6bb720 | 0x1810000 |
|
| 267 |
+
| 164 | blk.17.ffn_norm.weight | 0x6eecb720 | 0x4000 |
|
| 268 |
+
| 165 | blk.17.ffn_up.weight | 0x6eecf720 | 0x1810000 |
|
| 269 |
+
| 166 | blk.18.attn_k.weight | 0x706df720 | 0x1b8000 |
|
| 270 |
+
| 167 | blk.18.attn_norm.weight | 0x70897720 | 0x4000 |
|
| 271 |
+
| 168 | blk.18.attn_output.weight | 0x7089b720 | 0x6e0000 |
|
| 272 |
+
| 169 | blk.18.attn_q.weight | 0x70f7b720 | 0x6e0000 |
|
| 273 |
+
| 170 | blk.18.attn_v.weight | 0x7165b720 | 0x1b8000 |
|
| 274 |
+
| 171 | blk.18.ffn_down.weight | 0x71813720 | 0x1f80000 |
|
| 275 |
+
| 172 | blk.18.ffn_gate.weight | 0x73793720 | 0x1810000 |
|
| 276 |
+
| 173 | blk.18.ffn_norm.weight | 0x74fa3720 | 0x4000 |
|
| 277 |
+
| 174 | blk.18.ffn_up.weight | 0x74fa7720 | 0x1810000 |
|
| 278 |
+
| 175 | blk.19.attn_k.weight | 0x767b7720 | 0x1b8000 |
|
| 279 |
+
| 176 | blk.19.attn_norm.weight | 0x7696f720 | 0x4000 |
|
| 280 |
+
| 177 | blk.19.attn_output.weight | 0x76973720 | 0x6e0000 |
|
| 281 |
+
| 178 | blk.19.attn_q.weight | 0x77053720 | 0x6e0000 |
|
| 282 |
+
| 179 | blk.19.attn_v.weight | 0x77733720 | 0x1b8000 |
|
| 283 |
+
| 180 | blk.19.ffn_down.weight | 0x778eb720 | 0x1f80000 |
|
| 284 |
+
| 181 | blk.19.ffn_gate.weight | 0x7986b720 | 0x1810000 |
|
| 285 |
+
| 182 | blk.19.ffn_norm.weight | 0x7b07b720 | 0x4000 |
|
| 286 |
+
| 183 | blk.19.ffn_up.weight | 0x7b07f720 | 0x1810000 |
|
| 287 |
+
| 184 | blk.20.attn_k.weight | 0x7c88f720 | 0x1b8000 |
|
| 288 |
+
| 185 | blk.20.attn_norm.weight | 0x7ca47720 | 0x4000 |
|
| 289 |
+
| 186 | blk.20.attn_output.weight | 0x7ca4b720 | 0x6e0000 |
|
| 290 |
+
| 187 | blk.20.attn_q.weight | 0x7d12b720 | 0x6e0000 |
|
| 291 |
+
| 188 | blk.20.attn_v.weight | 0x7d80b720 | 0x1b8000 |
|
| 292 |
+
| 189 | blk.20.ffn_down.weight | 0x7d9c3720 | 0x1f80000 |
|
| 293 |
+
| 190 | blk.20.ffn_gate.weight | 0x7f943720 | 0x1810000 |
|
| 294 |
+
| 191 | blk.20.ffn_norm.weight | 0x81153720 | 0x4000 |
|
| 295 |
+
| 192 | blk.20.ffn_up.weight | 0x81157720 | 0x1810000 |
|
| 296 |
+
| 193 | blk.21.attn_k.weight | 0x82967720 | 0x1b8000 |
|
| 297 |
+
| 194 | blk.21.attn_norm.weight | 0x82b1f720 | 0x4000 |
|
| 298 |
+
| 195 | blk.21.attn_output.weight | 0x82b23720 | 0x6e0000 |
|
| 299 |
+
| 196 | blk.21.attn_q.weight | 0x83203720 | 0x6e0000 |
|
| 300 |
+
| 197 | blk.21.attn_v.weight | 0x838e3720 | 0x1b8000 |
|
| 301 |
+
| 198 | blk.21.ffn_down.weight | 0x83a9b720 | 0x1f80000 |
|
| 302 |
+
| 199 | blk.21.ffn_gate.weight | 0x85a1b720 | 0x1810000 |
|
| 303 |
+
| 200 | blk.21.ffn_norm.weight | 0x8722b720 | 0x4000 |
|
| 304 |
+
| 201 | blk.21.ffn_up.weight | 0x8722f720 | 0x1810000 |
|
| 305 |
+
| 202 | blk.22.attn_k.weight | 0x88a3f720 | 0x1b8000 |
|
| 306 |
+
| 203 | blk.22.attn_norm.weight | 0x88bf7720 | 0x4000 |
|
| 307 |
+
| 204 | blk.22.attn_output.weight | 0x88bfb720 | 0x6e0000 |
|
| 308 |
+
| 205 | blk.22.attn_q.weight | 0x892db720 | 0x6e0000 |
|
| 309 |
+
| 206 | blk.22.attn_v.weight | 0x899bb720 | 0x1b8000 |
|
| 310 |
+
| 207 | blk.22.ffn_down.weight | 0x89b73720 | 0x1f80000 |
|
| 311 |
+
| 208 | blk.22.ffn_gate.weight | 0x8baf3720 | 0x1810000 |
|
| 312 |
+
| 209 | blk.22.ffn_norm.weight | 0x8d303720 | 0x4000 |
|
| 313 |
+
| 210 | blk.22.ffn_up.weight | 0x8d307720 | 0x1810000 |
|
| 314 |
+
| 211 | blk.23.attn_k.weight | 0x8eb17720 | 0x1b8000 |
|
| 315 |
+
| 212 | blk.23.attn_norm.weight | 0x8eccf720 | 0x4000 |
|
| 316 |
+
| 213 | blk.23.attn_output.weight | 0x8ecd3720 | 0x6e0000 |
|
| 317 |
+
| 214 | blk.23.attn_q.weight | 0x8f3b3720 | 0x6e0000 |
|
| 318 |
+
| 215 | blk.23.attn_v.weight | 0x8fa93720 | 0x1b8000 |
|
| 319 |
+
| 216 | blk.23.ffn_down.weight | 0x8fc4b720 | 0x1f80000 |
|
| 320 |
+
| 217 | blk.23.ffn_gate.weight | 0x91bcb720 | 0x1810000 |
|
| 321 |
+
| 218 | blk.23.ffn_norm.weight | 0x933db720 | 0x4000 |
|
| 322 |
+
| 219 | blk.23.ffn_up.weight | 0x933df720 | 0x1810000 |
|
| 323 |
+
| 220 | blk.24.attn_k.weight | 0x94bef720 | 0x1b8000 |
|
| 324 |
+
| 221 | blk.24.attn_norm.weight | 0x94da7720 | 0x4000 |
|
| 325 |
+
| 222 | blk.24.attn_output.weight | 0x94dab720 | 0x6e0000 |
|
| 326 |
+
| 223 | blk.24.attn_q.weight | 0x9548b720 | 0x6e0000 |
|
| 327 |
+
| 224 | blk.24.attn_v.weight | 0x95b6b720 | 0x1b8000 |
|
| 328 |
+
| 225 | blk.24.ffn_down.weight | 0x95d23720 | 0x1f80000 |
|
| 329 |
+
| 226 | blk.24.ffn_gate.weight | 0x97ca3720 | 0x1810000 |
|
| 330 |
+
| 227 | blk.24.ffn_norm.weight | 0x994b3720 | 0x4000 |
|
| 331 |
+
| 228 | blk.24.ffn_up.weight | 0x994b7720 | 0x1810000 |
|
| 332 |
+
| 229 | blk.25.attn_k.weight | 0x9acc7720 | 0x1b8000 |
|
| 333 |
+
| 230 | blk.25.attn_norm.weight | 0x9ae7f720 | 0x4000 |
|
| 334 |
+
| 231 | blk.25.attn_output.weight | 0x9ae83720 | 0x6e0000 |
|
| 335 |
+
| 232 | blk.25.attn_q.weight | 0x9b563720 | 0x6e0000 |
|
| 336 |
+
| 233 | blk.25.attn_v.weight | 0x9bc43720 | 0x1b8000 |
|
| 337 |
+
| 234 | blk.25.ffn_down.weight | 0x9bdfb720 | 0x1f80000 |
|
| 338 |
+
| 235 | blk.25.ffn_gate.weight | 0x9dd7b720 | 0x1810000 |
|
| 339 |
+
| 236 | blk.25.ffn_norm.weight | 0x9f58b720 | 0x4000 |
|
| 340 |
+
| 237 | blk.25.ffn_up.weight | 0x9f58f720 | 0x1810000 |
|
| 341 |
+
| 238 | blk.26.attn_k.weight | 0xa0d9f720 | 0x1b8000 |
|
| 342 |
+
| 239 | blk.26.attn_norm.weight | 0xa0f57720 | 0x4000 |
|
| 343 |
+
| 240 | blk.26.attn_output.weight | 0xa0f5b720 | 0x6e0000 |
|
| 344 |
+
| 241 | blk.26.attn_q.weight | 0xa163b720 | 0x6e0000 |
|
| 345 |
+
| 242 | blk.26.attn_v.weight | 0xa1d1b720 | 0x1b8000 |
|
| 346 |
+
| 243 | blk.26.ffn_down.weight | 0xa1ed3720 | 0x1f80000 |
|
| 347 |
+
| 244 | blk.26.ffn_gate.weight | 0xa3e53720 | 0x1810000 |
|
| 348 |
+
| 245 | blk.26.ffn_norm.weight | 0xa5663720 | 0x4000 |
|
| 349 |
+
| 246 | blk.26.ffn_up.weight | 0xa5667720 | 0x1810000 |
|
| 350 |
+
| 247 | blk.27.attn_k.weight | 0xa6e77720 | 0x1b8000 |
|
| 351 |
+
| 248 | blk.27.attn_norm.weight | 0xa702f720 | 0x4000 |
|
| 352 |
+
| 249 | blk.27.attn_output.weight | 0xa7033720 | 0x6e0000 |
|
| 353 |
+
| 250 | blk.27.attn_q.weight | 0xa7713720 | 0x6e0000 |
|
| 354 |
+
| 251 | blk.27.attn_v.weight | 0xa7df3720 | 0x1b8000 |
|
| 355 |
+
| 252 | blk.27.ffn_down.weight | 0xa7fab720 | 0x1f80000 |
|
| 356 |
+
| 253 | blk.27.ffn_gate.weight | 0xa9f2b720 | 0x1810000 |
|
| 357 |
+
| 254 | blk.27.ffn_norm.weight | 0xab73b720 | 0x4000 |
|
| 358 |
+
| 255 | blk.27.ffn_up.weight | 0xab73f720 | 0x1810000 |
|
| 359 |
+
| 256 | blk.28.attn_k.weight | 0xacf4f720 | 0x1b8000 |
|
| 360 |
+
| 257 | blk.28.attn_norm.weight | 0xad107720 | 0x4000 |
|
| 361 |
+
| 258 | blk.28.attn_output.weight | 0xad10b720 | 0x6e0000 |
|
| 362 |
+
| 259 | blk.28.attn_q.weight | 0xad7eb720 | 0x6e0000 |
|
| 363 |
+
| 260 | blk.28.attn_v.weight | 0xadecb720 | 0x1b8000 |
|
| 364 |
+
| 261 | blk.28.ffn_down.weight | 0xae083720 | 0x1f80000 |
|
| 365 |
+
| 262 | blk.28.ffn_gate.weight | 0xb0003720 | 0x1810000 |
|
| 366 |
+
| 263 | blk.28.ffn_norm.weight | 0xb1813720 | 0x4000 |
|
| 367 |
+
| 264 | blk.28.ffn_up.weight | 0xb1817720 | 0x1810000 |
|
| 368 |
+
| 265 | blk.29.attn_k.weight | 0xb3027720 | 0x1b8000 |
|
| 369 |
+
| 266 | blk.29.attn_norm.weight | 0xb31df720 | 0x4000 |
|
| 370 |
+
| 267 | blk.29.attn_output.weight | 0xb31e3720 | 0x6e0000 |
|
| 371 |
+
| 268 | blk.29.attn_q.weight | 0xb38c3720 | 0x6e0000 |
|
| 372 |
+
| 269 | blk.29.attn_v.weight | 0xb3fa3720 | 0x1b8000 |
|
| 373 |
+
| 270 | blk.29.ffn_down.weight | 0xb415b720 | 0x1f80000 |
|
| 374 |
+
| 271 | blk.29.ffn_gate.weight | 0xb60db720 | 0x1810000 |
|
| 375 |
+
| 272 | blk.29.ffn_norm.weight | 0xb78eb720 | 0x4000 |
|
| 376 |
+
| 273 | blk.29.ffn_up.weight | 0xb78ef720 | 0x1810000 |
|
| 377 |
+
| 274 | blk.30.attn_k.weight | 0xb90ff720 | 0x1b8000 |
|
| 378 |
+
| 275 | blk.30.attn_norm.weight | 0xb92b7720 | 0x4000 |
|
| 379 |
+
| 276 | blk.30.attn_output.weight | 0xb92bb720 | 0x6e0000 |
|
| 380 |
+
| 277 | blk.30.attn_q.weight | 0xb999b720 | 0x6e0000 |
|
| 381 |
+
| 278 | blk.30.attn_v.weight | 0xba07b720 | 0x1b8000 |
|
| 382 |
+
| 279 | blk.30.ffn_down.weight | 0xba233720 | 0x1f80000 |
|
| 383 |
+
| 280 | blk.30.ffn_gate.weight | 0xbc1b3720 | 0x1810000 |
|
| 384 |
+
| 281 | blk.30.ffn_norm.weight | 0xbd9c3720 | 0x4000 |
|
| 385 |
+
| 282 | blk.30.ffn_up.weight | 0xbd9c7720 | 0x1810000 |
|
| 386 |
+
| 283 | blk.31.attn_k.weight | 0xbf1d7720 | 0x150000 |
|
| 387 |
+
| 284 | blk.31.attn_norm.weight | 0xbf327720 | 0x4000 |
|
| 388 |
+
| 285 | blk.31.attn_output.weight | 0xbf32b720 | 0x6e0000 |
|
| 389 |
+
| 286 | blk.31.attn_q.weight | 0xbfa0b720 | 0x540000 |
|
| 390 |
+
| 287 | blk.31.attn_v.weight | 0xbff4b720 | 0x1b8000 |
|
| 391 |
+
| 288 | blk.31.ffn_down.weight | 0xc0103720 | 0x1f80000 |
|
| 392 |
+
| 289 | blk.31.ffn_gate.weight | 0xc2083720 | 0x1810000 |
|
| 393 |
+
| 290 | blk.31.ffn_norm.weight | 0xc3893720 | 0x4000 |
|
| 394 |
+
| 291 | blk.31.ffn_up.weight | 0xc3897720 | 0x1810000 |
|
| 395 |
+
|
| 396 |
+
### <a name="base">Base Tensor Group : ~1B Elements</a>
|
| 397 |
+
|
| 398 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 399 |
+
|-----:|:-------------------|:---------------------------------|:------------------|:----------------------|:-----|
|
| 400 |
+
| 0 | output.weight | Output (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | Q3_K |
|
| 401 |
+
| 1 | output_norm.weight | Output Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 402 |
+
| 2 | rope_freqs.weight | Rope_Freqs (W) | ( 64) 64 | 64 x 1 x 1 x 1 | F32 |
|
| 403 |
+
| 3 | token_embd.weight | Token Embedding (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | Q2_K |
|
| 404 |
+
|
| 405 |
+
- Total elements in base: ( ~1B) 1050677312
|
| 406 |
+
- Percentage of total elements: 13.08%
|
| 407 |
+
|
| 408 |
+
|
| 409 |
+
### <a name="blk_0">Block 0 Tensor Group : ~218M Elements</a>
|
| 410 |
+
|
| 411 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 412 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 413 |
+
| 4 | blk.0.attn_k.weight | Block 0 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 414 |
+
| 5 | blk.0.attn_norm.weight | Block 0 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 415 |
+
| 6 | blk.0.attn_output.weight | Block 0 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 416 |
+
| 7 | blk.0.attn_q.weight | Block 0 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 417 |
+
| 8 | blk.0.attn_v.weight | Block 0 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 418 |
+
| 9 | blk.0.ffn_down.weight | Block 0 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q3_K |
|
| 419 |
+
| 10 | blk.0.ffn_gate.weight | Block 0 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 420 |
+
| 11 | blk.0.ffn_norm.weight | Block 0 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 421 |
+
| 12 | blk.0.ffn_up.weight | Block 0 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 422 |
+
|
| 423 |
+
- Total elements in blk.0: (~218M) 218112000
|
| 424 |
+
- Percentage of total elements: 2.72%
|
| 425 |
+
|
| 426 |
+
|
| 427 |
+
### <a name="blk_1">Block 1 Tensor Group : ~218M Elements</a>
|
| 428 |
+
|
| 429 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 430 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 431 |
+
| 13 | blk.1.attn_k.weight | Block 1 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 432 |
+
| 14 | blk.1.attn_norm.weight | Block 1 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 433 |
+
| 15 | blk.1.attn_output.weight | Block 1 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 434 |
+
| 16 | blk.1.attn_q.weight | Block 1 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 435 |
+
| 17 | blk.1.attn_v.weight | Block 1 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 436 |
+
| 18 | blk.1.ffn_down.weight | Block 1 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 437 |
+
| 19 | blk.1.ffn_gate.weight | Block 1 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 438 |
+
| 20 | blk.1.ffn_norm.weight | Block 1 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 439 |
+
| 21 | blk.1.ffn_up.weight | Block 1 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 440 |
+
|
| 441 |
+
- Total elements in blk.1: (~218M) 218112000
|
| 442 |
+
- Percentage of total elements: 2.72%
|
| 443 |
+
|
| 444 |
+
|
| 445 |
+
### <a name="blk_2">Block 2 Tensor Group : ~218M Elements</a>
|
| 446 |
+
|
| 447 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 448 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 449 |
+
| 22 | blk.2.attn_k.weight | Block 2 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 450 |
+
| 23 | blk.2.attn_norm.weight | Block 2 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 451 |
+
| 24 | blk.2.attn_output.weight | Block 2 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 452 |
+
| 25 | blk.2.attn_q.weight | Block 2 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 453 |
+
| 26 | blk.2.attn_v.weight | Block 2 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 454 |
+
| 27 | blk.2.ffn_down.weight | Block 2 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q3_K |
|
| 455 |
+
| 28 | blk.2.ffn_gate.weight | Block 2 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 456 |
+
| 29 | blk.2.ffn_norm.weight | Block 2 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 457 |
+
| 30 | blk.2.ffn_up.weight | Block 2 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 458 |
+
|
| 459 |
+
- Total elements in blk.2: (~218M) 218112000
|
| 460 |
+
- Percentage of total elements: 2.72%
|
| 461 |
+
|
| 462 |
+
|
| 463 |
+
### <a name="blk_3">Block 3 Tensor Group : ~218M Elements</a>
|
| 464 |
+
|
| 465 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 466 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 467 |
+
| 31 | blk.3.attn_k.weight | Block 3 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 468 |
+
| 32 | blk.3.attn_norm.weight | Block 3 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 469 |
+
| 33 | blk.3.attn_output.weight | Block 3 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 470 |
+
| 34 | blk.3.attn_q.weight | Block 3 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 471 |
+
| 35 | blk.3.attn_v.weight | Block 3 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 472 |
+
| 36 | blk.3.ffn_down.weight | Block 3 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q3_K |
|
| 473 |
+
| 37 | blk.3.ffn_gate.weight | Block 3 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 474 |
+
| 38 | blk.3.ffn_norm.weight | Block 3 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 475 |
+
| 39 | blk.3.ffn_up.weight | Block 3 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 476 |
+
|
| 477 |
+
- Total elements in blk.3: (~218M) 218112000
|
| 478 |
+
- Percentage of total elements: 2.72%
|
| 479 |
+
|
| 480 |
+
|
| 481 |
+
### <a name="blk_4">Block 4 Tensor Group : ~218M Elements</a>
|
| 482 |
+
|
| 483 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 484 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 485 |
+
| 40 | blk.4.attn_k.weight | Block 4 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 486 |
+
| 41 | blk.4.attn_norm.weight | Block 4 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 487 |
+
| 42 | blk.4.attn_output.weight | Block 4 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 488 |
+
| 43 | blk.4.attn_q.weight | Block 4 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 489 |
+
| 44 | blk.4.attn_v.weight | Block 4 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 490 |
+
| 45 | blk.4.ffn_down.weight | Block 4 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q3_K |
|
| 491 |
+
| 46 | blk.4.ffn_gate.weight | Block 4 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 492 |
+
| 47 | blk.4.ffn_norm.weight | Block 4 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 493 |
+
| 48 | blk.4.ffn_up.weight | Block 4 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 494 |
+
|
| 495 |
+
- Total elements in blk.4: (~218M) 218112000
|
| 496 |
+
- Percentage of total elements: 2.72%
|
| 497 |
+
|
| 498 |
+
|
| 499 |
+
### <a name="blk_5">Block 5 Tensor Group : ~218M Elements</a>
|
| 500 |
+
|
| 501 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 502 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 503 |
+
| 49 | blk.5.attn_k.weight | Block 5 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 504 |
+
| 50 | blk.5.attn_norm.weight | Block 5 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 505 |
+
| 51 | blk.5.attn_output.weight | Block 5 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 506 |
+
| 52 | blk.5.attn_q.weight | Block 5 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 507 |
+
| 53 | blk.5.attn_v.weight | Block 5 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 508 |
+
| 54 | blk.5.ffn_down.weight | Block 5 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q3_K |
|
| 509 |
+
| 55 | blk.5.ffn_gate.weight | Block 5 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 510 |
+
| 56 | blk.5.ffn_norm.weight | Block 5 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 511 |
+
| 57 | blk.5.ffn_up.weight | Block 5 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 512 |
+
|
| 513 |
+
- Total elements in blk.5: (~218M) 218112000
|
| 514 |
+
- Percentage of total elements: 2.72%
|
| 515 |
+
|
| 516 |
+
|
| 517 |
+
### <a name="blk_6">Block 6 Tensor Group : ~218M Elements</a>
|
| 518 |
+
|
| 519 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 520 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 521 |
+
| 58 | blk.6.attn_k.weight | Block 6 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 522 |
+
| 59 | blk.6.attn_norm.weight | Block 6 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 523 |
+
| 60 | blk.6.attn_output.weight | Block 6 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 524 |
+
| 61 | blk.6.attn_q.weight | Block 6 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 525 |
+
| 62 | blk.6.attn_v.weight | Block 6 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 526 |
+
| 63 | blk.6.ffn_down.weight | Block 6 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q3_K |
|
| 527 |
+
| 64 | blk.6.ffn_gate.weight | Block 6 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 528 |
+
| 65 | blk.6.ffn_norm.weight | Block 6 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 529 |
+
| 66 | blk.6.ffn_up.weight | Block 6 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 530 |
+
|
| 531 |
+
- Total elements in blk.6: (~218M) 218112000
|
| 532 |
+
- Percentage of total elements: 2.72%
|
| 533 |
+
|
| 534 |
+
|
| 535 |
+
### <a name="blk_7">Block 7 Tensor Group : ~218M Elements</a>
|
| 536 |
+
|
| 537 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 538 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 539 |
+
| 67 | blk.7.attn_k.weight | Block 7 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 540 |
+
| 68 | blk.7.attn_norm.weight | Block 7 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 541 |
+
| 69 | blk.7.attn_output.weight | Block 7 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 542 |
+
| 70 | blk.7.attn_q.weight | Block 7 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 543 |
+
| 71 | blk.7.attn_v.weight | Block 7 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 544 |
+
| 72 | blk.7.ffn_down.weight | Block 7 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q3_K |
|
| 545 |
+
| 73 | blk.7.ffn_gate.weight | Block 7 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 546 |
+
| 74 | blk.7.ffn_norm.weight | Block 7 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 547 |
+
| 75 | blk.7.ffn_up.weight | Block 7 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 548 |
+
|
| 549 |
+
- Total elements in blk.7: (~218M) 218112000
|
| 550 |
+
- Percentage of total elements: 2.72%
|
| 551 |
+
|
| 552 |
+
|
| 553 |
+
### <a name="blk_8">Block 8 Tensor Group : ~218M Elements</a>
|
| 554 |
+
|
| 555 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 556 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 557 |
+
| 76 | blk.8.attn_k.weight | Block 8 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 558 |
+
| 77 | blk.8.attn_norm.weight | Block 8 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 559 |
+
| 78 | blk.8.attn_output.weight | Block 8 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 560 |
+
| 79 | blk.8.attn_q.weight | Block 8 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 561 |
+
| 80 | blk.8.attn_v.weight | Block 8 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 562 |
+
| 81 | blk.8.ffn_down.weight | Block 8 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q3_K |
|
| 563 |
+
| 82 | blk.8.ffn_gate.weight | Block 8 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 564 |
+
| 83 | blk.8.ffn_norm.weight | Block 8 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 565 |
+
| 84 | blk.8.ffn_up.weight | Block 8 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 566 |
+
|
| 567 |
+
- Total elements in blk.8: (~218M) 218112000
|
| 568 |
+
- Percentage of total elements: 2.72%
|
| 569 |
+
|
| 570 |
+
|
| 571 |
+
### <a name="blk_9">Block 9 Tensor Group : ~218M Elements</a>
|
| 572 |
+
|
| 573 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 574 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 575 |
+
| 85 | blk.9.attn_k.weight | Block 9 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 576 |
+
| 86 | blk.9.attn_norm.weight | Block 9 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 577 |
+
| 87 | blk.9.attn_output.weight | Block 9 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 578 |
+
| 88 | blk.9.attn_q.weight | Block 9 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 579 |
+
| 89 | blk.9.attn_v.weight | Block 9 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 580 |
+
| 90 | blk.9.ffn_down.weight | Block 9 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q3_K |
|
| 581 |
+
| 91 | blk.9.ffn_gate.weight | Block 9 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 582 |
+
| 92 | blk.9.ffn_norm.weight | Block 9 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 583 |
+
| 93 | blk.9.ffn_up.weight | Block 9 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 584 |
+
|
| 585 |
+
- Total elements in blk.9: (~218M) 218112000
|
| 586 |
+
- Percentage of total elements: 2.72%
|
| 587 |
+
|
| 588 |
+
|
| 589 |
+
### <a name="blk_10">Block 10 Tensor Group : ~218M Elements</a>
|
| 590 |
+
|
| 591 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 592 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 593 |
+
| 94 | blk.10.attn_k.weight | Block 10 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 594 |
+
| 95 | blk.10.attn_norm.weight | Block 10 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 595 |
+
| 96 | blk.10.attn_output.weight | Block 10 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 596 |
+
| 97 | blk.10.attn_q.weight | Block 10 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 597 |
+
| 98 | blk.10.attn_v.weight | Block 10 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 598 |
+
| 99 | blk.10.ffn_down.weight | Block 10 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q3_K |
|
| 599 |
+
| 100 | blk.10.ffn_gate.weight | Block 10 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 600 |
+
| 101 | blk.10.ffn_norm.weight | Block 10 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 601 |
+
| 102 | blk.10.ffn_up.weight | Block 10 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 602 |
+
|
| 603 |
+
- Total elements in blk.10: (~218M) 218112000
|
| 604 |
+
- Percentage of total elements: 2.72%
|
| 605 |
+
|
| 606 |
+
|
| 607 |
+
### <a name="blk_11">Block 11 Tensor Group : ~218M Elements</a>
|
| 608 |
+
|
| 609 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 610 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 611 |
+
| 103 | blk.11.attn_k.weight | Block 11 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 612 |
+
| 104 | blk.11.attn_norm.weight | Block 11 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 613 |
+
| 105 | blk.11.attn_output.weight | Block 11 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 614 |
+
| 106 | blk.11.attn_q.weight | Block 11 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 615 |
+
| 107 | blk.11.attn_v.weight | Block 11 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 616 |
+
| 108 | blk.11.ffn_down.weight | Block 11 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q3_K |
|
| 617 |
+
| 109 | blk.11.ffn_gate.weight | Block 11 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 618 |
+
| 110 | blk.11.ffn_norm.weight | Block 11 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 619 |
+
| 111 | blk.11.ffn_up.weight | Block 11 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 620 |
+
|
| 621 |
+
- Total elements in blk.11: (~218M) 218112000
|
| 622 |
+
- Percentage of total elements: 2.72%
|
| 623 |
+
|
| 624 |
+
|
| 625 |
+
### <a name="blk_12">Block 12 Tensor Group : ~218M Elements</a>
|
| 626 |
+
|
| 627 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 628 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 629 |
+
| 112 | blk.12.attn_k.weight | Block 12 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 630 |
+
| 113 | blk.12.attn_norm.weight | Block 12 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 631 |
+
| 114 | blk.12.attn_output.weight | Block 12 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 632 |
+
| 115 | blk.12.attn_q.weight | Block 12 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 633 |
+
| 116 | blk.12.attn_v.weight | Block 12 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 634 |
+
| 117 | blk.12.ffn_down.weight | Block 12 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q3_K |
|
| 635 |
+
| 118 | blk.12.ffn_gate.weight | Block 12 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 636 |
+
| 119 | blk.12.ffn_norm.weight | Block 12 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 637 |
+
| 120 | blk.12.ffn_up.weight | Block 12 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 638 |
+
|
| 639 |
+
- Total elements in blk.12: (~218M) 218112000
|
| 640 |
+
- Percentage of total elements: 2.72%
|
| 641 |
+
|
| 642 |
+
|
| 643 |
+
### <a name="blk_13">Block 13 Tensor Group : ~218M Elements</a>
|
| 644 |
+
|
| 645 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 646 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 647 |
+
| 121 | blk.13.attn_k.weight | Block 13 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 648 |
+
| 122 | blk.13.attn_norm.weight | Block 13 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 649 |
+
| 123 | blk.13.attn_output.weight | Block 13 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 650 |
+
| 124 | blk.13.attn_q.weight | Block 13 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 651 |
+
| 125 | blk.13.attn_v.weight | Block 13 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 652 |
+
| 126 | blk.13.ffn_down.weight | Block 13 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q3_K |
|
| 653 |
+
| 127 | blk.13.ffn_gate.weight | Block 13 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 654 |
+
| 128 | blk.13.ffn_norm.weight | Block 13 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 655 |
+
| 129 | blk.13.ffn_up.weight | Block 13 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 656 |
+
|
| 657 |
+
- Total elements in blk.13: (~218M) 218112000
|
| 658 |
+
- Percentage of total elements: 2.72%
|
| 659 |
+
|
| 660 |
+
|
| 661 |
+
### <a name="blk_14">Block 14 Tensor Group : ~218M Elements</a>
|
| 662 |
+
|
| 663 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 664 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 665 |
+
| 130 | blk.14.attn_k.weight | Block 14 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 666 |
+
| 131 | blk.14.attn_norm.weight | Block 14 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 667 |
+
| 132 | blk.14.attn_output.weight | Block 14 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 668 |
+
| 133 | blk.14.attn_q.weight | Block 14 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 669 |
+
| 134 | blk.14.attn_v.weight | Block 14 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 670 |
+
| 135 | blk.14.ffn_down.weight | Block 14 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q3_K |
|
| 671 |
+
| 136 | blk.14.ffn_gate.weight | Block 14 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 672 |
+
| 137 | blk.14.ffn_norm.weight | Block 14 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 673 |
+
| 138 | blk.14.ffn_up.weight | Block 14 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 674 |
+
|
| 675 |
+
- Total elements in blk.14: (~218M) 218112000
|
| 676 |
+
- Percentage of total elements: 2.72%
|
| 677 |
+
|
| 678 |
+
|
| 679 |
+
### <a name="blk_15">Block 15 Tensor Group : ~218M Elements</a>
|
| 680 |
+
|
| 681 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 682 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 683 |
+
| 139 | blk.15.attn_k.weight | Block 15 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 684 |
+
| 140 | blk.15.attn_norm.weight | Block 15 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 685 |
+
| 141 | blk.15.attn_output.weight | Block 15 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 686 |
+
| 142 | blk.15.attn_q.weight | Block 15 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 687 |
+
| 143 | blk.15.attn_v.weight | Block 15 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 688 |
+
| 144 | blk.15.ffn_down.weight | Block 15 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q3_K |
|
| 689 |
+
| 145 | blk.15.ffn_gate.weight | Block 15 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 690 |
+
| 146 | blk.15.ffn_norm.weight | Block 15 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 691 |
+
| 147 | blk.15.ffn_up.weight | Block 15 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q2_K |
|
| 692 |
+
|
| 693 |
+
- Total elements in blk.15: (~218M) 218112000
|
| 694 |
+
- Percentage of total elements: 2.72%
|
| 695 |
+
|
| 696 |
+
|
| 697 |
+
### <a name="blk_16">Block 16 Tensor Group : ~218M Elements</a>
|
| 698 |
+
|
| 699 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 700 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 701 |
+
| 148 | blk.16.attn_k.weight | Block 16 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 702 |
+
| 149 | blk.16.attn_norm.weight | Block 16 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 703 |
+
| 150 | blk.16.attn_output.weight | Block 16 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 704 |
+
| 151 | blk.16.attn_q.weight | Block 16 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 705 |
+
| 152 | blk.16.attn_v.weight | Block 16 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 706 |
+
| 153 | blk.16.ffn_down.weight | Block 16 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q3_K |
|
| 707 |
+
| 154 | blk.16.ffn_gate.weight | Block 16 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 708 |
+
| 155 | blk.16.ffn_norm.weight | Block 16 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 709 |
+
| 156 | blk.16.ffn_up.weight | Block 16 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 710 |
+
|
| 711 |
+
- Total elements in blk.16: (~218M) 218112000
|
| 712 |
+
- Percentage of total elements: 2.72%
|
| 713 |
+
|
| 714 |
+
|
| 715 |
+
### <a name="blk_17">Block 17 Tensor Group : ~218M Elements</a>
|
| 716 |
+
|
| 717 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 718 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 719 |
+
| 157 | blk.17.attn_k.weight | Block 17 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 720 |
+
| 158 | blk.17.attn_norm.weight | Block 17 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 721 |
+
| 159 | blk.17.attn_output.weight | Block 17 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 722 |
+
| 160 | blk.17.attn_q.weight | Block 17 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 723 |
+
| 161 | blk.17.attn_v.weight | Block 17 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 724 |
+
| 162 | blk.17.ffn_down.weight | Block 17 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 725 |
+
| 163 | blk.17.ffn_gate.weight | Block 17 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 726 |
+
| 164 | blk.17.ffn_norm.weight | Block 17 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 727 |
+
| 165 | blk.17.ffn_up.weight | Block 17 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 728 |
+
|
| 729 |
+
- Total elements in blk.17: (~218M) 218112000
|
| 730 |
+
- Percentage of total elements: 2.72%
|
| 731 |
+
|
| 732 |
+
|
| 733 |
+
### <a name="blk_18">Block 18 Tensor Group : ~218M Elements</a>
|
| 734 |
+
|
| 735 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 736 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 737 |
+
| 166 | blk.18.attn_k.weight | Block 18 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 738 |
+
| 167 | blk.18.attn_norm.weight | Block 18 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 739 |
+
| 168 | blk.18.attn_output.weight | Block 18 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 740 |
+
| 169 | blk.18.attn_q.weight | Block 18 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 741 |
+
| 170 | blk.18.attn_v.weight | Block 18 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 742 |
+
| 171 | blk.18.ffn_down.weight | Block 18 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 743 |
+
| 172 | blk.18.ffn_gate.weight | Block 18 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 744 |
+
| 173 | blk.18.ffn_norm.weight | Block 18 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 745 |
+
| 174 | blk.18.ffn_up.weight | Block 18 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 746 |
+
|
| 747 |
+
- Total elements in blk.18: (~218M) 218112000
|
| 748 |
+
- Percentage of total elements: 2.72%
|
| 749 |
+
|
| 750 |
+
|
| 751 |
+
### <a name="blk_19">Block 19 Tensor Group : ~218M Elements</a>
|
| 752 |
+
|
| 753 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 754 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 755 |
+
| 175 | blk.19.attn_k.weight | Block 19 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 756 |
+
| 176 | blk.19.attn_norm.weight | Block 19 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 757 |
+
| 177 | blk.19.attn_output.weight | Block 19 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 758 |
+
| 178 | blk.19.attn_q.weight | Block 19 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 759 |
+
| 179 | blk.19.attn_v.weight | Block 19 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 760 |
+
| 180 | blk.19.ffn_down.weight | Block 19 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 761 |
+
| 181 | blk.19.ffn_gate.weight | Block 19 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 762 |
+
| 182 | blk.19.ffn_norm.weight | Block 19 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 763 |
+
| 183 | blk.19.ffn_up.weight | Block 19 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 764 |
+
|
| 765 |
+
- Total elements in blk.19: (~218M) 218112000
|
| 766 |
+
- Percentage of total elements: 2.72%
|
| 767 |
+
|
| 768 |
+
|
| 769 |
+
### <a name="blk_20">Block 20 Tensor Group : ~218M Elements</a>
|
| 770 |
+
|
| 771 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 772 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 773 |
+
| 184 | blk.20.attn_k.weight | Block 20 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 774 |
+
| 185 | blk.20.attn_norm.weight | Block 20 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 775 |
+
| 186 | blk.20.attn_output.weight | Block 20 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 776 |
+
| 187 | blk.20.attn_q.weight | Block 20 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 777 |
+
| 188 | blk.20.attn_v.weight | Block 20 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 778 |
+
| 189 | blk.20.ffn_down.weight | Block 20 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 779 |
+
| 190 | blk.20.ffn_gate.weight | Block 20 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 780 |
+
| 191 | blk.20.ffn_norm.weight | Block 20 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 781 |
+
| 192 | blk.20.ffn_up.weight | Block 20 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 782 |
+
|
| 783 |
+
- Total elements in blk.20: (~218M) 218112000
|
| 784 |
+
- Percentage of total elements: 2.72%
|
| 785 |
+
|
| 786 |
+
|
| 787 |
+
### <a name="blk_21">Block 21 Tensor Group : ~218M Elements</a>
|
| 788 |
+
|
| 789 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 790 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 791 |
+
| 193 | blk.21.attn_k.weight | Block 21 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 792 |
+
| 194 | blk.21.attn_norm.weight | Block 21 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 793 |
+
| 195 | blk.21.attn_output.weight | Block 21 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 794 |
+
| 196 | blk.21.attn_q.weight | Block 21 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 795 |
+
| 197 | blk.21.attn_v.weight | Block 21 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 796 |
+
| 198 | blk.21.ffn_down.weight | Block 21 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 797 |
+
| 199 | blk.21.ffn_gate.weight | Block 21 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 798 |
+
| 200 | blk.21.ffn_norm.weight | Block 21 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 799 |
+
| 201 | blk.21.ffn_up.weight | Block 21 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 800 |
+
|
| 801 |
+
- Total elements in blk.21: (~218M) 218112000
|
| 802 |
+
- Percentage of total elements: 2.72%
|
| 803 |
+
|
| 804 |
+
|
| 805 |
+
### <a name="blk_22">Block 22 Tensor Group : ~218M Elements</a>
|
| 806 |
+
|
| 807 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 808 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 809 |
+
| 202 | blk.22.attn_k.weight | Block 22 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 810 |
+
| 203 | blk.22.attn_norm.weight | Block 22 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 811 |
+
| 204 | blk.22.attn_output.weight | Block 22 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 812 |
+
| 205 | blk.22.attn_q.weight | Block 22 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 813 |
+
| 206 | blk.22.attn_v.weight | Block 22 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 814 |
+
| 207 | blk.22.ffn_down.weight | Block 22 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 815 |
+
| 208 | blk.22.ffn_gate.weight | Block 22 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 816 |
+
| 209 | blk.22.ffn_norm.weight | Block 22 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 817 |
+
| 210 | blk.22.ffn_up.weight | Block 22 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 818 |
+
|
| 819 |
+
- Total elements in blk.22: (~218M) 218112000
|
| 820 |
+
- Percentage of total elements: 2.72%
|
| 821 |
+
|
| 822 |
+
|
| 823 |
+
### <a name="blk_23">Block 23 Tensor Group : ~218M Elements</a>
|
| 824 |
+
|
| 825 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 826 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 827 |
+
| 211 | blk.23.attn_k.weight | Block 23 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 828 |
+
| 212 | blk.23.attn_norm.weight | Block 23 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 829 |
+
| 213 | blk.23.attn_output.weight | Block 23 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 830 |
+
| 214 | blk.23.attn_q.weight | Block 23 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 831 |
+
| 215 | blk.23.attn_v.weight | Block 23 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 832 |
+
| 216 | blk.23.ffn_down.weight | Block 23 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 833 |
+
| 217 | blk.23.ffn_gate.weight | Block 23 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 834 |
+
| 218 | blk.23.ffn_norm.weight | Block 23 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 835 |
+
| 219 | blk.23.ffn_up.weight | Block 23 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 836 |
+
|
| 837 |
+
- Total elements in blk.23: (~218M) 218112000
|
| 838 |
+
- Percentage of total elements: 2.72%
|
| 839 |
+
|
| 840 |
+
|
| 841 |
+
### <a name="blk_24">Block 24 Tensor Group : ~218M Elements</a>
|
| 842 |
+
|
| 843 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 844 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 845 |
+
| 220 | blk.24.attn_k.weight | Block 24 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 846 |
+
| 221 | blk.24.attn_norm.weight | Block 24 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 847 |
+
| 222 | blk.24.attn_output.weight | Block 24 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 848 |
+
| 223 | blk.24.attn_q.weight | Block 24 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 849 |
+
| 224 | blk.24.attn_v.weight | Block 24 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 850 |
+
| 225 | blk.24.ffn_down.weight | Block 24 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 851 |
+
| 226 | blk.24.ffn_gate.weight | Block 24 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 852 |
+
| 227 | blk.24.ffn_norm.weight | Block 24 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 853 |
+
| 228 | blk.24.ffn_up.weight | Block 24 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 854 |
+
|
| 855 |
+
- Total elements in blk.24: (~218M) 218112000
|
| 856 |
+
- Percentage of total elements: 2.72%
|
| 857 |
+
|
| 858 |
+
|
| 859 |
+
### <a name="blk_25">Block 25 Tensor Group : ~218M Elements</a>
|
| 860 |
+
|
| 861 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 862 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 863 |
+
| 229 | blk.25.attn_k.weight | Block 25 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 864 |
+
| 230 | blk.25.attn_norm.weight | Block 25 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 865 |
+
| 231 | blk.25.attn_output.weight | Block 25 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 866 |
+
| 232 | blk.25.attn_q.weight | Block 25 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 867 |
+
| 233 | blk.25.attn_v.weight | Block 25 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 868 |
+
| 234 | blk.25.ffn_down.weight | Block 25 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 869 |
+
| 235 | blk.25.ffn_gate.weight | Block 25 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 870 |
+
| 236 | blk.25.ffn_norm.weight | Block 25 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 871 |
+
| 237 | blk.25.ffn_up.weight | Block 25 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 872 |
+
|
| 873 |
+
- Total elements in blk.25: (~218M) 218112000
|
| 874 |
+
- Percentage of total elements: 2.72%
|
| 875 |
+
|
| 876 |
+
|
| 877 |
+
### <a name="blk_26">Block 26 Tensor Group : ~218M Elements</a>
|
| 878 |
+
|
| 879 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 880 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 881 |
+
| 238 | blk.26.attn_k.weight | Block 26 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 882 |
+
| 239 | blk.26.attn_norm.weight | Block 26 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 883 |
+
| 240 | blk.26.attn_output.weight | Block 26 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 884 |
+
| 241 | blk.26.attn_q.weight | Block 26 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 885 |
+
| 242 | blk.26.attn_v.weight | Block 26 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 886 |
+
| 243 | blk.26.ffn_down.weight | Block 26 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 887 |
+
| 244 | blk.26.ffn_gate.weight | Block 26 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 888 |
+
| 245 | blk.26.ffn_norm.weight | Block 26 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 889 |
+
| 246 | blk.26.ffn_up.weight | Block 26 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 890 |
+
|
| 891 |
+
- Total elements in blk.26: (~218M) 218112000
|
| 892 |
+
- Percentage of total elements: 2.72%
|
| 893 |
+
|
| 894 |
+
|
| 895 |
+
### <a name="blk_27">Block 27 Tensor Group : ~218M Elements</a>
|
| 896 |
+
|
| 897 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 898 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 899 |
+
| 247 | blk.27.attn_k.weight | Block 27 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 900 |
+
| 248 | blk.27.attn_norm.weight | Block 27 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 901 |
+
| 249 | blk.27.attn_output.weight | Block 27 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 902 |
+
| 250 | blk.27.attn_q.weight | Block 27 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 903 |
+
| 251 | blk.27.attn_v.weight | Block 27 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 904 |
+
| 252 | blk.27.ffn_down.weight | Block 27 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 905 |
+
| 253 | blk.27.ffn_gate.weight | Block 27 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 906 |
+
| 254 | blk.27.ffn_norm.weight | Block 27 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 907 |
+
| 255 | blk.27.ffn_up.weight | Block 27 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 908 |
+
|
| 909 |
+
- Total elements in blk.27: (~218M) 218112000
|
| 910 |
+
- Percentage of total elements: 2.72%
|
| 911 |
+
|
| 912 |
+
|
| 913 |
+
### <a name="blk_28">Block 28 Tensor Group : ~218M Elements</a>
|
| 914 |
+
|
| 915 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 916 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 917 |
+
| 256 | blk.28.attn_k.weight | Block 28 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 918 |
+
| 257 | blk.28.attn_norm.weight | Block 28 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 919 |
+
| 258 | blk.28.attn_output.weight | Block 28 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 920 |
+
| 259 | blk.28.attn_q.weight | Block 28 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 921 |
+
| 260 | blk.28.attn_v.weight | Block 28 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 922 |
+
| 261 | blk.28.ffn_down.weight | Block 28 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 923 |
+
| 262 | blk.28.ffn_gate.weight | Block 28 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 924 |
+
| 263 | blk.28.ffn_norm.weight | Block 28 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 925 |
+
| 264 | blk.28.ffn_up.weight | Block 28 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 926 |
+
|
| 927 |
+
- Total elements in blk.28: (~218M) 218112000
|
| 928 |
+
- Percentage of total elements: 2.72%
|
| 929 |
+
|
| 930 |
+
|
| 931 |
+
### <a name="blk_29">Block 29 Tensor Group : ~218M Elements</a>
|
| 932 |
+
|
| 933 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 934 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 935 |
+
| 265 | blk.29.attn_k.weight | Block 29 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 936 |
+
| 266 | blk.29.attn_norm.weight | Block 29 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 937 |
+
| 267 | blk.29.attn_output.weight | Block 29 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 938 |
+
| 268 | blk.29.attn_q.weight | Block 29 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 939 |
+
| 269 | blk.29.attn_v.weight | Block 29 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 940 |
+
| 270 | blk.29.ffn_down.weight | Block 29 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 941 |
+
| 271 | blk.29.ffn_gate.weight | Block 29 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 942 |
+
| 272 | blk.29.ffn_norm.weight | Block 29 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 943 |
+
| 273 | blk.29.ffn_up.weight | Block 29 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 944 |
+
|
| 945 |
+
- Total elements in blk.29: (~218M) 218112000
|
| 946 |
+
- Percentage of total elements: 2.72%
|
| 947 |
+
|
| 948 |
+
|
| 949 |
+
### <a name="blk_30">Block 30 Tensor Group : ~218M Elements</a>
|
| 950 |
+
|
| 951 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 952 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 953 |
+
| 274 | blk.30.attn_k.weight | Block 30 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 954 |
+
| 275 | blk.30.attn_norm.weight | Block 30 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 955 |
+
| 276 | blk.30.attn_output.weight | Block 30 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 956 |
+
| 277 | blk.30.attn_q.weight | Block 30 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 957 |
+
| 278 | blk.30.attn_v.weight | Block 30 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 958 |
+
| 279 | blk.30.ffn_down.weight | Block 30 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 959 |
+
| 280 | blk.30.ffn_gate.weight | Block 30 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 960 |
+
| 281 | blk.30.ffn_norm.weight | Block 30 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 961 |
+
| 282 | blk.30.ffn_up.weight | Block 30 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 962 |
+
|
| 963 |
+
- Total elements in blk.30: (~218M) 218112000
|
| 964 |
+
- Percentage of total elements: 2.72%
|
| 965 |
+
|
| 966 |
+
|
| 967 |
+
### <a name="blk_31">Block 31 Tensor Group : ~218M Elements</a>
|
| 968 |
+
|
| 969 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 970 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 971 |
+
| 283 | blk.31.attn_k.weight | Block 31 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q2_K |
|
| 972 |
+
| 284 | blk.31.attn_norm.weight | Block 31 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 973 |
+
| 285 | blk.31.attn_output.weight | Block 31 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 974 |
+
| 286 | blk.31.attn_q.weight | Block 31 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q2_K |
|
| 975 |
+
| 287 | blk.31.attn_v.weight | Block 31 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 976 |
+
| 288 | blk.31.ffn_down.weight | Block 31 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 977 |
+
| 289 | blk.31.ffn_gate.weight | Block 31 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 978 |
+
| 290 | blk.31.ffn_norm.weight | Block 31 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 979 |
+
| 291 | blk.31.ffn_up.weight | Block 31 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 980 |
+
|
| 981 |
+
- Total elements in blk.31: (~218M) 218112000
|
| 982 |
+
- Percentage of total elements: 2.72%
|
scores/Watt-Tool-8B-Q4_K_M.md
ADDED
|
@@ -0,0 +1,982 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Watt-Tool-8B-Q4_K_M.gguf - GGUF Internal File Dump
|
| 2 |
+
|
| 3 |
+
- Endian: LITTLE endian
|
| 4 |
+
|
| 5 |
+
## Key Value Metadata Store
|
| 6 |
+
|
| 7 |
+
There are 43 key-value pairs in this file
|
| 8 |
+
|
| 9 |
+
| POS | TYPE | Count | Key | Value |
|
| 10 |
+
|----:|:---------|-------:|:---------------------------------------|:--------------------------------------------------------------------|
|
| 11 |
+
| 1 | UINT32 | 1 | GGUF.version | 3 |
|
| 12 |
+
| 2 | UINT64 | 1 | GGUF.tensor_count | 292 |
|
| 13 |
+
| 3 | UINT64 | 1 | GGUF.kv_count | 40 |
|
| 14 |
+
| 4 | STRING | 1 | general.architecture | `llama` |
|
| 15 |
+
| 5 | STRING | 1 | general.type | `model` |
|
| 16 |
+
| 6 | STRING | 1 | general.name | `Watt Tool 8B GGUF` |
|
| 17 |
+
| 7 | STRING | 1 | general.finetune | `GGUF` |
|
| 18 |
+
| 8 | STRING | 1 | general.basename | `Watt-Tool` |
|
| 19 |
+
| 9 | STRING | 1 | general.size_label | `8B` |
|
| 20 |
+
| 10 | STRING | 1 | general.license | `apache-2.0` |
|
| 21 |
+
| 11 | UINT32 | 1 | general.base_model.count | 1 |
|
| 22 |
+
| 12 | STRING | 1 | general.base_model.0.name | `Llama 3.1 8B Instruct` |
|
| 23 |
+
| 13 | STRING | 1 | general.base_model.0.organization | `Meta Llama` |
|
| 24 |
+
| 14 | STRING | 1 | general.base_model.0.repo_url | `https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct` |
|
| 25 |
+
| 15 | [STRING] | 4 | general.tags | [ `function-calling`, `tool-use`, `llama`, `bfcl` ] |
|
| 26 |
+
| 16 | [STRING] | 1 | general.languages | [ `en` ] |
|
| 27 |
+
| 17 | UINT32 | 1 | llama.block_count | 32 |
|
| 28 |
+
| 18 | UINT32 | 1 | llama.context_length | 131072 |
|
| 29 |
+
| 19 | UINT32 | 1 | llama.embedding_length | 4096 |
|
| 30 |
+
| 20 | UINT32 | 1 | llama.feed_forward_length | 14336 |
|
| 31 |
+
| 21 | UINT32 | 1 | llama.attention.head_count | 32 |
|
| 32 |
+
| 22 | UINT32 | 1 | llama.attention.head_count_kv | 8 |
|
| 33 |
+
| 23 | FLOAT32 | 1 | llama.rope.freq_base | 500000.0 |
|
| 34 |
+
| 24 | FLOAT32 | 1 | llama.attention.layer_norm_rms_epsilon | 1e-05 |
|
| 35 |
+
| 25 | UINT32 | 1 | llama.attention.key_length | 128 |
|
| 36 |
+
| 26 | UINT32 | 1 | llama.attention.value_length | 128 |
|
| 37 |
+
| 27 | UINT32 | 1 | llama.vocab_size | 128256 |
|
| 38 |
+
| 28 | UINT32 | 1 | llama.rope.dimension_count | 128 |
|
| 39 |
+
| 29 | STRING | 1 | tokenizer.ggml.model | `gpt2` |
|
| 40 |
+
| 30 | STRING | 1 | tokenizer.ggml.pre | `llama-bpe` |
|
| 41 |
+
| 31 | [STRING] | 128256 | tokenizer.ggml.tokens | [ `!`, `"`, `#`, `$`, `%`, ... ] |
|
| 42 |
+
| 32 | [INT32] | 128256 | tokenizer.ggml.token_type | [ 1, 1, 1, 1, 1, 1, 1, ... ] |
|
| 43 |
+
| 33 | [STRING] | 280147 | tokenizer.ggml.merges | [ `Ġ Ġ`, `Ġ ĠĠĠ`, `ĠĠ ĠĠ`, `ĠĠĠ Ġ`, `i n`, ... ] |
|
| 44 |
+
| 34 | UINT32 | 1 | tokenizer.ggml.bos_token_id | 128000 |
|
| 45 |
+
| 35 | UINT32 | 1 | tokenizer.ggml.eos_token_id | 128009 |
|
| 46 |
+
| 36 | UINT32 | 1 | tokenizer.ggml.padding_token_id | 128009 |
|
| 47 |
+
| 37 | STRING | 1 | tokenizer.chat_template | `{{ '<|begin_of_text|>' }}{% if`...`d|>' }}{% endif %}{% endfor %}` |
|
| 48 |
+
| 38 | UINT32 | 1 | general.quantization_version | 2 |
|
| 49 |
+
| 39 | UINT32 | 1 | general.file_type | 15 |
|
| 50 |
+
| 40 | STRING | 1 | quantize.imatrix.file | `./imatrix/imatrix-Watt-Tool-8B-small.dat` |
|
| 51 |
+
| 41 | STRING | 1 | quantize.imatrix.dataset | `../../datasets/imatrix/calibration_eur_small.txt` |
|
| 52 |
+
| 42 | INT32 | 1 | quantize.imatrix.entries_count | 225 |
|
| 53 |
+
| 43 | INT32 | 1 | quantize.imatrix.chunks_count | 962 |
|
| 54 |
+
|
| 55 |
+
## Tensors Overview ~8B Elements
|
| 56 |
+
|
| 57 |
+
Total number of elements in all tensors: 8030261312 Elements
|
| 58 |
+
|
| 59 |
+
- [Watt-Tool-8B-Q4\_K\_M.gguf - GGUF Internal File Dump](#watt-tool-8b-q4_k_mgguf---gguf-internal-file-dump)
|
| 60 |
+
- [Key Value Metadata Store](#key-value-metadata-store)
|
| 61 |
+
- [Tensors Overview ~8B Elements](#tensors-overview-8b-elements)
|
| 62 |
+
- [Tensor Data Offset](#tensor-data-offset)
|
| 63 |
+
- [Base Tensor Group : ~1B Elements](#base-tensor-group--1b-elements)
|
| 64 |
+
- [Block 0 Tensor Group : ~218M Elements](#block-0-tensor-group--218m-elements)
|
| 65 |
+
- [Block 1 Tensor Group : ~218M Elements](#block-1-tensor-group--218m-elements)
|
| 66 |
+
- [Block 2 Tensor Group : ~218M Elements](#block-2-tensor-group--218m-elements)
|
| 67 |
+
- [Block 3 Tensor Group : ~218M Elements](#block-3-tensor-group--218m-elements)
|
| 68 |
+
- [Block 4 Tensor Group : ~218M Elements](#block-4-tensor-group--218m-elements)
|
| 69 |
+
- [Block 5 Tensor Group : ~218M Elements](#block-5-tensor-group--218m-elements)
|
| 70 |
+
- [Block 6 Tensor Group : ~218M Elements](#block-6-tensor-group--218m-elements)
|
| 71 |
+
- [Block 7 Tensor Group : ~218M Elements](#block-7-tensor-group--218m-elements)
|
| 72 |
+
- [Block 8 Tensor Group : ~218M Elements](#block-8-tensor-group--218m-elements)
|
| 73 |
+
- [Block 9 Tensor Group : ~218M Elements](#block-9-tensor-group--218m-elements)
|
| 74 |
+
- [Block 10 Tensor Group : ~218M Elements](#block-10-tensor-group--218m-elements)
|
| 75 |
+
- [Block 11 Tensor Group : ~218M Elements](#block-11-tensor-group--218m-elements)
|
| 76 |
+
- [Block 12 Tensor Group : ~218M Elements](#block-12-tensor-group--218m-elements)
|
| 77 |
+
- [Block 13 Tensor Group : ~218M Elements](#block-13-tensor-group--218m-elements)
|
| 78 |
+
- [Block 14 Tensor Group : ~218M Elements](#block-14-tensor-group--218m-elements)
|
| 79 |
+
- [Block 15 Tensor Group : ~218M Elements](#block-15-tensor-group--218m-elements)
|
| 80 |
+
- [Block 16 Tensor Group : ~218M Elements](#block-16-tensor-group--218m-elements)
|
| 81 |
+
- [Block 17 Tensor Group : ~218M Elements](#block-17-tensor-group--218m-elements)
|
| 82 |
+
- [Block 18 Tensor Group : ~218M Elements](#block-18-tensor-group--218m-elements)
|
| 83 |
+
- [Block 19 Tensor Group : ~218M Elements](#block-19-tensor-group--218m-elements)
|
| 84 |
+
- [Block 20 Tensor Group : ~218M Elements](#block-20-tensor-group--218m-elements)
|
| 85 |
+
- [Block 21 Tensor Group : ~218M Elements](#block-21-tensor-group--218m-elements)
|
| 86 |
+
- [Block 22 Tensor Group : ~218M Elements](#block-22-tensor-group--218m-elements)
|
| 87 |
+
- [Block 23 Tensor Group : ~218M Elements](#block-23-tensor-group--218m-elements)
|
| 88 |
+
- [Block 24 Tensor Group : ~218M Elements](#block-24-tensor-group--218m-elements)
|
| 89 |
+
- [Block 25 Tensor Group : ~218M Elements](#block-25-tensor-group--218m-elements)
|
| 90 |
+
- [Block 26 Tensor Group : ~218M Elements](#block-26-tensor-group--218m-elements)
|
| 91 |
+
- [Block 27 Tensor Group : ~218M Elements](#block-27-tensor-group--218m-elements)
|
| 92 |
+
- [Block 28 Tensor Group : ~218M Elements](#block-28-tensor-group--218m-elements)
|
| 93 |
+
- [Block 29 Tensor Group : ~218M Elements](#block-29-tensor-group--218m-elements)
|
| 94 |
+
- [Block 30 Tensor Group : ~218M Elements](#block-30-tensor-group--218m-elements)
|
| 95 |
+
- [Block 31 Tensor Group : ~218M Elements](#block-31-tensor-group--218m-elements)
|
| 96 |
+
|
| 97 |
+
### Tensor Data Offset
|
| 98 |
+
|
| 99 |
+
This table contains the offset and data segment relative to start of file
|
| 100 |
+
|
| 101 |
+
| T_ID | Tensor Layer Name | Data Offset (B) | Data Size (B) |
|
| 102 |
+
|-----:|:--------------------------|-----------------:|-----------------:|
|
| 103 |
+
| 0 | output.weight | 0x779620 | 0x119d0000 |
|
| 104 |
+
| 1 | output_norm.weight | 0x12149620 | 0x4000 |
|
| 105 |
+
| 2 | rope_freqs.weight | 0x1214d620 | 0x100 |
|
| 106 |
+
| 3 | token_embd.weight | 0x1214d720 | 0xd746000 |
|
| 107 |
+
| 4 | blk.0.attn_k.weight | 0x1f893720 | 0x1b8000 |
|
| 108 |
+
| 5 | blk.0.attn_norm.weight | 0x1fa4b720 | 0x4000 |
|
| 109 |
+
| 6 | blk.0.attn_output.weight | 0x1fa4f720 | 0x900000 |
|
| 110 |
+
| 7 | blk.0.attn_q.weight | 0x2034f720 | 0x6e0000 |
|
| 111 |
+
| 8 | blk.0.attn_v.weight | 0x20a2f720 | 0x240000 |
|
| 112 |
+
| 9 | blk.0.ffn_down.weight | 0x20c6f720 | 0x2680000 |
|
| 113 |
+
| 10 | blk.0.ffn_gate.weight | 0x232ef720 | 0x1810000 |
|
| 114 |
+
| 11 | blk.0.ffn_norm.weight | 0x24aff720 | 0x4000 |
|
| 115 |
+
| 12 | blk.0.ffn_up.weight | 0x24b03720 | 0x1810000 |
|
| 116 |
+
| 13 | blk.1.attn_k.weight | 0x26313720 | 0x1b8000 |
|
| 117 |
+
| 14 | blk.1.attn_norm.weight | 0x264cb720 | 0x4000 |
|
| 118 |
+
| 15 | blk.1.attn_output.weight | 0x264cf720 | 0x900000 |
|
| 119 |
+
| 16 | blk.1.attn_q.weight | 0x26dcf720 | 0x6e0000 |
|
| 120 |
+
| 17 | blk.1.attn_v.weight | 0x274af720 | 0x240000 |
|
| 121 |
+
| 18 | blk.1.ffn_down.weight | 0x276ef720 | 0x2680000 |
|
| 122 |
+
| 19 | blk.1.ffn_gate.weight | 0x29d6f720 | 0x1810000 |
|
| 123 |
+
| 20 | blk.1.ffn_norm.weight | 0x2b57f720 | 0x4000 |
|
| 124 |
+
| 21 | blk.1.ffn_up.weight | 0x2b583720 | 0x1810000 |
|
| 125 |
+
| 22 | blk.2.attn_k.weight | 0x2cd93720 | 0x1b8000 |
|
| 126 |
+
| 23 | blk.2.attn_norm.weight | 0x2cf4b720 | 0x4000 |
|
| 127 |
+
| 24 | blk.2.attn_output.weight | 0x2cf4f720 | 0x900000 |
|
| 128 |
+
| 25 | blk.2.attn_q.weight | 0x2d84f720 | 0x6e0000 |
|
| 129 |
+
| 26 | blk.2.attn_v.weight | 0x2df2f720 | 0x240000 |
|
| 130 |
+
| 27 | blk.2.ffn_down.weight | 0x2e16f720 | 0x2680000 |
|
| 131 |
+
| 28 | blk.2.ffn_gate.weight | 0x307ef720 | 0x1810000 |
|
| 132 |
+
| 29 | blk.2.ffn_norm.weight | 0x31fff720 | 0x4000 |
|
| 133 |
+
| 30 | blk.2.ffn_up.weight | 0x32003720 | 0x1810000 |
|
| 134 |
+
| 31 | blk.3.attn_k.weight | 0x33813720 | 0x1b8000 |
|
| 135 |
+
| 32 | blk.3.attn_norm.weight | 0x339cb720 | 0x4000 |
|
| 136 |
+
| 33 | blk.3.attn_output.weight | 0x339cf720 | 0x900000 |
|
| 137 |
+
| 34 | blk.3.attn_q.weight | 0x342cf720 | 0x6e0000 |
|
| 138 |
+
| 35 | blk.3.attn_v.weight | 0x349af720 | 0x240000 |
|
| 139 |
+
| 36 | blk.3.ffn_down.weight | 0x34bef720 | 0x2680000 |
|
| 140 |
+
| 37 | blk.3.ffn_gate.weight | 0x3726f720 | 0x1810000 |
|
| 141 |
+
| 38 | blk.3.ffn_norm.weight | 0x38a7f720 | 0x4000 |
|
| 142 |
+
| 39 | blk.3.ffn_up.weight | 0x38a83720 | 0x1810000 |
|
| 143 |
+
| 40 | blk.4.attn_k.weight | 0x3a293720 | 0x1b8000 |
|
| 144 |
+
| 41 | blk.4.attn_norm.weight | 0x3a44b720 | 0x4000 |
|
| 145 |
+
| 42 | blk.4.attn_output.weight | 0x3a44f720 | 0x900000 |
|
| 146 |
+
| 43 | blk.4.attn_q.weight | 0x3ad4f720 | 0x6e0000 |
|
| 147 |
+
| 44 | blk.4.attn_v.weight | 0x3b42f720 | 0x240000 |
|
| 148 |
+
| 45 | blk.4.ffn_down.weight | 0x3b66f720 | 0x2680000 |
|
| 149 |
+
| 46 | blk.4.ffn_gate.weight | 0x3dcef720 | 0x1810000 |
|
| 150 |
+
| 47 | blk.4.ffn_norm.weight | 0x3f4ff720 | 0x4000 |
|
| 151 |
+
| 48 | blk.4.ffn_up.weight | 0x3f503720 | 0x1810000 |
|
| 152 |
+
| 49 | blk.5.attn_k.weight | 0x40d13720 | 0x1b8000 |
|
| 153 |
+
| 50 | blk.5.attn_norm.weight | 0x40ecb720 | 0x4000 |
|
| 154 |
+
| 51 | blk.5.attn_output.weight | 0x40ecf720 | 0x900000 |
|
| 155 |
+
| 52 | blk.5.attn_q.weight | 0x417cf720 | 0x6e0000 |
|
| 156 |
+
| 53 | blk.5.attn_v.weight | 0x41eaf720 | 0x240000 |
|
| 157 |
+
| 54 | blk.5.ffn_down.weight | 0x420ef720 | 0x2680000 |
|
| 158 |
+
| 55 | blk.5.ffn_gate.weight | 0x4476f720 | 0x1810000 |
|
| 159 |
+
| 56 | blk.5.ffn_norm.weight | 0x45f7f720 | 0x4000 |
|
| 160 |
+
| 57 | blk.5.ffn_up.weight | 0x45f83720 | 0x1810000 |
|
| 161 |
+
| 58 | blk.6.attn_k.weight | 0x47793720 | 0x1b8000 |
|
| 162 |
+
| 59 | blk.6.attn_norm.weight | 0x4794b720 | 0x4000 |
|
| 163 |
+
| 60 | blk.6.attn_output.weight | 0x4794f720 | 0x900000 |
|
| 164 |
+
| 61 | blk.6.attn_q.weight | 0x4824f720 | 0x6e0000 |
|
| 165 |
+
| 62 | blk.6.attn_v.weight | 0x4892f720 | 0x240000 |
|
| 166 |
+
| 63 | blk.6.ffn_down.weight | 0x48b6f720 | 0x2680000 |
|
| 167 |
+
| 64 | blk.6.ffn_gate.weight | 0x4b1ef720 | 0x1810000 |
|
| 168 |
+
| 65 | blk.6.ffn_norm.weight | 0x4c9ff720 | 0x4000 |
|
| 169 |
+
| 66 | blk.6.ffn_up.weight | 0x4ca03720 | 0x1810000 |
|
| 170 |
+
| 67 | blk.7.attn_k.weight | 0x4e213720 | 0x1b8000 |
|
| 171 |
+
| 68 | blk.7.attn_norm.weight | 0x4e3cb720 | 0x4000 |
|
| 172 |
+
| 69 | blk.7.attn_output.weight | 0x4e3cf720 | 0x900000 |
|
| 173 |
+
| 70 | blk.7.attn_q.weight | 0x4eccf720 | 0x6e0000 |
|
| 174 |
+
| 71 | blk.7.attn_v.weight | 0x4f3af720 | 0x240000 |
|
| 175 |
+
| 72 | blk.7.ffn_down.weight | 0x4f5ef720 | 0x2680000 |
|
| 176 |
+
| 73 | blk.7.ffn_gate.weight | 0x51c6f720 | 0x1810000 |
|
| 177 |
+
| 74 | blk.7.ffn_norm.weight | 0x5347f720 | 0x4000 |
|
| 178 |
+
| 75 | blk.7.ffn_up.weight | 0x53483720 | 0x1810000 |
|
| 179 |
+
| 76 | blk.8.attn_k.weight | 0x54c93720 | 0x1b8000 |
|
| 180 |
+
| 77 | blk.8.attn_norm.weight | 0x54e4b720 | 0x4000 |
|
| 181 |
+
| 78 | blk.8.attn_output.weight | 0x54e4f720 | 0x900000 |
|
| 182 |
+
| 79 | blk.8.attn_q.weight | 0x5574f720 | 0x6e0000 |
|
| 183 |
+
| 80 | blk.8.attn_v.weight | 0x55e2f720 | 0x240000 |
|
| 184 |
+
| 81 | blk.8.ffn_down.weight | 0x5606f720 | 0x2680000 |
|
| 185 |
+
| 82 | blk.8.ffn_gate.weight | 0x586ef720 | 0x1810000 |
|
| 186 |
+
| 83 | blk.8.ffn_norm.weight | 0x59eff720 | 0x4000 |
|
| 187 |
+
| 84 | blk.8.ffn_up.weight | 0x59f03720 | 0x1810000 |
|
| 188 |
+
| 85 | blk.9.attn_k.weight | 0x5b713720 | 0x1b8000 |
|
| 189 |
+
| 86 | blk.9.attn_norm.weight | 0x5b8cb720 | 0x4000 |
|
| 190 |
+
| 87 | blk.9.attn_output.weight | 0x5b8cf720 | 0x900000 |
|
| 191 |
+
| 88 | blk.9.attn_q.weight | 0x5c1cf720 | 0x6e0000 |
|
| 192 |
+
| 89 | blk.9.attn_v.weight | 0x5c8af720 | 0x240000 |
|
| 193 |
+
| 90 | blk.9.ffn_down.weight | 0x5caef720 | 0x2680000 |
|
| 194 |
+
| 91 | blk.9.ffn_gate.weight | 0x5f16f720 | 0x1810000 |
|
| 195 |
+
| 92 | blk.9.ffn_norm.weight | 0x6097f720 | 0x4000 |
|
| 196 |
+
| 93 | blk.9.ffn_up.weight | 0x60983720 | 0x1810000 |
|
| 197 |
+
| 94 | blk.10.attn_k.weight | 0x62193720 | 0x1b8000 |
|
| 198 |
+
| 95 | blk.10.attn_norm.weight | 0x6234b720 | 0x4000 |
|
| 199 |
+
| 96 | blk.10.attn_output.weight | 0x6234f720 | 0x900000 |
|
| 200 |
+
| 97 | blk.10.attn_q.weight | 0x62c4f720 | 0x6e0000 |
|
| 201 |
+
| 98 | blk.10.attn_v.weight | 0x6332f720 | 0x240000 |
|
| 202 |
+
| 99 | blk.10.ffn_down.weight | 0x6356f720 | 0x2680000 |
|
| 203 |
+
| 100 | blk.10.ffn_gate.weight | 0x65bef720 | 0x1810000 |
|
| 204 |
+
| 101 | blk.10.ffn_norm.weight | 0x673ff720 | 0x4000 |
|
| 205 |
+
| 102 | blk.10.ffn_up.weight | 0x67403720 | 0x1810000 |
|
| 206 |
+
| 103 | blk.11.attn_k.weight | 0x68c13720 | 0x1b8000 |
|
| 207 |
+
| 104 | blk.11.attn_norm.weight | 0x68dcb720 | 0x4000 |
|
| 208 |
+
| 105 | blk.11.attn_output.weight | 0x68dcf720 | 0x900000 |
|
| 209 |
+
| 106 | blk.11.attn_q.weight | 0x696cf720 | 0x6e0000 |
|
| 210 |
+
| 107 | blk.11.attn_v.weight | 0x69daf720 | 0x240000 |
|
| 211 |
+
| 108 | blk.11.ffn_down.weight | 0x69fef720 | 0x2680000 |
|
| 212 |
+
| 109 | blk.11.ffn_gate.weight | 0x6c66f720 | 0x1810000 |
|
| 213 |
+
| 110 | blk.11.ffn_norm.weight | 0x6de7f720 | 0x4000 |
|
| 214 |
+
| 111 | blk.11.ffn_up.weight | 0x6de83720 | 0x1810000 |
|
| 215 |
+
| 112 | blk.12.attn_k.weight | 0x6f693720 | 0x1b8000 |
|
| 216 |
+
| 113 | blk.12.attn_norm.weight | 0x6f84b720 | 0x4000 |
|
| 217 |
+
| 114 | blk.12.attn_output.weight | 0x6f84f720 | 0x900000 |
|
| 218 |
+
| 115 | blk.12.attn_q.weight | 0x7014f720 | 0x6e0000 |
|
| 219 |
+
| 116 | blk.12.attn_v.weight | 0x7082f720 | 0x240000 |
|
| 220 |
+
| 117 | blk.12.ffn_down.weight | 0x70a6f720 | 0x2680000 |
|
| 221 |
+
| 118 | blk.12.ffn_gate.weight | 0x730ef720 | 0x1810000 |
|
| 222 |
+
| 119 | blk.12.ffn_norm.weight | 0x748ff720 | 0x4000 |
|
| 223 |
+
| 120 | blk.12.ffn_up.weight | 0x74903720 | 0x1810000 |
|
| 224 |
+
| 121 | blk.13.attn_k.weight | 0x76113720 | 0x240000 |
|
| 225 |
+
| 122 | blk.13.attn_norm.weight | 0x76353720 | 0x4000 |
|
| 226 |
+
| 123 | blk.13.attn_output.weight | 0x76357720 | 0x900000 |
|
| 227 |
+
| 124 | blk.13.attn_q.weight | 0x76c57720 | 0x900000 |
|
| 228 |
+
| 125 | blk.13.attn_v.weight | 0x77557720 | 0x2c0000 |
|
| 229 |
+
| 126 | blk.13.ffn_down.weight | 0x77817720 | 0x2680000 |
|
| 230 |
+
| 127 | blk.13.ffn_gate.weight | 0x79e97720 | 0x1810000 |
|
| 231 |
+
| 128 | blk.13.ffn_norm.weight | 0x7b6a7720 | 0x4000 |
|
| 232 |
+
| 129 | blk.13.ffn_up.weight | 0x7b6ab720 | 0x1810000 |
|
| 233 |
+
| 130 | blk.14.attn_k.weight | 0x7cebb720 | 0x240000 |
|
| 234 |
+
| 131 | blk.14.attn_norm.weight | 0x7d0fb720 | 0x4000 |
|
| 235 |
+
| 132 | blk.14.attn_output.weight | 0x7d0ff720 | 0x900000 |
|
| 236 |
+
| 133 | blk.14.attn_q.weight | 0x7d9ff720 | 0x900000 |
|
| 237 |
+
| 134 | blk.14.attn_v.weight | 0x7e2ff720 | 0x2c0000 |
|
| 238 |
+
| 135 | blk.14.ffn_down.weight | 0x7e5bf720 | 0x2680000 |
|
| 239 |
+
| 136 | blk.14.ffn_gate.weight | 0x80c3f720 | 0x1810000 |
|
| 240 |
+
| 137 | blk.14.ffn_norm.weight | 0x8244f720 | 0x4000 |
|
| 241 |
+
| 138 | blk.14.ffn_up.weight | 0x82453720 | 0x1810000 |
|
| 242 |
+
| 139 | blk.15.attn_k.weight | 0x83c63720 | 0x1b8000 |
|
| 243 |
+
| 140 | blk.15.attn_norm.weight | 0x83e1b720 | 0x4000 |
|
| 244 |
+
| 141 | blk.15.attn_output.weight | 0x83e1f720 | 0x900000 |
|
| 245 |
+
| 142 | blk.15.attn_q.weight | 0x8471f720 | 0x6e0000 |
|
| 246 |
+
| 143 | blk.15.attn_v.weight | 0x84dff720 | 0x240000 |
|
| 247 |
+
| 144 | blk.15.ffn_down.weight | 0x8503f720 | 0x2680000 |
|
| 248 |
+
| 145 | blk.15.ffn_gate.weight | 0x876bf720 | 0x1810000 |
|
| 249 |
+
| 146 | blk.15.ffn_norm.weight | 0x88ecf720 | 0x4000 |
|
| 250 |
+
| 147 | blk.15.ffn_up.weight | 0x88ed3720 | 0x1810000 |
|
| 251 |
+
| 148 | blk.16.attn_k.weight | 0x8a6e3720 | 0x240000 |
|
| 252 |
+
| 149 | blk.16.attn_norm.weight | 0x8a923720 | 0x4000 |
|
| 253 |
+
| 150 | blk.16.attn_output.weight | 0x8a927720 | 0x900000 |
|
| 254 |
+
| 151 | blk.16.attn_q.weight | 0x8b227720 | 0x900000 |
|
| 255 |
+
| 152 | blk.16.attn_v.weight | 0x8bb27720 | 0x2c0000 |
|
| 256 |
+
| 153 | blk.16.ffn_down.weight | 0x8bde7720 | 0x2680000 |
|
| 257 |
+
| 154 | blk.16.ffn_gate.weight | 0x8e467720 | 0x1f80000 |
|
| 258 |
+
| 155 | blk.16.ffn_norm.weight | 0x903e7720 | 0x4000 |
|
| 259 |
+
| 156 | blk.16.ffn_up.weight | 0x903eb720 | 0x1f80000 |
|
| 260 |
+
| 157 | blk.17.attn_k.weight | 0x9236b720 | 0x1b8000 |
|
| 261 |
+
| 158 | blk.17.attn_norm.weight | 0x92523720 | 0x4000 |
|
| 262 |
+
| 159 | blk.17.attn_output.weight | 0x92527720 | 0x900000 |
|
| 263 |
+
| 160 | blk.17.attn_q.weight | 0x92e27720 | 0x6e0000 |
|
| 264 |
+
| 161 | blk.17.attn_v.weight | 0x93507720 | 0x240000 |
|
| 265 |
+
| 162 | blk.17.ffn_down.weight | 0x93747720 | 0x2680000 |
|
| 266 |
+
| 163 | blk.17.ffn_gate.weight | 0x95dc7720 | 0x1f80000 |
|
| 267 |
+
| 164 | blk.17.ffn_norm.weight | 0x97d47720 | 0x4000 |
|
| 268 |
+
| 165 | blk.17.ffn_up.weight | 0x97d4b720 | 0x1f80000 |
|
| 269 |
+
| 166 | blk.18.attn_k.weight | 0x99ccb720 | 0x240000 |
|
| 270 |
+
| 167 | blk.18.attn_norm.weight | 0x99f0b720 | 0x4000 |
|
| 271 |
+
| 168 | blk.18.attn_output.weight | 0x99f0f720 | 0x900000 |
|
| 272 |
+
| 169 | blk.18.attn_q.weight | 0x9a80f720 | 0x900000 |
|
| 273 |
+
| 170 | blk.18.attn_v.weight | 0x9b10f720 | 0x2c0000 |
|
| 274 |
+
| 171 | blk.18.ffn_down.weight | 0x9b3cf720 | 0x2680000 |
|
| 275 |
+
| 172 | blk.18.ffn_gate.weight | 0x9da4f720 | 0x1f80000 |
|
| 276 |
+
| 173 | blk.18.ffn_norm.weight | 0x9f9cf720 | 0x4000 |
|
| 277 |
+
| 174 | blk.18.ffn_up.weight | 0x9f9d3720 | 0x1f80000 |
|
| 278 |
+
| 175 | blk.19.attn_k.weight | 0xa1953720 | 0x240000 |
|
| 279 |
+
| 176 | blk.19.attn_norm.weight | 0xa1b93720 | 0x4000 |
|
| 280 |
+
| 177 | blk.19.attn_output.weight | 0xa1b97720 | 0x900000 |
|
| 281 |
+
| 178 | blk.19.attn_q.weight | 0xa2497720 | 0x900000 |
|
| 282 |
+
| 179 | blk.19.attn_v.weight | 0xa2d97720 | 0x2c0000 |
|
| 283 |
+
| 180 | blk.19.ffn_down.weight | 0xa3057720 | 0x2680000 |
|
| 284 |
+
| 181 | blk.19.ffn_gate.weight | 0xa56d7720 | 0x1f80000 |
|
| 285 |
+
| 182 | blk.19.ffn_norm.weight | 0xa7657720 | 0x4000 |
|
| 286 |
+
| 183 | blk.19.ffn_up.weight | 0xa765b720 | 0x1f80000 |
|
| 287 |
+
| 184 | blk.20.attn_k.weight | 0xa95db720 | 0x240000 |
|
| 288 |
+
| 185 | blk.20.attn_norm.weight | 0xa981b720 | 0x4000 |
|
| 289 |
+
| 186 | blk.20.attn_output.weight | 0xa981f720 | 0x900000 |
|
| 290 |
+
| 187 | blk.20.attn_q.weight | 0xaa11f720 | 0x900000 |
|
| 291 |
+
| 188 | blk.20.attn_v.weight | 0xaaa1f720 | 0x2c0000 |
|
| 292 |
+
| 189 | blk.20.ffn_down.weight | 0xaacdf720 | 0x2680000 |
|
| 293 |
+
| 190 | blk.20.ffn_gate.weight | 0xad35f720 | 0x1f80000 |
|
| 294 |
+
| 191 | blk.20.ffn_norm.weight | 0xaf2df720 | 0x4000 |
|
| 295 |
+
| 192 | blk.20.ffn_up.weight | 0xaf2e3720 | 0x1f80000 |
|
| 296 |
+
| 193 | blk.21.attn_k.weight | 0xb1263720 | 0x240000 |
|
| 297 |
+
| 194 | blk.21.attn_norm.weight | 0xb14a3720 | 0x4000 |
|
| 298 |
+
| 195 | blk.21.attn_output.weight | 0xb14a7720 | 0x900000 |
|
| 299 |
+
| 196 | blk.21.attn_q.weight | 0xb1da7720 | 0x900000 |
|
| 300 |
+
| 197 | blk.21.attn_v.weight | 0xb26a7720 | 0x2c0000 |
|
| 301 |
+
| 198 | blk.21.ffn_down.weight | 0xb2967720 | 0x2680000 |
|
| 302 |
+
| 199 | blk.21.ffn_gate.weight | 0xb4fe7720 | 0x1f80000 |
|
| 303 |
+
| 200 | blk.21.ffn_norm.weight | 0xb6f67720 | 0x4000 |
|
| 304 |
+
| 201 | blk.21.ffn_up.weight | 0xb6f6b720 | 0x1f80000 |
|
| 305 |
+
| 202 | blk.22.attn_k.weight | 0xb8eeb720 | 0x240000 |
|
| 306 |
+
| 203 | blk.22.attn_norm.weight | 0xb912b720 | 0x4000 |
|
| 307 |
+
| 204 | blk.22.attn_output.weight | 0xb912f720 | 0x900000 |
|
| 308 |
+
| 205 | blk.22.attn_q.weight | 0xb9a2f720 | 0x900000 |
|
| 309 |
+
| 206 | blk.22.attn_v.weight | 0xba32f720 | 0x2c0000 |
|
| 310 |
+
| 207 | blk.22.ffn_down.weight | 0xba5ef720 | 0x2680000 |
|
| 311 |
+
| 208 | blk.22.ffn_gate.weight | 0xbcc6f720 | 0x1f80000 |
|
| 312 |
+
| 209 | blk.22.ffn_norm.weight | 0xbebef720 | 0x4000 |
|
| 313 |
+
| 210 | blk.22.ffn_up.weight | 0xbebf3720 | 0x1f80000 |
|
| 314 |
+
| 211 | blk.23.attn_k.weight | 0xc0b73720 | 0x240000 |
|
| 315 |
+
| 212 | blk.23.attn_norm.weight | 0xc0db3720 | 0x4000 |
|
| 316 |
+
| 213 | blk.23.attn_output.weight | 0xc0db7720 | 0x900000 |
|
| 317 |
+
| 214 | blk.23.attn_q.weight | 0xc16b7720 | 0x900000 |
|
| 318 |
+
| 215 | blk.23.attn_v.weight | 0xc1fb7720 | 0x2c0000 |
|
| 319 |
+
| 216 | blk.23.ffn_down.weight | 0xc2277720 | 0x2680000 |
|
| 320 |
+
| 217 | blk.23.ffn_gate.weight | 0xc48f7720 | 0x1f80000 |
|
| 321 |
+
| 218 | blk.23.ffn_norm.weight | 0xc6877720 | 0x4000 |
|
| 322 |
+
| 219 | blk.23.ffn_up.weight | 0xc687b720 | 0x1f80000 |
|
| 323 |
+
| 220 | blk.24.attn_k.weight | 0xc87fb720 | 0x240000 |
|
| 324 |
+
| 221 | blk.24.attn_norm.weight | 0xc8a3b720 | 0x4000 |
|
| 325 |
+
| 222 | blk.24.attn_output.weight | 0xc8a3f720 | 0x900000 |
|
| 326 |
+
| 223 | blk.24.attn_q.weight | 0xc933f720 | 0x900000 |
|
| 327 |
+
| 224 | blk.24.attn_v.weight | 0xc9c3f720 | 0x2c0000 |
|
| 328 |
+
| 225 | blk.24.ffn_down.weight | 0xc9eff720 | 0x2680000 |
|
| 329 |
+
| 226 | blk.24.ffn_gate.weight | 0xcc57f720 | 0x1f80000 |
|
| 330 |
+
| 227 | blk.24.ffn_norm.weight | 0xce4ff720 | 0x4000 |
|
| 331 |
+
| 228 | blk.24.ffn_up.weight | 0xce503720 | 0x1f80000 |
|
| 332 |
+
| 229 | blk.25.attn_k.weight | 0xd0483720 | 0x240000 |
|
| 333 |
+
| 230 | blk.25.attn_norm.weight | 0xd06c3720 | 0x4000 |
|
| 334 |
+
| 231 | blk.25.attn_output.weight | 0xd06c7720 | 0x900000 |
|
| 335 |
+
| 232 | blk.25.attn_q.weight | 0xd0fc7720 | 0x900000 |
|
| 336 |
+
| 233 | blk.25.attn_v.weight | 0xd18c7720 | 0x2c0000 |
|
| 337 |
+
| 234 | blk.25.ffn_down.weight | 0xd1b87720 | 0x2680000 |
|
| 338 |
+
| 235 | blk.25.ffn_gate.weight | 0xd4207720 | 0x1f80000 |
|
| 339 |
+
| 236 | blk.25.ffn_norm.weight | 0xd6187720 | 0x4000 |
|
| 340 |
+
| 237 | blk.25.ffn_up.weight | 0xd618b720 | 0x1f80000 |
|
| 341 |
+
| 238 | blk.26.attn_k.weight | 0xd810b720 | 0x240000 |
|
| 342 |
+
| 239 | blk.26.attn_norm.weight | 0xd834b720 | 0x4000 |
|
| 343 |
+
| 240 | blk.26.attn_output.weight | 0xd834f720 | 0x900000 |
|
| 344 |
+
| 241 | blk.26.attn_q.weight | 0xd8c4f720 | 0x900000 |
|
| 345 |
+
| 242 | blk.26.attn_v.weight | 0xd954f720 | 0x2c0000 |
|
| 346 |
+
| 243 | blk.26.ffn_down.weight | 0xd980f720 | 0x2680000 |
|
| 347 |
+
| 244 | blk.26.ffn_gate.weight | 0xdbe8f720 | 0x1f80000 |
|
| 348 |
+
| 245 | blk.26.ffn_norm.weight | 0xdde0f720 | 0x4000 |
|
| 349 |
+
| 246 | blk.26.ffn_up.weight | 0xdde13720 | 0x1f80000 |
|
| 350 |
+
| 247 | blk.27.attn_k.weight | 0xdfd93720 | 0x240000 |
|
| 351 |
+
| 248 | blk.27.attn_norm.weight | 0xdffd3720 | 0x4000 |
|
| 352 |
+
| 249 | blk.27.attn_output.weight | 0xdffd7720 | 0x900000 |
|
| 353 |
+
| 250 | blk.27.attn_q.weight | 0xe08d7720 | 0x900000 |
|
| 354 |
+
| 251 | blk.27.attn_v.weight | 0xe11d7720 | 0x2c0000 |
|
| 355 |
+
| 252 | blk.27.ffn_down.weight | 0xe1497720 | 0x2680000 |
|
| 356 |
+
| 253 | blk.27.ffn_gate.weight | 0xe3b17720 | 0x1f80000 |
|
| 357 |
+
| 254 | blk.27.ffn_norm.weight | 0xe5a97720 | 0x4000 |
|
| 358 |
+
| 255 | blk.27.ffn_up.weight | 0xe5a9b720 | 0x1f80000 |
|
| 359 |
+
| 256 | blk.28.attn_k.weight | 0xe7a1b720 | 0x240000 |
|
| 360 |
+
| 257 | blk.28.attn_norm.weight | 0xe7c5b720 | 0x4000 |
|
| 361 |
+
| 258 | blk.28.attn_output.weight | 0xe7c5f720 | 0x900000 |
|
| 362 |
+
| 259 | blk.28.attn_q.weight | 0xe855f720 | 0x900000 |
|
| 363 |
+
| 260 | blk.28.attn_v.weight | 0xe8e5f720 | 0x2c0000 |
|
| 364 |
+
| 261 | blk.28.ffn_down.weight | 0xe911f720 | 0x2680000 |
|
| 365 |
+
| 262 | blk.28.ffn_gate.weight | 0xeb79f720 | 0x1f80000 |
|
| 366 |
+
| 263 | blk.28.ffn_norm.weight | 0xed71f720 | 0x4000 |
|
| 367 |
+
| 264 | blk.28.ffn_up.weight | 0xed723720 | 0x1f80000 |
|
| 368 |
+
| 265 | blk.29.attn_k.weight | 0xef6a3720 | 0x240000 |
|
| 369 |
+
| 266 | blk.29.attn_norm.weight | 0xef8e3720 | 0x4000 |
|
| 370 |
+
| 267 | blk.29.attn_output.weight | 0xef8e7720 | 0x900000 |
|
| 371 |
+
| 268 | blk.29.attn_q.weight | 0xf01e7720 | 0x900000 |
|
| 372 |
+
| 269 | blk.29.attn_v.weight | 0xf0ae7720 | 0x2c0000 |
|
| 373 |
+
| 270 | blk.29.ffn_down.weight | 0xf0da7720 | 0x2680000 |
|
| 374 |
+
| 271 | blk.29.ffn_gate.weight | 0xf3427720 | 0x1f80000 |
|
| 375 |
+
| 272 | blk.29.ffn_norm.weight | 0xf53a7720 | 0x4000 |
|
| 376 |
+
| 273 | blk.29.ffn_up.weight | 0xf53ab720 | 0x1f80000 |
|
| 377 |
+
| 274 | blk.30.attn_k.weight | 0xf732b720 | 0x240000 |
|
| 378 |
+
| 275 | blk.30.attn_norm.weight | 0xf756b720 | 0x4000 |
|
| 379 |
+
| 276 | blk.30.attn_output.weight | 0xf756f720 | 0x900000 |
|
| 380 |
+
| 277 | blk.30.attn_q.weight | 0xf7e6f720 | 0x900000 |
|
| 381 |
+
| 278 | blk.30.attn_v.weight | 0xf876f720 | 0x2c0000 |
|
| 382 |
+
| 279 | blk.30.ffn_down.weight | 0xf8a2f720 | 0x2680000 |
|
| 383 |
+
| 280 | blk.30.ffn_gate.weight | 0xfb0af720 | 0x1f80000 |
|
| 384 |
+
| 281 | blk.30.ffn_norm.weight | 0xfd02f720 | 0x4000 |
|
| 385 |
+
| 282 | blk.30.ffn_up.weight | 0xfd033720 | 0x1f80000 |
|
| 386 |
+
| 283 | blk.31.attn_k.weight | 0xfefb3720 | 0x1b8000 |
|
| 387 |
+
| 284 | blk.31.attn_norm.weight | 0xff16b720 | 0x4000 |
|
| 388 |
+
| 285 | blk.31.attn_output.weight | 0xff16f720 | 0x900000 |
|
| 389 |
+
| 286 | blk.31.attn_q.weight | 0xffa6f720 | 0x6e0000 |
|
| 390 |
+
| 287 | blk.31.attn_v.weight | 0x10014f720 | 0x240000 |
|
| 391 |
+
| 288 | blk.31.ffn_down.weight | 0x10038f720 | 0x2680000 |
|
| 392 |
+
| 289 | blk.31.ffn_gate.weight | 0x102a0f720 | 0x1f80000 |
|
| 393 |
+
| 290 | blk.31.ffn_norm.weight | 0x10498f720 | 0x4000 |
|
| 394 |
+
| 291 | blk.31.ffn_up.weight | 0x104993720 | 0x1f80000 |
|
| 395 |
+
|
| 396 |
+
### <a name="base">Base Tensor Group : ~1B Elements</a>
|
| 397 |
+
|
| 398 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 399 |
+
|-----:|:-------------------|:---------------------------------|:------------------|:----------------------|:-----|
|
| 400 |
+
| 0 | output.weight | Output (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | Q4_K |
|
| 401 |
+
| 1 | output_norm.weight | Output Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 402 |
+
| 2 | rope_freqs.weight | Rope_Freqs (W) | ( 64) 64 | 64 x 1 x 1 x 1 | F32 |
|
| 403 |
+
| 3 | token_embd.weight | Token Embedding (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | Q3_K |
|
| 404 |
+
|
| 405 |
+
- Total elements in base: ( ~1B) 1050677312
|
| 406 |
+
- Percentage of total elements: 13.08%
|
| 407 |
+
|
| 408 |
+
|
| 409 |
+
### <a name="blk_0">Block 0 Tensor Group : ~218M Elements</a>
|
| 410 |
+
|
| 411 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 412 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 413 |
+
| 4 | blk.0.attn_k.weight | Block 0 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 414 |
+
| 5 | blk.0.attn_norm.weight | Block 0 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 415 |
+
| 6 | blk.0.attn_output.weight | Block 0 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 416 |
+
| 7 | blk.0.attn_q.weight | Block 0 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 417 |
+
| 8 | blk.0.attn_v.weight | Block 0 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 418 |
+
| 9 | blk.0.ffn_down.weight | Block 0 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 419 |
+
| 10 | blk.0.ffn_gate.weight | Block 0 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 420 |
+
| 11 | blk.0.ffn_norm.weight | Block 0 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 421 |
+
| 12 | blk.0.ffn_up.weight | Block 0 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 422 |
+
|
| 423 |
+
- Total elements in blk.0: (~218M) 218112000
|
| 424 |
+
- Percentage of total elements: 2.72%
|
| 425 |
+
|
| 426 |
+
|
| 427 |
+
### <a name="blk_1">Block 1 Tensor Group : ~218M Elements</a>
|
| 428 |
+
|
| 429 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 430 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 431 |
+
| 13 | blk.1.attn_k.weight | Block 1 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 432 |
+
| 14 | blk.1.attn_norm.weight | Block 1 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 433 |
+
| 15 | blk.1.attn_output.weight | Block 1 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 434 |
+
| 16 | blk.1.attn_q.weight | Block 1 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 435 |
+
| 17 | blk.1.attn_v.weight | Block 1 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 436 |
+
| 18 | blk.1.ffn_down.weight | Block 1 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 437 |
+
| 19 | blk.1.ffn_gate.weight | Block 1 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 438 |
+
| 20 | blk.1.ffn_norm.weight | Block 1 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 439 |
+
| 21 | blk.1.ffn_up.weight | Block 1 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 440 |
+
|
| 441 |
+
- Total elements in blk.1: (~218M) 218112000
|
| 442 |
+
- Percentage of total elements: 2.72%
|
| 443 |
+
|
| 444 |
+
|
| 445 |
+
### <a name="blk_2">Block 2 Tensor Group : ~218M Elements</a>
|
| 446 |
+
|
| 447 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 448 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 449 |
+
| 22 | blk.2.attn_k.weight | Block 2 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 450 |
+
| 23 | blk.2.attn_norm.weight | Block 2 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 451 |
+
| 24 | blk.2.attn_output.weight | Block 2 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 452 |
+
| 25 | blk.2.attn_q.weight | Block 2 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 453 |
+
| 26 | blk.2.attn_v.weight | Block 2 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 454 |
+
| 27 | blk.2.ffn_down.weight | Block 2 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 455 |
+
| 28 | blk.2.ffn_gate.weight | Block 2 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 456 |
+
| 29 | blk.2.ffn_norm.weight | Block 2 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 457 |
+
| 30 | blk.2.ffn_up.weight | Block 2 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 458 |
+
|
| 459 |
+
- Total elements in blk.2: (~218M) 218112000
|
| 460 |
+
- Percentage of total elements: 2.72%
|
| 461 |
+
|
| 462 |
+
|
| 463 |
+
### <a name="blk_3">Block 3 Tensor Group : ~218M Elements</a>
|
| 464 |
+
|
| 465 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 466 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 467 |
+
| 31 | blk.3.attn_k.weight | Block 3 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 468 |
+
| 32 | blk.3.attn_norm.weight | Block 3 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 469 |
+
| 33 | blk.3.attn_output.weight | Block 3 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 470 |
+
| 34 | blk.3.attn_q.weight | Block 3 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 471 |
+
| 35 | blk.3.attn_v.weight | Block 3 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 472 |
+
| 36 | blk.3.ffn_down.weight | Block 3 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 473 |
+
| 37 | blk.3.ffn_gate.weight | Block 3 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 474 |
+
| 38 | blk.3.ffn_norm.weight | Block 3 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 475 |
+
| 39 | blk.3.ffn_up.weight | Block 3 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 476 |
+
|
| 477 |
+
- Total elements in blk.3: (~218M) 218112000
|
| 478 |
+
- Percentage of total elements: 2.72%
|
| 479 |
+
|
| 480 |
+
|
| 481 |
+
### <a name="blk_4">Block 4 Tensor Group : ~218M Elements</a>
|
| 482 |
+
|
| 483 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 484 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 485 |
+
| 40 | blk.4.attn_k.weight | Block 4 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 486 |
+
| 41 | blk.4.attn_norm.weight | Block 4 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 487 |
+
| 42 | blk.4.attn_output.weight | Block 4 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 488 |
+
| 43 | blk.4.attn_q.weight | Block 4 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 489 |
+
| 44 | blk.4.attn_v.weight | Block 4 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 490 |
+
| 45 | blk.4.ffn_down.weight | Block 4 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 491 |
+
| 46 | blk.4.ffn_gate.weight | Block 4 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 492 |
+
| 47 | blk.4.ffn_norm.weight | Block 4 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 493 |
+
| 48 | blk.4.ffn_up.weight | Block 4 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 494 |
+
|
| 495 |
+
- Total elements in blk.4: (~218M) 218112000
|
| 496 |
+
- Percentage of total elements: 2.72%
|
| 497 |
+
|
| 498 |
+
|
| 499 |
+
### <a name="blk_5">Block 5 Tensor Group : ~218M Elements</a>
|
| 500 |
+
|
| 501 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 502 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 503 |
+
| 49 | blk.5.attn_k.weight | Block 5 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 504 |
+
| 50 | blk.5.attn_norm.weight | Block 5 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 505 |
+
| 51 | blk.5.attn_output.weight | Block 5 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 506 |
+
| 52 | blk.5.attn_q.weight | Block 5 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 507 |
+
| 53 | blk.5.attn_v.weight | Block 5 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 508 |
+
| 54 | blk.5.ffn_down.weight | Block 5 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 509 |
+
| 55 | blk.5.ffn_gate.weight | Block 5 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 510 |
+
| 56 | blk.5.ffn_norm.weight | Block 5 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 511 |
+
| 57 | blk.5.ffn_up.weight | Block 5 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 512 |
+
|
| 513 |
+
- Total elements in blk.5: (~218M) 218112000
|
| 514 |
+
- Percentage of total elements: 2.72%
|
| 515 |
+
|
| 516 |
+
|
| 517 |
+
### <a name="blk_6">Block 6 Tensor Group : ~218M Elements</a>
|
| 518 |
+
|
| 519 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 520 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 521 |
+
| 58 | blk.6.attn_k.weight | Block 6 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 522 |
+
| 59 | blk.6.attn_norm.weight | Block 6 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 523 |
+
| 60 | blk.6.attn_output.weight | Block 6 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 524 |
+
| 61 | blk.6.attn_q.weight | Block 6 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 525 |
+
| 62 | blk.6.attn_v.weight | Block 6 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 526 |
+
| 63 | blk.6.ffn_down.weight | Block 6 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 527 |
+
| 64 | blk.6.ffn_gate.weight | Block 6 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 528 |
+
| 65 | blk.6.ffn_norm.weight | Block 6 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 529 |
+
| 66 | blk.6.ffn_up.weight | Block 6 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 530 |
+
|
| 531 |
+
- Total elements in blk.6: (~218M) 218112000
|
| 532 |
+
- Percentage of total elements: 2.72%
|
| 533 |
+
|
| 534 |
+
|
| 535 |
+
### <a name="blk_7">Block 7 Tensor Group : ~218M Elements</a>
|
| 536 |
+
|
| 537 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 538 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 539 |
+
| 67 | blk.7.attn_k.weight | Block 7 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 540 |
+
| 68 | blk.7.attn_norm.weight | Block 7 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 541 |
+
| 69 | blk.7.attn_output.weight | Block 7 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 542 |
+
| 70 | blk.7.attn_q.weight | Block 7 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 543 |
+
| 71 | blk.7.attn_v.weight | Block 7 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 544 |
+
| 72 | blk.7.ffn_down.weight | Block 7 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 545 |
+
| 73 | blk.7.ffn_gate.weight | Block 7 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 546 |
+
| 74 | blk.7.ffn_norm.weight | Block 7 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 547 |
+
| 75 | blk.7.ffn_up.weight | Block 7 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 548 |
+
|
| 549 |
+
- Total elements in blk.7: (~218M) 218112000
|
| 550 |
+
- Percentage of total elements: 2.72%
|
| 551 |
+
|
| 552 |
+
|
| 553 |
+
### <a name="blk_8">Block 8 Tensor Group : ~218M Elements</a>
|
| 554 |
+
|
| 555 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 556 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 557 |
+
| 76 | blk.8.attn_k.weight | Block 8 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 558 |
+
| 77 | blk.8.attn_norm.weight | Block 8 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 559 |
+
| 78 | blk.8.attn_output.weight | Block 8 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 560 |
+
| 79 | blk.8.attn_q.weight | Block 8 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 561 |
+
| 80 | blk.8.attn_v.weight | Block 8 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 562 |
+
| 81 | blk.8.ffn_down.weight | Block 8 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 563 |
+
| 82 | blk.8.ffn_gate.weight | Block 8 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 564 |
+
| 83 | blk.8.ffn_norm.weight | Block 8 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 565 |
+
| 84 | blk.8.ffn_up.weight | Block 8 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 566 |
+
|
| 567 |
+
- Total elements in blk.8: (~218M) 218112000
|
| 568 |
+
- Percentage of total elements: 2.72%
|
| 569 |
+
|
| 570 |
+
|
| 571 |
+
### <a name="blk_9">Block 9 Tensor Group : ~218M Elements</a>
|
| 572 |
+
|
| 573 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 574 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 575 |
+
| 85 | blk.9.attn_k.weight | Block 9 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 576 |
+
| 86 | blk.9.attn_norm.weight | Block 9 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 577 |
+
| 87 | blk.9.attn_output.weight | Block 9 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 578 |
+
| 88 | blk.9.attn_q.weight | Block 9 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 579 |
+
| 89 | blk.9.attn_v.weight | Block 9 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 580 |
+
| 90 | blk.9.ffn_down.weight | Block 9 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 581 |
+
| 91 | blk.9.ffn_gate.weight | Block 9 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 582 |
+
| 92 | blk.9.ffn_norm.weight | Block 9 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 583 |
+
| 93 | blk.9.ffn_up.weight | Block 9 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 584 |
+
|
| 585 |
+
- Total elements in blk.9: (~218M) 218112000
|
| 586 |
+
- Percentage of total elements: 2.72%
|
| 587 |
+
|
| 588 |
+
|
| 589 |
+
### <a name="blk_10">Block 10 Tensor Group : ~218M Elements</a>
|
| 590 |
+
|
| 591 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 592 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 593 |
+
| 94 | blk.10.attn_k.weight | Block 10 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 594 |
+
| 95 | blk.10.attn_norm.weight | Block 10 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 595 |
+
| 96 | blk.10.attn_output.weight | Block 10 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 596 |
+
| 97 | blk.10.attn_q.weight | Block 10 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 597 |
+
| 98 | blk.10.attn_v.weight | Block 10 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 598 |
+
| 99 | blk.10.ffn_down.weight | Block 10 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 599 |
+
| 100 | blk.10.ffn_gate.weight | Block 10 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 600 |
+
| 101 | blk.10.ffn_norm.weight | Block 10 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 601 |
+
| 102 | blk.10.ffn_up.weight | Block 10 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 602 |
+
|
| 603 |
+
- Total elements in blk.10: (~218M) 218112000
|
| 604 |
+
- Percentage of total elements: 2.72%
|
| 605 |
+
|
| 606 |
+
|
| 607 |
+
### <a name="blk_11">Block 11 Tensor Group : ~218M Elements</a>
|
| 608 |
+
|
| 609 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 610 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 611 |
+
| 103 | blk.11.attn_k.weight | Block 11 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 612 |
+
| 104 | blk.11.attn_norm.weight | Block 11 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 613 |
+
| 105 | blk.11.attn_output.weight | Block 11 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 614 |
+
| 106 | blk.11.attn_q.weight | Block 11 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 615 |
+
| 107 | blk.11.attn_v.weight | Block 11 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 616 |
+
| 108 | blk.11.ffn_down.weight | Block 11 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 617 |
+
| 109 | blk.11.ffn_gate.weight | Block 11 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 618 |
+
| 110 | blk.11.ffn_norm.weight | Block 11 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 619 |
+
| 111 | blk.11.ffn_up.weight | Block 11 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 620 |
+
|
| 621 |
+
- Total elements in blk.11: (~218M) 218112000
|
| 622 |
+
- Percentage of total elements: 2.72%
|
| 623 |
+
|
| 624 |
+
|
| 625 |
+
### <a name="blk_12">Block 12 Tensor Group : ~218M Elements</a>
|
| 626 |
+
|
| 627 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 628 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 629 |
+
| 112 | blk.12.attn_k.weight | Block 12 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 630 |
+
| 113 | blk.12.attn_norm.weight | Block 12 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 631 |
+
| 114 | blk.12.attn_output.weight | Block 12 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 632 |
+
| 115 | blk.12.attn_q.weight | Block 12 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 633 |
+
| 116 | blk.12.attn_v.weight | Block 12 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 634 |
+
| 117 | blk.12.ffn_down.weight | Block 12 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 635 |
+
| 118 | blk.12.ffn_gate.weight | Block 12 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 636 |
+
| 119 | blk.12.ffn_norm.weight | Block 12 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 637 |
+
| 120 | blk.12.ffn_up.weight | Block 12 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 638 |
+
|
| 639 |
+
- Total elements in blk.12: (~218M) 218112000
|
| 640 |
+
- Percentage of total elements: 2.72%
|
| 641 |
+
|
| 642 |
+
|
| 643 |
+
### <a name="blk_13">Block 13 Tensor Group : ~218M Elements</a>
|
| 644 |
+
|
| 645 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 646 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 647 |
+
| 121 | blk.13.attn_k.weight | Block 13 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 648 |
+
| 122 | blk.13.attn_norm.weight | Block 13 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 649 |
+
| 123 | blk.13.attn_output.weight | Block 13 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 650 |
+
| 124 | blk.13.attn_q.weight | Block 13 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 651 |
+
| 125 | blk.13.attn_v.weight | Block 13 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 652 |
+
| 126 | blk.13.ffn_down.weight | Block 13 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 653 |
+
| 127 | blk.13.ffn_gate.weight | Block 13 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 654 |
+
| 128 | blk.13.ffn_norm.weight | Block 13 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 655 |
+
| 129 | blk.13.ffn_up.weight | Block 13 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 656 |
+
|
| 657 |
+
- Total elements in blk.13: (~218M) 218112000
|
| 658 |
+
- Percentage of total elements: 2.72%
|
| 659 |
+
|
| 660 |
+
|
| 661 |
+
### <a name="blk_14">Block 14 Tensor Group : ~218M Elements</a>
|
| 662 |
+
|
| 663 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 664 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 665 |
+
| 130 | blk.14.attn_k.weight | Block 14 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 666 |
+
| 131 | blk.14.attn_norm.weight | Block 14 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 667 |
+
| 132 | blk.14.attn_output.weight | Block 14 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 668 |
+
| 133 | blk.14.attn_q.weight | Block 14 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 669 |
+
| 134 | blk.14.attn_v.weight | Block 14 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 670 |
+
| 135 | blk.14.ffn_down.weight | Block 14 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 671 |
+
| 136 | blk.14.ffn_gate.weight | Block 14 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 672 |
+
| 137 | blk.14.ffn_norm.weight | Block 14 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 673 |
+
| 138 | blk.14.ffn_up.weight | Block 14 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 674 |
+
|
| 675 |
+
- Total elements in blk.14: (~218M) 218112000
|
| 676 |
+
- Percentage of total elements: 2.72%
|
| 677 |
+
|
| 678 |
+
|
| 679 |
+
### <a name="blk_15">Block 15 Tensor Group : ~218M Elements</a>
|
| 680 |
+
|
| 681 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 682 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 683 |
+
| 139 | blk.15.attn_k.weight | Block 15 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 684 |
+
| 140 | blk.15.attn_norm.weight | Block 15 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 685 |
+
| 141 | blk.15.attn_output.weight | Block 15 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 686 |
+
| 142 | blk.15.attn_q.weight | Block 15 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 687 |
+
| 143 | blk.15.attn_v.weight | Block 15 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 688 |
+
| 144 | blk.15.ffn_down.weight | Block 15 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 689 |
+
| 145 | blk.15.ffn_gate.weight | Block 15 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 690 |
+
| 146 | blk.15.ffn_norm.weight | Block 15 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 691 |
+
| 147 | blk.15.ffn_up.weight | Block 15 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 692 |
+
|
| 693 |
+
- Total elements in blk.15: (~218M) 218112000
|
| 694 |
+
- Percentage of total elements: 2.72%
|
| 695 |
+
|
| 696 |
+
|
| 697 |
+
### <a name="blk_16">Block 16 Tensor Group : ~218M Elements</a>
|
| 698 |
+
|
| 699 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 700 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 701 |
+
| 148 | blk.16.attn_k.weight | Block 16 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 702 |
+
| 149 | blk.16.attn_norm.weight | Block 16 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 703 |
+
| 150 | blk.16.attn_output.weight | Block 16 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 704 |
+
| 151 | blk.16.attn_q.weight | Block 16 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 705 |
+
| 152 | blk.16.attn_v.weight | Block 16 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 706 |
+
| 153 | blk.16.ffn_down.weight | Block 16 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 707 |
+
| 154 | blk.16.ffn_gate.weight | Block 16 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 708 |
+
| 155 | blk.16.ffn_norm.weight | Block 16 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 709 |
+
| 156 | blk.16.ffn_up.weight | Block 16 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 710 |
+
|
| 711 |
+
- Total elements in blk.16: (~218M) 218112000
|
| 712 |
+
- Percentage of total elements: 2.72%
|
| 713 |
+
|
| 714 |
+
|
| 715 |
+
### <a name="blk_17">Block 17 Tensor Group : ~218M Elements</a>
|
| 716 |
+
|
| 717 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 718 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 719 |
+
| 157 | blk.17.attn_k.weight | Block 17 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 720 |
+
| 158 | blk.17.attn_norm.weight | Block 17 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 721 |
+
| 159 | blk.17.attn_output.weight | Block 17 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 722 |
+
| 160 | blk.17.attn_q.weight | Block 17 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 723 |
+
| 161 | blk.17.attn_v.weight | Block 17 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 724 |
+
| 162 | blk.17.ffn_down.weight | Block 17 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 725 |
+
| 163 | blk.17.ffn_gate.weight | Block 17 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 726 |
+
| 164 | blk.17.ffn_norm.weight | Block 17 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 727 |
+
| 165 | blk.17.ffn_up.weight | Block 17 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 728 |
+
|
| 729 |
+
- Total elements in blk.17: (~218M) 218112000
|
| 730 |
+
- Percentage of total elements: 2.72%
|
| 731 |
+
|
| 732 |
+
|
| 733 |
+
### <a name="blk_18">Block 18 Tensor Group : ~218M Elements</a>
|
| 734 |
+
|
| 735 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 736 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 737 |
+
| 166 | blk.18.attn_k.weight | Block 18 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 738 |
+
| 167 | blk.18.attn_norm.weight | Block 18 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 739 |
+
| 168 | blk.18.attn_output.weight | Block 18 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 740 |
+
| 169 | blk.18.attn_q.weight | Block 18 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 741 |
+
| 170 | blk.18.attn_v.weight | Block 18 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 742 |
+
| 171 | blk.18.ffn_down.weight | Block 18 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 743 |
+
| 172 | blk.18.ffn_gate.weight | Block 18 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 744 |
+
| 173 | blk.18.ffn_norm.weight | Block 18 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 745 |
+
| 174 | blk.18.ffn_up.weight | Block 18 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 746 |
+
|
| 747 |
+
- Total elements in blk.18: (~218M) 218112000
|
| 748 |
+
- Percentage of total elements: 2.72%
|
| 749 |
+
|
| 750 |
+
|
| 751 |
+
### <a name="blk_19">Block 19 Tensor Group : ~218M Elements</a>
|
| 752 |
+
|
| 753 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 754 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 755 |
+
| 175 | blk.19.attn_k.weight | Block 19 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 756 |
+
| 176 | blk.19.attn_norm.weight | Block 19 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 757 |
+
| 177 | blk.19.attn_output.weight | Block 19 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 758 |
+
| 178 | blk.19.attn_q.weight | Block 19 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 759 |
+
| 179 | blk.19.attn_v.weight | Block 19 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 760 |
+
| 180 | blk.19.ffn_down.weight | Block 19 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 761 |
+
| 181 | blk.19.ffn_gate.weight | Block 19 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 762 |
+
| 182 | blk.19.ffn_norm.weight | Block 19 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 763 |
+
| 183 | blk.19.ffn_up.weight | Block 19 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 764 |
+
|
| 765 |
+
- Total elements in blk.19: (~218M) 218112000
|
| 766 |
+
- Percentage of total elements: 2.72%
|
| 767 |
+
|
| 768 |
+
|
| 769 |
+
### <a name="blk_20">Block 20 Tensor Group : ~218M Elements</a>
|
| 770 |
+
|
| 771 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 772 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 773 |
+
| 184 | blk.20.attn_k.weight | Block 20 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 774 |
+
| 185 | blk.20.attn_norm.weight | Block 20 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 775 |
+
| 186 | blk.20.attn_output.weight | Block 20 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 776 |
+
| 187 | blk.20.attn_q.weight | Block 20 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 777 |
+
| 188 | blk.20.attn_v.weight | Block 20 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 778 |
+
| 189 | blk.20.ffn_down.weight | Block 20 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 779 |
+
| 190 | blk.20.ffn_gate.weight | Block 20 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 780 |
+
| 191 | blk.20.ffn_norm.weight | Block 20 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 781 |
+
| 192 | blk.20.ffn_up.weight | Block 20 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 782 |
+
|
| 783 |
+
- Total elements in blk.20: (~218M) 218112000
|
| 784 |
+
- Percentage of total elements: 2.72%
|
| 785 |
+
|
| 786 |
+
|
| 787 |
+
### <a name="blk_21">Block 21 Tensor Group : ~218M Elements</a>
|
| 788 |
+
|
| 789 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 790 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 791 |
+
| 193 | blk.21.attn_k.weight | Block 21 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 792 |
+
| 194 | blk.21.attn_norm.weight | Block 21 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 793 |
+
| 195 | blk.21.attn_output.weight | Block 21 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 794 |
+
| 196 | blk.21.attn_q.weight | Block 21 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 795 |
+
| 197 | blk.21.attn_v.weight | Block 21 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 796 |
+
| 198 | blk.21.ffn_down.weight | Block 21 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 797 |
+
| 199 | blk.21.ffn_gate.weight | Block 21 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 798 |
+
| 200 | blk.21.ffn_norm.weight | Block 21 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 799 |
+
| 201 | blk.21.ffn_up.weight | Block 21 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 800 |
+
|
| 801 |
+
- Total elements in blk.21: (~218M) 218112000
|
| 802 |
+
- Percentage of total elements: 2.72%
|
| 803 |
+
|
| 804 |
+
|
| 805 |
+
### <a name="blk_22">Block 22 Tensor Group : ~218M Elements</a>
|
| 806 |
+
|
| 807 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 808 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 809 |
+
| 202 | blk.22.attn_k.weight | Block 22 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 810 |
+
| 203 | blk.22.attn_norm.weight | Block 22 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 811 |
+
| 204 | blk.22.attn_output.weight | Block 22 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 812 |
+
| 205 | blk.22.attn_q.weight | Block 22 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 813 |
+
| 206 | blk.22.attn_v.weight | Block 22 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 814 |
+
| 207 | blk.22.ffn_down.weight | Block 22 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 815 |
+
| 208 | blk.22.ffn_gate.weight | Block 22 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 816 |
+
| 209 | blk.22.ffn_norm.weight | Block 22 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 817 |
+
| 210 | blk.22.ffn_up.weight | Block 22 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 818 |
+
|
| 819 |
+
- Total elements in blk.22: (~218M) 218112000
|
| 820 |
+
- Percentage of total elements: 2.72%
|
| 821 |
+
|
| 822 |
+
|
| 823 |
+
### <a name="blk_23">Block 23 Tensor Group : ~218M Elements</a>
|
| 824 |
+
|
| 825 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 826 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 827 |
+
| 211 | blk.23.attn_k.weight | Block 23 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 828 |
+
| 212 | blk.23.attn_norm.weight | Block 23 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 829 |
+
| 213 | blk.23.attn_output.weight | Block 23 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 830 |
+
| 214 | blk.23.attn_q.weight | Block 23 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 831 |
+
| 215 | blk.23.attn_v.weight | Block 23 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 832 |
+
| 216 | blk.23.ffn_down.weight | Block 23 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 833 |
+
| 217 | blk.23.ffn_gate.weight | Block 23 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 834 |
+
| 218 | blk.23.ffn_norm.weight | Block 23 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 835 |
+
| 219 | blk.23.ffn_up.weight | Block 23 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 836 |
+
|
| 837 |
+
- Total elements in blk.23: (~218M) 218112000
|
| 838 |
+
- Percentage of total elements: 2.72%
|
| 839 |
+
|
| 840 |
+
|
| 841 |
+
### <a name="blk_24">Block 24 Tensor Group : ~218M Elements</a>
|
| 842 |
+
|
| 843 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 844 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 845 |
+
| 220 | blk.24.attn_k.weight | Block 24 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 846 |
+
| 221 | blk.24.attn_norm.weight | Block 24 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 847 |
+
| 222 | blk.24.attn_output.weight | Block 24 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 848 |
+
| 223 | blk.24.attn_q.weight | Block 24 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 849 |
+
| 224 | blk.24.attn_v.weight | Block 24 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 850 |
+
| 225 | blk.24.ffn_down.weight | Block 24 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 851 |
+
| 226 | blk.24.ffn_gate.weight | Block 24 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 852 |
+
| 227 | blk.24.ffn_norm.weight | Block 24 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 853 |
+
| 228 | blk.24.ffn_up.weight | Block 24 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 854 |
+
|
| 855 |
+
- Total elements in blk.24: (~218M) 218112000
|
| 856 |
+
- Percentage of total elements: 2.72%
|
| 857 |
+
|
| 858 |
+
|
| 859 |
+
### <a name="blk_25">Block 25 Tensor Group : ~218M Elements</a>
|
| 860 |
+
|
| 861 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 862 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 863 |
+
| 229 | blk.25.attn_k.weight | Block 25 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 864 |
+
| 230 | blk.25.attn_norm.weight | Block 25 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 865 |
+
| 231 | blk.25.attn_output.weight | Block 25 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 866 |
+
| 232 | blk.25.attn_q.weight | Block 25 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 867 |
+
| 233 | blk.25.attn_v.weight | Block 25 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 868 |
+
| 234 | blk.25.ffn_down.weight | Block 25 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 869 |
+
| 235 | blk.25.ffn_gate.weight | Block 25 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 870 |
+
| 236 | blk.25.ffn_norm.weight | Block 25 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 871 |
+
| 237 | blk.25.ffn_up.weight | Block 25 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 872 |
+
|
| 873 |
+
- Total elements in blk.25: (~218M) 218112000
|
| 874 |
+
- Percentage of total elements: 2.72%
|
| 875 |
+
|
| 876 |
+
|
| 877 |
+
### <a name="blk_26">Block 26 Tensor Group : ~218M Elements</a>
|
| 878 |
+
|
| 879 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 880 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 881 |
+
| 238 | blk.26.attn_k.weight | Block 26 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 882 |
+
| 239 | blk.26.attn_norm.weight | Block 26 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 883 |
+
| 240 | blk.26.attn_output.weight | Block 26 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 884 |
+
| 241 | blk.26.attn_q.weight | Block 26 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 885 |
+
| 242 | blk.26.attn_v.weight | Block 26 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 886 |
+
| 243 | blk.26.ffn_down.weight | Block 26 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 887 |
+
| 244 | blk.26.ffn_gate.weight | Block 26 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 888 |
+
| 245 | blk.26.ffn_norm.weight | Block 26 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 889 |
+
| 246 | blk.26.ffn_up.weight | Block 26 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 890 |
+
|
| 891 |
+
- Total elements in blk.26: (~218M) 218112000
|
| 892 |
+
- Percentage of total elements: 2.72%
|
| 893 |
+
|
| 894 |
+
|
| 895 |
+
### <a name="blk_27">Block 27 Tensor Group : ~218M Elements</a>
|
| 896 |
+
|
| 897 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 898 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 899 |
+
| 247 | blk.27.attn_k.weight | Block 27 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 900 |
+
| 248 | blk.27.attn_norm.weight | Block 27 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 901 |
+
| 249 | blk.27.attn_output.weight | Block 27 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 902 |
+
| 250 | blk.27.attn_q.weight | Block 27 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 903 |
+
| 251 | blk.27.attn_v.weight | Block 27 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 904 |
+
| 252 | blk.27.ffn_down.weight | Block 27 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 905 |
+
| 253 | blk.27.ffn_gate.weight | Block 27 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 906 |
+
| 254 | blk.27.ffn_norm.weight | Block 27 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 907 |
+
| 255 | blk.27.ffn_up.weight | Block 27 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 908 |
+
|
| 909 |
+
- Total elements in blk.27: (~218M) 218112000
|
| 910 |
+
- Percentage of total elements: 2.72%
|
| 911 |
+
|
| 912 |
+
|
| 913 |
+
### <a name="blk_28">Block 28 Tensor Group : ~218M Elements</a>
|
| 914 |
+
|
| 915 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 916 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 917 |
+
| 256 | blk.28.attn_k.weight | Block 28 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 918 |
+
| 257 | blk.28.attn_norm.weight | Block 28 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 919 |
+
| 258 | blk.28.attn_output.weight | Block 28 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 920 |
+
| 259 | blk.28.attn_q.weight | Block 28 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 921 |
+
| 260 | blk.28.attn_v.weight | Block 28 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 922 |
+
| 261 | blk.28.ffn_down.weight | Block 28 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 923 |
+
| 262 | blk.28.ffn_gate.weight | Block 28 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 924 |
+
| 263 | blk.28.ffn_norm.weight | Block 28 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 925 |
+
| 264 | blk.28.ffn_up.weight | Block 28 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 926 |
+
|
| 927 |
+
- Total elements in blk.28: (~218M) 218112000
|
| 928 |
+
- Percentage of total elements: 2.72%
|
| 929 |
+
|
| 930 |
+
|
| 931 |
+
### <a name="blk_29">Block 29 Tensor Group : ~218M Elements</a>
|
| 932 |
+
|
| 933 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 934 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 935 |
+
| 265 | blk.29.attn_k.weight | Block 29 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 936 |
+
| 266 | blk.29.attn_norm.weight | Block 29 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 937 |
+
| 267 | blk.29.attn_output.weight | Block 29 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 938 |
+
| 268 | blk.29.attn_q.weight | Block 29 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 939 |
+
| 269 | blk.29.attn_v.weight | Block 29 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 940 |
+
| 270 | blk.29.ffn_down.weight | Block 29 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 941 |
+
| 271 | blk.29.ffn_gate.weight | Block 29 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 942 |
+
| 272 | blk.29.ffn_norm.weight | Block 29 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 943 |
+
| 273 | blk.29.ffn_up.weight | Block 29 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 944 |
+
|
| 945 |
+
- Total elements in blk.29: (~218M) 218112000
|
| 946 |
+
- Percentage of total elements: 2.72%
|
| 947 |
+
|
| 948 |
+
|
| 949 |
+
### <a name="blk_30">Block 30 Tensor Group : ~218M Elements</a>
|
| 950 |
+
|
| 951 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 952 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 953 |
+
| 274 | blk.30.attn_k.weight | Block 30 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 954 |
+
| 275 | blk.30.attn_norm.weight | Block 30 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 955 |
+
| 276 | blk.30.attn_output.weight | Block 30 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 956 |
+
| 277 | blk.30.attn_q.weight | Block 30 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 957 |
+
| 278 | blk.30.attn_v.weight | Block 30 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 958 |
+
| 279 | blk.30.ffn_down.weight | Block 30 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 959 |
+
| 280 | blk.30.ffn_gate.weight | Block 30 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 960 |
+
| 281 | blk.30.ffn_norm.weight | Block 30 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 961 |
+
| 282 | blk.30.ffn_up.weight | Block 30 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 962 |
+
|
| 963 |
+
- Total elements in blk.30: (~218M) 218112000
|
| 964 |
+
- Percentage of total elements: 2.72%
|
| 965 |
+
|
| 966 |
+
|
| 967 |
+
### <a name="blk_31">Block 31 Tensor Group : ~218M Elements</a>
|
| 968 |
+
|
| 969 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 970 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 971 |
+
| 283 | blk.31.attn_k.weight | Block 31 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 972 |
+
| 284 | blk.31.attn_norm.weight | Block 31 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 973 |
+
| 285 | blk.31.attn_output.weight | Block 31 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 974 |
+
| 286 | blk.31.attn_q.weight | Block 31 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 975 |
+
| 287 | blk.31.attn_v.weight | Block 31 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 976 |
+
| 288 | blk.31.ffn_down.weight | Block 31 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 977 |
+
| 289 | blk.31.ffn_gate.weight | Block 31 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 978 |
+
| 290 | blk.31.ffn_norm.weight | Block 31 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 979 |
+
| 291 | blk.31.ffn_up.weight | Block 31 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 980 |
+
|
| 981 |
+
- Total elements in blk.31: (~218M) 218112000
|
| 982 |
+
- Percentage of total elements: 2.72%
|
scores/Watt-Tool-8B-Q4_K_S.md
ADDED
|
@@ -0,0 +1,982 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Watt-Tool-8B-Q4_K_S.gguf - GGUF Internal File Dump
|
| 2 |
+
|
| 3 |
+
- Endian: LITTLE endian
|
| 4 |
+
|
| 5 |
+
## Key Value Metadata Store
|
| 6 |
+
|
| 7 |
+
There are 43 key-value pairs in this file
|
| 8 |
+
|
| 9 |
+
| POS | TYPE | Count | Key | Value |
|
| 10 |
+
|----:|:---------|-------:|:---------------------------------------|:--------------------------------------------------------------------|
|
| 11 |
+
| 1 | UINT32 | 1 | GGUF.version | 3 |
|
| 12 |
+
| 2 | UINT64 | 1 | GGUF.tensor_count | 292 |
|
| 13 |
+
| 3 | UINT64 | 1 | GGUF.kv_count | 40 |
|
| 14 |
+
| 4 | STRING | 1 | general.architecture | `llama` |
|
| 15 |
+
| 5 | STRING | 1 | general.type | `model` |
|
| 16 |
+
| 6 | STRING | 1 | general.name | `Watt Tool 8B GGUF` |
|
| 17 |
+
| 7 | STRING | 1 | general.finetune | `GGUF` |
|
| 18 |
+
| 8 | STRING | 1 | general.basename | `Watt-Tool` |
|
| 19 |
+
| 9 | STRING | 1 | general.size_label | `8B` |
|
| 20 |
+
| 10 | STRING | 1 | general.license | `apache-2.0` |
|
| 21 |
+
| 11 | UINT32 | 1 | general.base_model.count | 1 |
|
| 22 |
+
| 12 | STRING | 1 | general.base_model.0.name | `Llama 3.1 8B Instruct` |
|
| 23 |
+
| 13 | STRING | 1 | general.base_model.0.organization | `Meta Llama` |
|
| 24 |
+
| 14 | STRING | 1 | general.base_model.0.repo_url | `https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct` |
|
| 25 |
+
| 15 | [STRING] | 4 | general.tags | [ `function-calling`, `tool-use`, `llama`, `bfcl` ] |
|
| 26 |
+
| 16 | [STRING] | 1 | general.languages | [ `en` ] |
|
| 27 |
+
| 17 | UINT32 | 1 | llama.block_count | 32 |
|
| 28 |
+
| 18 | UINT32 | 1 | llama.context_length | 131072 |
|
| 29 |
+
| 19 | UINT32 | 1 | llama.embedding_length | 4096 |
|
| 30 |
+
| 20 | UINT32 | 1 | llama.feed_forward_length | 14336 |
|
| 31 |
+
| 21 | UINT32 | 1 | llama.attention.head_count | 32 |
|
| 32 |
+
| 22 | UINT32 | 1 | llama.attention.head_count_kv | 8 |
|
| 33 |
+
| 23 | FLOAT32 | 1 | llama.rope.freq_base | 500000.0 |
|
| 34 |
+
| 24 | FLOAT32 | 1 | llama.attention.layer_norm_rms_epsilon | 1e-05 |
|
| 35 |
+
| 25 | UINT32 | 1 | llama.attention.key_length | 128 |
|
| 36 |
+
| 26 | UINT32 | 1 | llama.attention.value_length | 128 |
|
| 37 |
+
| 27 | UINT32 | 1 | llama.vocab_size | 128256 |
|
| 38 |
+
| 28 | UINT32 | 1 | llama.rope.dimension_count | 128 |
|
| 39 |
+
| 29 | STRING | 1 | tokenizer.ggml.model | `gpt2` |
|
| 40 |
+
| 30 | STRING | 1 | tokenizer.ggml.pre | `llama-bpe` |
|
| 41 |
+
| 31 | [STRING] | 128256 | tokenizer.ggml.tokens | [ `!`, `"`, `#`, `$`, `%`, ... ] |
|
| 42 |
+
| 32 | [INT32] | 128256 | tokenizer.ggml.token_type | [ 1, 1, 1, 1, 1, 1, 1, ... ] |
|
| 43 |
+
| 33 | [STRING] | 280147 | tokenizer.ggml.merges | [ `Ġ Ġ`, `Ġ ĠĠĠ`, `ĠĠ ĠĠ`, `ĠĠĠ Ġ`, `i n`, ... ] |
|
| 44 |
+
| 34 | UINT32 | 1 | tokenizer.ggml.bos_token_id | 128000 |
|
| 45 |
+
| 35 | UINT32 | 1 | tokenizer.ggml.eos_token_id | 128009 |
|
| 46 |
+
| 36 | UINT32 | 1 | tokenizer.ggml.padding_token_id | 128009 |
|
| 47 |
+
| 37 | STRING | 1 | tokenizer.chat_template | `{{ '<|begin_of_text|>' }}{% if`...`d|>' }}{% endif %}{% endfor %}` |
|
| 48 |
+
| 38 | UINT32 | 1 | general.quantization_version | 2 |
|
| 49 |
+
| 39 | UINT32 | 1 | general.file_type | 14 |
|
| 50 |
+
| 40 | STRING | 1 | quantize.imatrix.file | `./imatrix/imatrix-Watt-Tool-8B-small.dat` |
|
| 51 |
+
| 41 | STRING | 1 | quantize.imatrix.dataset | `../../datasets/imatrix/calibration_eur_small.txt` |
|
| 52 |
+
| 42 | INT32 | 1 | quantize.imatrix.entries_count | 225 |
|
| 53 |
+
| 43 | INT32 | 1 | quantize.imatrix.chunks_count | 962 |
|
| 54 |
+
|
| 55 |
+
## Tensors Overview ~8B Elements
|
| 56 |
+
|
| 57 |
+
Total number of elements in all tensors: 8030261312 Elements
|
| 58 |
+
|
| 59 |
+
- [Watt-Tool-8B-Q4\_K\_S.gguf - GGUF Internal File Dump](#watt-tool-8b-q4_k_sgguf---gguf-internal-file-dump)
|
| 60 |
+
- [Key Value Metadata Store](#key-value-metadata-store)
|
| 61 |
+
- [Tensors Overview ~8B Elements](#tensors-overview-8b-elements)
|
| 62 |
+
- [Tensor Data Offset](#tensor-data-offset)
|
| 63 |
+
- [Base Tensor Group : ~1B Elements](#base-tensor-group--1b-elements)
|
| 64 |
+
- [Block 0 Tensor Group : ~218M Elements](#block-0-tensor-group--218m-elements)
|
| 65 |
+
- [Block 1 Tensor Group : ~218M Elements](#block-1-tensor-group--218m-elements)
|
| 66 |
+
- [Block 2 Tensor Group : ~218M Elements](#block-2-tensor-group--218m-elements)
|
| 67 |
+
- [Block 3 Tensor Group : ~218M Elements](#block-3-tensor-group--218m-elements)
|
| 68 |
+
- [Block 4 Tensor Group : ~218M Elements](#block-4-tensor-group--218m-elements)
|
| 69 |
+
- [Block 5 Tensor Group : ~218M Elements](#block-5-tensor-group--218m-elements)
|
| 70 |
+
- [Block 6 Tensor Group : ~218M Elements](#block-6-tensor-group--218m-elements)
|
| 71 |
+
- [Block 7 Tensor Group : ~218M Elements](#block-7-tensor-group--218m-elements)
|
| 72 |
+
- [Block 8 Tensor Group : ~218M Elements](#block-8-tensor-group--218m-elements)
|
| 73 |
+
- [Block 9 Tensor Group : ~218M Elements](#block-9-tensor-group--218m-elements)
|
| 74 |
+
- [Block 10 Tensor Group : ~218M Elements](#block-10-tensor-group--218m-elements)
|
| 75 |
+
- [Block 11 Tensor Group : ~218M Elements](#block-11-tensor-group--218m-elements)
|
| 76 |
+
- [Block 12 Tensor Group : ~218M Elements](#block-12-tensor-group--218m-elements)
|
| 77 |
+
- [Block 13 Tensor Group : ~218M Elements](#block-13-tensor-group--218m-elements)
|
| 78 |
+
- [Block 14 Tensor Group : ~218M Elements](#block-14-tensor-group--218m-elements)
|
| 79 |
+
- [Block 15 Tensor Group : ~218M Elements](#block-15-tensor-group--218m-elements)
|
| 80 |
+
- [Block 16 Tensor Group : ~218M Elements](#block-16-tensor-group--218m-elements)
|
| 81 |
+
- [Block 17 Tensor Group : ~218M Elements](#block-17-tensor-group--218m-elements)
|
| 82 |
+
- [Block 18 Tensor Group : ~218M Elements](#block-18-tensor-group--218m-elements)
|
| 83 |
+
- [Block 19 Tensor Group : ~218M Elements](#block-19-tensor-group--218m-elements)
|
| 84 |
+
- [Block 20 Tensor Group : ~218M Elements](#block-20-tensor-group--218m-elements)
|
| 85 |
+
- [Block 21 Tensor Group : ~218M Elements](#block-21-tensor-group--218m-elements)
|
| 86 |
+
- [Block 22 Tensor Group : ~218M Elements](#block-22-tensor-group--218m-elements)
|
| 87 |
+
- [Block 23 Tensor Group : ~218M Elements](#block-23-tensor-group--218m-elements)
|
| 88 |
+
- [Block 24 Tensor Group : ~218M Elements](#block-24-tensor-group--218m-elements)
|
| 89 |
+
- [Block 25 Tensor Group : ~218M Elements](#block-25-tensor-group--218m-elements)
|
| 90 |
+
- [Block 26 Tensor Group : ~218M Elements](#block-26-tensor-group--218m-elements)
|
| 91 |
+
- [Block 27 Tensor Group : ~218M Elements](#block-27-tensor-group--218m-elements)
|
| 92 |
+
- [Block 28 Tensor Group : ~218M Elements](#block-28-tensor-group--218m-elements)
|
| 93 |
+
- [Block 29 Tensor Group : ~218M Elements](#block-29-tensor-group--218m-elements)
|
| 94 |
+
- [Block 30 Tensor Group : ~218M Elements](#block-30-tensor-group--218m-elements)
|
| 95 |
+
- [Block 31 Tensor Group : ~218M Elements](#block-31-tensor-group--218m-elements)
|
| 96 |
+
|
| 97 |
+
### Tensor Data Offset
|
| 98 |
+
|
| 99 |
+
This table contains the offset and data segment relative to start of file
|
| 100 |
+
|
| 101 |
+
| T_ID | Tensor Layer Name | Data Offset (B) | Data Size (B) |
|
| 102 |
+
|-----:|:--------------------------|-----------------:|-----------------:|
|
| 103 |
+
| 0 | output.weight | 0x779620 | 0x119d0000 |
|
| 104 |
+
| 1 | output_norm.weight | 0x12149620 | 0x4000 |
|
| 105 |
+
| 2 | rope_freqs.weight | 0x1214d620 | 0x100 |
|
| 106 |
+
| 3 | token_embd.weight | 0x1214d720 | 0xd746000 |
|
| 107 |
+
| 4 | blk.0.attn_k.weight | 0x1f893720 | 0x1b8000 |
|
| 108 |
+
| 5 | blk.0.attn_norm.weight | 0x1fa4b720 | 0x4000 |
|
| 109 |
+
| 6 | blk.0.attn_output.weight | 0x1fa4f720 | 0x900000 |
|
| 110 |
+
| 7 | blk.0.attn_q.weight | 0x2034f720 | 0x6e0000 |
|
| 111 |
+
| 8 | blk.0.attn_v.weight | 0x20a2f720 | 0x240000 |
|
| 112 |
+
| 9 | blk.0.ffn_down.weight | 0x20c6f720 | 0x1f80000 |
|
| 113 |
+
| 10 | blk.0.ffn_gate.weight | 0x22bef720 | 0x1810000 |
|
| 114 |
+
| 11 | blk.0.ffn_norm.weight | 0x243ff720 | 0x4000 |
|
| 115 |
+
| 12 | blk.0.ffn_up.weight | 0x24403720 | 0x1810000 |
|
| 116 |
+
| 13 | blk.1.attn_k.weight | 0x25c13720 | 0x1b8000 |
|
| 117 |
+
| 14 | blk.1.attn_norm.weight | 0x25dcb720 | 0x4000 |
|
| 118 |
+
| 15 | blk.1.attn_output.weight | 0x25dcf720 | 0x900000 |
|
| 119 |
+
| 16 | blk.1.attn_q.weight | 0x266cf720 | 0x6e0000 |
|
| 120 |
+
| 17 | blk.1.attn_v.weight | 0x26daf720 | 0x240000 |
|
| 121 |
+
| 18 | blk.1.ffn_down.weight | 0x26fef720 | 0x2680000 |
|
| 122 |
+
| 19 | blk.1.ffn_gate.weight | 0x2966f720 | 0x1810000 |
|
| 123 |
+
| 20 | blk.1.ffn_norm.weight | 0x2ae7f720 | 0x4000 |
|
| 124 |
+
| 21 | blk.1.ffn_up.weight | 0x2ae83720 | 0x1810000 |
|
| 125 |
+
| 22 | blk.2.attn_k.weight | 0x2c693720 | 0x1b8000 |
|
| 126 |
+
| 23 | blk.2.attn_norm.weight | 0x2c84b720 | 0x4000 |
|
| 127 |
+
| 24 | blk.2.attn_output.weight | 0x2c84f720 | 0x900000 |
|
| 128 |
+
| 25 | blk.2.attn_q.weight | 0x2d14f720 | 0x6e0000 |
|
| 129 |
+
| 26 | blk.2.attn_v.weight | 0x2d82f720 | 0x240000 |
|
| 130 |
+
| 27 | blk.2.ffn_down.weight | 0x2da6f720 | 0x1f80000 |
|
| 131 |
+
| 28 | blk.2.ffn_gate.weight | 0x2f9ef720 | 0x1810000 |
|
| 132 |
+
| 29 | blk.2.ffn_norm.weight | 0x311ff720 | 0x4000 |
|
| 133 |
+
| 30 | blk.2.ffn_up.weight | 0x31203720 | 0x1810000 |
|
| 134 |
+
| 31 | blk.3.attn_k.weight | 0x32a13720 | 0x1b8000 |
|
| 135 |
+
| 32 | blk.3.attn_norm.weight | 0x32bcb720 | 0x4000 |
|
| 136 |
+
| 33 | blk.3.attn_output.weight | 0x32bcf720 | 0x900000 |
|
| 137 |
+
| 34 | blk.3.attn_q.weight | 0x334cf720 | 0x6e0000 |
|
| 138 |
+
| 35 | blk.3.attn_v.weight | 0x33baf720 | 0x240000 |
|
| 139 |
+
| 36 | blk.3.ffn_down.weight | 0x33def720 | 0x1f80000 |
|
| 140 |
+
| 37 | blk.3.ffn_gate.weight | 0x35d6f720 | 0x1810000 |
|
| 141 |
+
| 38 | blk.3.ffn_norm.weight | 0x3757f720 | 0x4000 |
|
| 142 |
+
| 39 | blk.3.ffn_up.weight | 0x37583720 | 0x1810000 |
|
| 143 |
+
| 40 | blk.4.attn_k.weight | 0x38d93720 | 0x1b8000 |
|
| 144 |
+
| 41 | blk.4.attn_norm.weight | 0x38f4b720 | 0x4000 |
|
| 145 |
+
| 42 | blk.4.attn_output.weight | 0x38f4f720 | 0x900000 |
|
| 146 |
+
| 43 | blk.4.attn_q.weight | 0x3984f720 | 0x6e0000 |
|
| 147 |
+
| 44 | blk.4.attn_v.weight | 0x39f2f720 | 0x240000 |
|
| 148 |
+
| 45 | blk.4.ffn_down.weight | 0x3a16f720 | 0x1f80000 |
|
| 149 |
+
| 46 | blk.4.ffn_gate.weight | 0x3c0ef720 | 0x1810000 |
|
| 150 |
+
| 47 | blk.4.ffn_norm.weight | 0x3d8ff720 | 0x4000 |
|
| 151 |
+
| 48 | blk.4.ffn_up.weight | 0x3d903720 | 0x1810000 |
|
| 152 |
+
| 49 | blk.5.attn_k.weight | 0x3f113720 | 0x1b8000 |
|
| 153 |
+
| 50 | blk.5.attn_norm.weight | 0x3f2cb720 | 0x4000 |
|
| 154 |
+
| 51 | blk.5.attn_output.weight | 0x3f2cf720 | 0x900000 |
|
| 155 |
+
| 52 | blk.5.attn_q.weight | 0x3fbcf720 | 0x6e0000 |
|
| 156 |
+
| 53 | blk.5.attn_v.weight | 0x402af720 | 0x240000 |
|
| 157 |
+
| 54 | blk.5.ffn_down.weight | 0x404ef720 | 0x1f80000 |
|
| 158 |
+
| 55 | blk.5.ffn_gate.weight | 0x4246f720 | 0x1810000 |
|
| 159 |
+
| 56 | blk.5.ffn_norm.weight | 0x43c7f720 | 0x4000 |
|
| 160 |
+
| 57 | blk.5.ffn_up.weight | 0x43c83720 | 0x1810000 |
|
| 161 |
+
| 58 | blk.6.attn_k.weight | 0x45493720 | 0x1b8000 |
|
| 162 |
+
| 59 | blk.6.attn_norm.weight | 0x4564b720 | 0x4000 |
|
| 163 |
+
| 60 | blk.6.attn_output.weight | 0x4564f720 | 0x900000 |
|
| 164 |
+
| 61 | blk.6.attn_q.weight | 0x45f4f720 | 0x6e0000 |
|
| 165 |
+
| 62 | blk.6.attn_v.weight | 0x4662f720 | 0x240000 |
|
| 166 |
+
| 63 | blk.6.ffn_down.weight | 0x4686f720 | 0x1f80000 |
|
| 167 |
+
| 64 | blk.6.ffn_gate.weight | 0x487ef720 | 0x1810000 |
|
| 168 |
+
| 65 | blk.6.ffn_norm.weight | 0x49fff720 | 0x4000 |
|
| 169 |
+
| 66 | blk.6.ffn_up.weight | 0x4a003720 | 0x1810000 |
|
| 170 |
+
| 67 | blk.7.attn_k.weight | 0x4b813720 | 0x1b8000 |
|
| 171 |
+
| 68 | blk.7.attn_norm.weight | 0x4b9cb720 | 0x4000 |
|
| 172 |
+
| 69 | blk.7.attn_output.weight | 0x4b9cf720 | 0x900000 |
|
| 173 |
+
| 70 | blk.7.attn_q.weight | 0x4c2cf720 | 0x6e0000 |
|
| 174 |
+
| 71 | blk.7.attn_v.weight | 0x4c9af720 | 0x240000 |
|
| 175 |
+
| 72 | blk.7.ffn_down.weight | 0x4cbef720 | 0x1f80000 |
|
| 176 |
+
| 73 | blk.7.ffn_gate.weight | 0x4eb6f720 | 0x1810000 |
|
| 177 |
+
| 74 | blk.7.ffn_norm.weight | 0x5037f720 | 0x4000 |
|
| 178 |
+
| 75 | blk.7.ffn_up.weight | 0x50383720 | 0x1810000 |
|
| 179 |
+
| 76 | blk.8.attn_k.weight | 0x51b93720 | 0x1b8000 |
|
| 180 |
+
| 77 | blk.8.attn_norm.weight | 0x51d4b720 | 0x4000 |
|
| 181 |
+
| 78 | blk.8.attn_output.weight | 0x51d4f720 | 0x900000 |
|
| 182 |
+
| 79 | blk.8.attn_q.weight | 0x5264f720 | 0x6e0000 |
|
| 183 |
+
| 80 | blk.8.attn_v.weight | 0x52d2f720 | 0x240000 |
|
| 184 |
+
| 81 | blk.8.ffn_down.weight | 0x52f6f720 | 0x1f80000 |
|
| 185 |
+
| 82 | blk.8.ffn_gate.weight | 0x54eef720 | 0x1810000 |
|
| 186 |
+
| 83 | blk.8.ffn_norm.weight | 0x566ff720 | 0x4000 |
|
| 187 |
+
| 84 | blk.8.ffn_up.weight | 0x56703720 | 0x1810000 |
|
| 188 |
+
| 85 | blk.9.attn_k.weight | 0x57f13720 | 0x1b8000 |
|
| 189 |
+
| 86 | blk.9.attn_norm.weight | 0x580cb720 | 0x4000 |
|
| 190 |
+
| 87 | blk.9.attn_output.weight | 0x580cf720 | 0x900000 |
|
| 191 |
+
| 88 | blk.9.attn_q.weight | 0x589cf720 | 0x6e0000 |
|
| 192 |
+
| 89 | blk.9.attn_v.weight | 0x590af720 | 0x240000 |
|
| 193 |
+
| 90 | blk.9.ffn_down.weight | 0x592ef720 | 0x1f80000 |
|
| 194 |
+
| 91 | blk.9.ffn_gate.weight | 0x5b26f720 | 0x1810000 |
|
| 195 |
+
| 92 | blk.9.ffn_norm.weight | 0x5ca7f720 | 0x4000 |
|
| 196 |
+
| 93 | blk.9.ffn_up.weight | 0x5ca83720 | 0x1810000 |
|
| 197 |
+
| 94 | blk.10.attn_k.weight | 0x5e293720 | 0x1b8000 |
|
| 198 |
+
| 95 | blk.10.attn_norm.weight | 0x5e44b720 | 0x4000 |
|
| 199 |
+
| 96 | blk.10.attn_output.weight | 0x5e44f720 | 0x900000 |
|
| 200 |
+
| 97 | blk.10.attn_q.weight | 0x5ed4f720 | 0x6e0000 |
|
| 201 |
+
| 98 | blk.10.attn_v.weight | 0x5f42f720 | 0x240000 |
|
| 202 |
+
| 99 | blk.10.ffn_down.weight | 0x5f66f720 | 0x1f80000 |
|
| 203 |
+
| 100 | blk.10.ffn_gate.weight | 0x615ef720 | 0x1810000 |
|
| 204 |
+
| 101 | blk.10.ffn_norm.weight | 0x62dff720 | 0x4000 |
|
| 205 |
+
| 102 | blk.10.ffn_up.weight | 0x62e03720 | 0x1810000 |
|
| 206 |
+
| 103 | blk.11.attn_k.weight | 0x64613720 | 0x1b8000 |
|
| 207 |
+
| 104 | blk.11.attn_norm.weight | 0x647cb720 | 0x4000 |
|
| 208 |
+
| 105 | blk.11.attn_output.weight | 0x647cf720 | 0x900000 |
|
| 209 |
+
| 106 | blk.11.attn_q.weight | 0x650cf720 | 0x6e0000 |
|
| 210 |
+
| 107 | blk.11.attn_v.weight | 0x657af720 | 0x240000 |
|
| 211 |
+
| 108 | blk.11.ffn_down.weight | 0x659ef720 | 0x1f80000 |
|
| 212 |
+
| 109 | blk.11.ffn_gate.weight | 0x6796f720 | 0x1810000 |
|
| 213 |
+
| 110 | blk.11.ffn_norm.weight | 0x6917f720 | 0x4000 |
|
| 214 |
+
| 111 | blk.11.ffn_up.weight | 0x69183720 | 0x1810000 |
|
| 215 |
+
| 112 | blk.12.attn_k.weight | 0x6a993720 | 0x1b8000 |
|
| 216 |
+
| 113 | blk.12.attn_norm.weight | 0x6ab4b720 | 0x4000 |
|
| 217 |
+
| 114 | blk.12.attn_output.weight | 0x6ab4f720 | 0x900000 |
|
| 218 |
+
| 115 | blk.12.attn_q.weight | 0x6b44f720 | 0x6e0000 |
|
| 219 |
+
| 116 | blk.12.attn_v.weight | 0x6bb2f720 | 0x240000 |
|
| 220 |
+
| 117 | blk.12.ffn_down.weight | 0x6bd6f720 | 0x1f80000 |
|
| 221 |
+
| 118 | blk.12.ffn_gate.weight | 0x6dcef720 | 0x1810000 |
|
| 222 |
+
| 119 | blk.12.ffn_norm.weight | 0x6f4ff720 | 0x4000 |
|
| 223 |
+
| 120 | blk.12.ffn_up.weight | 0x6f503720 | 0x1810000 |
|
| 224 |
+
| 121 | blk.13.attn_k.weight | 0x70d13720 | 0x240000 |
|
| 225 |
+
| 122 | blk.13.attn_norm.weight | 0x70f53720 | 0x4000 |
|
| 226 |
+
| 123 | blk.13.attn_output.weight | 0x70f57720 | 0x900000 |
|
| 227 |
+
| 124 | blk.13.attn_q.weight | 0x71857720 | 0x900000 |
|
| 228 |
+
| 125 | blk.13.attn_v.weight | 0x72157720 | 0x240000 |
|
| 229 |
+
| 126 | blk.13.ffn_down.weight | 0x72397720 | 0x1f80000 |
|
| 230 |
+
| 127 | blk.13.ffn_gate.weight | 0x74317720 | 0x1810000 |
|
| 231 |
+
| 128 | blk.13.ffn_norm.weight | 0x75b27720 | 0x4000 |
|
| 232 |
+
| 129 | blk.13.ffn_up.weight | 0x75b2b720 | 0x1810000 |
|
| 233 |
+
| 130 | blk.14.attn_k.weight | 0x7733b720 | 0x240000 |
|
| 234 |
+
| 131 | blk.14.attn_norm.weight | 0x7757b720 | 0x4000 |
|
| 235 |
+
| 132 | blk.14.attn_output.weight | 0x7757f720 | 0x900000 |
|
| 236 |
+
| 133 | blk.14.attn_q.weight | 0x77e7f720 | 0x900000 |
|
| 237 |
+
| 134 | blk.14.attn_v.weight | 0x7877f720 | 0x240000 |
|
| 238 |
+
| 135 | blk.14.ffn_down.weight | 0x789bf720 | 0x1f80000 |
|
| 239 |
+
| 136 | blk.14.ffn_gate.weight | 0x7a93f720 | 0x1810000 |
|
| 240 |
+
| 137 | blk.14.ffn_norm.weight | 0x7c14f720 | 0x4000 |
|
| 241 |
+
| 138 | blk.14.ffn_up.weight | 0x7c153720 | 0x1810000 |
|
| 242 |
+
| 139 | blk.15.attn_k.weight | 0x7d963720 | 0x1b8000 |
|
| 243 |
+
| 140 | blk.15.attn_norm.weight | 0x7db1b720 | 0x4000 |
|
| 244 |
+
| 141 | blk.15.attn_output.weight | 0x7db1f720 | 0x900000 |
|
| 245 |
+
| 142 | blk.15.attn_q.weight | 0x7e41f720 | 0x6e0000 |
|
| 246 |
+
| 143 | blk.15.attn_v.weight | 0x7eaff720 | 0x240000 |
|
| 247 |
+
| 144 | blk.15.ffn_down.weight | 0x7ed3f720 | 0x1f80000 |
|
| 248 |
+
| 145 | blk.15.ffn_gate.weight | 0x80cbf720 | 0x1810000 |
|
| 249 |
+
| 146 | blk.15.ffn_norm.weight | 0x824cf720 | 0x4000 |
|
| 250 |
+
| 147 | blk.15.ffn_up.weight | 0x824d3720 | 0x1810000 |
|
| 251 |
+
| 148 | blk.16.attn_k.weight | 0x83ce3720 | 0x240000 |
|
| 252 |
+
| 149 | blk.16.attn_norm.weight | 0x83f23720 | 0x4000 |
|
| 253 |
+
| 150 | blk.16.attn_output.weight | 0x83f27720 | 0x900000 |
|
| 254 |
+
| 151 | blk.16.attn_q.weight | 0x84827720 | 0x900000 |
|
| 255 |
+
| 152 | blk.16.attn_v.weight | 0x85127720 | 0x240000 |
|
| 256 |
+
| 153 | blk.16.ffn_down.weight | 0x85367720 | 0x1f80000 |
|
| 257 |
+
| 154 | blk.16.ffn_gate.weight | 0x872e7720 | 0x1f80000 |
|
| 258 |
+
| 155 | blk.16.ffn_norm.weight | 0x89267720 | 0x4000 |
|
| 259 |
+
| 156 | blk.16.ffn_up.weight | 0x8926b720 | 0x1f80000 |
|
| 260 |
+
| 157 | blk.17.attn_k.weight | 0x8b1eb720 | 0x1b8000 |
|
| 261 |
+
| 158 | blk.17.attn_norm.weight | 0x8b3a3720 | 0x4000 |
|
| 262 |
+
| 159 | blk.17.attn_output.weight | 0x8b3a7720 | 0x900000 |
|
| 263 |
+
| 160 | blk.17.attn_q.weight | 0x8bca7720 | 0x6e0000 |
|
| 264 |
+
| 161 | blk.17.attn_v.weight | 0x8c387720 | 0x240000 |
|
| 265 |
+
| 162 | blk.17.ffn_down.weight | 0x8c5c7720 | 0x2680000 |
|
| 266 |
+
| 163 | blk.17.ffn_gate.weight | 0x8ec47720 | 0x1f80000 |
|
| 267 |
+
| 164 | blk.17.ffn_norm.weight | 0x90bc7720 | 0x4000 |
|
| 268 |
+
| 165 | blk.17.ffn_up.weight | 0x90bcb720 | 0x1f80000 |
|
| 269 |
+
| 166 | blk.18.attn_k.weight | 0x92b4b720 | 0x240000 |
|
| 270 |
+
| 167 | blk.18.attn_norm.weight | 0x92d8b720 | 0x4000 |
|
| 271 |
+
| 168 | blk.18.attn_output.weight | 0x92d8f720 | 0x900000 |
|
| 272 |
+
| 169 | blk.18.attn_q.weight | 0x9368f720 | 0x900000 |
|
| 273 |
+
| 170 | blk.18.attn_v.weight | 0x93f8f720 | 0x240000 |
|
| 274 |
+
| 171 | blk.18.ffn_down.weight | 0x941cf720 | 0x2680000 |
|
| 275 |
+
| 172 | blk.18.ffn_gate.weight | 0x9684f720 | 0x1f80000 |
|
| 276 |
+
| 173 | blk.18.ffn_norm.weight | 0x987cf720 | 0x4000 |
|
| 277 |
+
| 174 | blk.18.ffn_up.weight | 0x987d3720 | 0x1f80000 |
|
| 278 |
+
| 175 | blk.19.attn_k.weight | 0x9a753720 | 0x240000 |
|
| 279 |
+
| 176 | blk.19.attn_norm.weight | 0x9a993720 | 0x4000 |
|
| 280 |
+
| 177 | blk.19.attn_output.weight | 0x9a997720 | 0x900000 |
|
| 281 |
+
| 178 | blk.19.attn_q.weight | 0x9b297720 | 0x900000 |
|
| 282 |
+
| 179 | blk.19.attn_v.weight | 0x9bb97720 | 0x240000 |
|
| 283 |
+
| 180 | blk.19.ffn_down.weight | 0x9bdd7720 | 0x2680000 |
|
| 284 |
+
| 181 | blk.19.ffn_gate.weight | 0x9e457720 | 0x1f80000 |
|
| 285 |
+
| 182 | blk.19.ffn_norm.weight | 0xa03d7720 | 0x4000 |
|
| 286 |
+
| 183 | blk.19.ffn_up.weight | 0xa03db720 | 0x1f80000 |
|
| 287 |
+
| 184 | blk.20.attn_k.weight | 0xa235b720 | 0x240000 |
|
| 288 |
+
| 185 | blk.20.attn_norm.weight | 0xa259b720 | 0x4000 |
|
| 289 |
+
| 186 | blk.20.attn_output.weight | 0xa259f720 | 0x900000 |
|
| 290 |
+
| 187 | blk.20.attn_q.weight | 0xa2e9f720 | 0x900000 |
|
| 291 |
+
| 188 | blk.20.attn_v.weight | 0xa379f720 | 0x240000 |
|
| 292 |
+
| 189 | blk.20.ffn_down.weight | 0xa39df720 | 0x2680000 |
|
| 293 |
+
| 190 | blk.20.ffn_gate.weight | 0xa605f720 | 0x1f80000 |
|
| 294 |
+
| 191 | blk.20.ffn_norm.weight | 0xa7fdf720 | 0x4000 |
|
| 295 |
+
| 192 | blk.20.ffn_up.weight | 0xa7fe3720 | 0x1f80000 |
|
| 296 |
+
| 193 | blk.21.attn_k.weight | 0xa9f63720 | 0x240000 |
|
| 297 |
+
| 194 | blk.21.attn_norm.weight | 0xaa1a3720 | 0x4000 |
|
| 298 |
+
| 195 | blk.21.attn_output.weight | 0xaa1a7720 | 0x900000 |
|
| 299 |
+
| 196 | blk.21.attn_q.weight | 0xaaaa7720 | 0x900000 |
|
| 300 |
+
| 197 | blk.21.attn_v.weight | 0xab3a7720 | 0x240000 |
|
| 301 |
+
| 198 | blk.21.ffn_down.weight | 0xab5e7720 | 0x2680000 |
|
| 302 |
+
| 199 | blk.21.ffn_gate.weight | 0xadc67720 | 0x1f80000 |
|
| 303 |
+
| 200 | blk.21.ffn_norm.weight | 0xafbe7720 | 0x4000 |
|
| 304 |
+
| 201 | blk.21.ffn_up.weight | 0xafbeb720 | 0x1f80000 |
|
| 305 |
+
| 202 | blk.22.attn_k.weight | 0xb1b6b720 | 0x240000 |
|
| 306 |
+
| 203 | blk.22.attn_norm.weight | 0xb1dab720 | 0x4000 |
|
| 307 |
+
| 204 | blk.22.attn_output.weight | 0xb1daf720 | 0x900000 |
|
| 308 |
+
| 205 | blk.22.attn_q.weight | 0xb26af720 | 0x900000 |
|
| 309 |
+
| 206 | blk.22.attn_v.weight | 0xb2faf720 | 0x240000 |
|
| 310 |
+
| 207 | blk.22.ffn_down.weight | 0xb31ef720 | 0x2680000 |
|
| 311 |
+
| 208 | blk.22.ffn_gate.weight | 0xb586f720 | 0x1f80000 |
|
| 312 |
+
| 209 | blk.22.ffn_norm.weight | 0xb77ef720 | 0x4000 |
|
| 313 |
+
| 210 | blk.22.ffn_up.weight | 0xb77f3720 | 0x1f80000 |
|
| 314 |
+
| 211 | blk.23.attn_k.weight | 0xb9773720 | 0x240000 |
|
| 315 |
+
| 212 | blk.23.attn_norm.weight | 0xb99b3720 | 0x4000 |
|
| 316 |
+
| 213 | blk.23.attn_output.weight | 0xb99b7720 | 0x900000 |
|
| 317 |
+
| 214 | blk.23.attn_q.weight | 0xba2b7720 | 0x900000 |
|
| 318 |
+
| 215 | blk.23.attn_v.weight | 0xbabb7720 | 0x240000 |
|
| 319 |
+
| 216 | blk.23.ffn_down.weight | 0xbadf7720 | 0x2680000 |
|
| 320 |
+
| 217 | blk.23.ffn_gate.weight | 0xbd477720 | 0x1f80000 |
|
| 321 |
+
| 218 | blk.23.ffn_norm.weight | 0xbf3f7720 | 0x4000 |
|
| 322 |
+
| 219 | blk.23.ffn_up.weight | 0xbf3fb720 | 0x1f80000 |
|
| 323 |
+
| 220 | blk.24.attn_k.weight | 0xc137b720 | 0x240000 |
|
| 324 |
+
| 221 | blk.24.attn_norm.weight | 0xc15bb720 | 0x4000 |
|
| 325 |
+
| 222 | blk.24.attn_output.weight | 0xc15bf720 | 0x900000 |
|
| 326 |
+
| 223 | blk.24.attn_q.weight | 0xc1ebf720 | 0x900000 |
|
| 327 |
+
| 224 | blk.24.attn_v.weight | 0xc27bf720 | 0x240000 |
|
| 328 |
+
| 225 | blk.24.ffn_down.weight | 0xc29ff720 | 0x2680000 |
|
| 329 |
+
| 226 | blk.24.ffn_gate.weight | 0xc507f720 | 0x1f80000 |
|
| 330 |
+
| 227 | blk.24.ffn_norm.weight | 0xc6fff720 | 0x4000 |
|
| 331 |
+
| 228 | blk.24.ffn_up.weight | 0xc7003720 | 0x1f80000 |
|
| 332 |
+
| 229 | blk.25.attn_k.weight | 0xc8f83720 | 0x240000 |
|
| 333 |
+
| 230 | blk.25.attn_norm.weight | 0xc91c3720 | 0x4000 |
|
| 334 |
+
| 231 | blk.25.attn_output.weight | 0xc91c7720 | 0x900000 |
|
| 335 |
+
| 232 | blk.25.attn_q.weight | 0xc9ac7720 | 0x900000 |
|
| 336 |
+
| 233 | blk.25.attn_v.weight | 0xca3c7720 | 0x240000 |
|
| 337 |
+
| 234 | blk.25.ffn_down.weight | 0xca607720 | 0x2680000 |
|
| 338 |
+
| 235 | blk.25.ffn_gate.weight | 0xccc87720 | 0x1f80000 |
|
| 339 |
+
| 236 | blk.25.ffn_norm.weight | 0xcec07720 | 0x4000 |
|
| 340 |
+
| 237 | blk.25.ffn_up.weight | 0xcec0b720 | 0x1f80000 |
|
| 341 |
+
| 238 | blk.26.attn_k.weight | 0xd0b8b720 | 0x240000 |
|
| 342 |
+
| 239 | blk.26.attn_norm.weight | 0xd0dcb720 | 0x4000 |
|
| 343 |
+
| 240 | blk.26.attn_output.weight | 0xd0dcf720 | 0x900000 |
|
| 344 |
+
| 241 | blk.26.attn_q.weight | 0xd16cf720 | 0x900000 |
|
| 345 |
+
| 242 | blk.26.attn_v.weight | 0xd1fcf720 | 0x240000 |
|
| 346 |
+
| 243 | blk.26.ffn_down.weight | 0xd220f720 | 0x2680000 |
|
| 347 |
+
| 244 | blk.26.ffn_gate.weight | 0xd488f720 | 0x1f80000 |
|
| 348 |
+
| 245 | blk.26.ffn_norm.weight | 0xd680f720 | 0x4000 |
|
| 349 |
+
| 246 | blk.26.ffn_up.weight | 0xd6813720 | 0x1f80000 |
|
| 350 |
+
| 247 | blk.27.attn_k.weight | 0xd8793720 | 0x240000 |
|
| 351 |
+
| 248 | blk.27.attn_norm.weight | 0xd89d3720 | 0x4000 |
|
| 352 |
+
| 249 | blk.27.attn_output.weight | 0xd89d7720 | 0x900000 |
|
| 353 |
+
| 250 | blk.27.attn_q.weight | 0xd92d7720 | 0x900000 |
|
| 354 |
+
| 251 | blk.27.attn_v.weight | 0xd9bd7720 | 0x240000 |
|
| 355 |
+
| 252 | blk.27.ffn_down.weight | 0xd9e17720 | 0x2680000 |
|
| 356 |
+
| 253 | blk.27.ffn_gate.weight | 0xdc497720 | 0x1f80000 |
|
| 357 |
+
| 254 | blk.27.ffn_norm.weight | 0xde417720 | 0x4000 |
|
| 358 |
+
| 255 | blk.27.ffn_up.weight | 0xde41b720 | 0x1f80000 |
|
| 359 |
+
| 256 | blk.28.attn_k.weight | 0xe039b720 | 0x240000 |
|
| 360 |
+
| 257 | blk.28.attn_norm.weight | 0xe05db720 | 0x4000 |
|
| 361 |
+
| 258 | blk.28.attn_output.weight | 0xe05df720 | 0x900000 |
|
| 362 |
+
| 259 | blk.28.attn_q.weight | 0xe0edf720 | 0x900000 |
|
| 363 |
+
| 260 | blk.28.attn_v.weight | 0xe17df720 | 0x240000 |
|
| 364 |
+
| 261 | blk.28.ffn_down.weight | 0xe1a1f720 | 0x2680000 |
|
| 365 |
+
| 262 | blk.28.ffn_gate.weight | 0xe409f720 | 0x1f80000 |
|
| 366 |
+
| 263 | blk.28.ffn_norm.weight | 0xe601f720 | 0x4000 |
|
| 367 |
+
| 264 | blk.28.ffn_up.weight | 0xe6023720 | 0x1f80000 |
|
| 368 |
+
| 265 | blk.29.attn_k.weight | 0xe7fa3720 | 0x240000 |
|
| 369 |
+
| 266 | blk.29.attn_norm.weight | 0xe81e3720 | 0x4000 |
|
| 370 |
+
| 267 | blk.29.attn_output.weight | 0xe81e7720 | 0x900000 |
|
| 371 |
+
| 268 | blk.29.attn_q.weight | 0xe8ae7720 | 0x900000 |
|
| 372 |
+
| 269 | blk.29.attn_v.weight | 0xe93e7720 | 0x240000 |
|
| 373 |
+
| 270 | blk.29.ffn_down.weight | 0xe9627720 | 0x2680000 |
|
| 374 |
+
| 271 | blk.29.ffn_gate.weight | 0xebca7720 | 0x1f80000 |
|
| 375 |
+
| 272 | blk.29.ffn_norm.weight | 0xedc27720 | 0x4000 |
|
| 376 |
+
| 273 | blk.29.ffn_up.weight | 0xedc2b720 | 0x1f80000 |
|
| 377 |
+
| 274 | blk.30.attn_k.weight | 0xefbab720 | 0x240000 |
|
| 378 |
+
| 275 | blk.30.attn_norm.weight | 0xefdeb720 | 0x4000 |
|
| 379 |
+
| 276 | blk.30.attn_output.weight | 0xefdef720 | 0x900000 |
|
| 380 |
+
| 277 | blk.30.attn_q.weight | 0xf06ef720 | 0x900000 |
|
| 381 |
+
| 278 | blk.30.attn_v.weight | 0xf0fef720 | 0x240000 |
|
| 382 |
+
| 279 | blk.30.ffn_down.weight | 0xf122f720 | 0x2680000 |
|
| 383 |
+
| 280 | blk.30.ffn_gate.weight | 0xf38af720 | 0x1f80000 |
|
| 384 |
+
| 281 | blk.30.ffn_norm.weight | 0xf582f720 | 0x4000 |
|
| 385 |
+
| 282 | blk.30.ffn_up.weight | 0xf5833720 | 0x1f80000 |
|
| 386 |
+
| 283 | blk.31.attn_k.weight | 0xf77b3720 | 0x1b8000 |
|
| 387 |
+
| 284 | blk.31.attn_norm.weight | 0xf796b720 | 0x4000 |
|
| 388 |
+
| 285 | blk.31.attn_output.weight | 0xf796f720 | 0x900000 |
|
| 389 |
+
| 286 | blk.31.attn_q.weight | 0xf826f720 | 0x6e0000 |
|
| 390 |
+
| 287 | blk.31.attn_v.weight | 0xf894f720 | 0x240000 |
|
| 391 |
+
| 288 | blk.31.ffn_down.weight | 0xf8b8f720 | 0x2680000 |
|
| 392 |
+
| 289 | blk.31.ffn_gate.weight | 0xfb20f720 | 0x1f80000 |
|
| 393 |
+
| 290 | blk.31.ffn_norm.weight | 0xfd18f720 | 0x4000 |
|
| 394 |
+
| 291 | blk.31.ffn_up.weight | 0xfd193720 | 0x1f80000 |
|
| 395 |
+
|
| 396 |
+
### <a name="base">Base Tensor Group : ~1B Elements</a>
|
| 397 |
+
|
| 398 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 399 |
+
|-----:|:-------------------|:---------------------------------|:------------------|:----------------------|:-----|
|
| 400 |
+
| 0 | output.weight | Output (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | Q4_K |
|
| 401 |
+
| 1 | output_norm.weight | Output Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 402 |
+
| 2 | rope_freqs.weight | Rope_Freqs (W) | ( 64) 64 | 64 x 1 x 1 x 1 | F32 |
|
| 403 |
+
| 3 | token_embd.weight | Token Embedding (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | Q3_K |
|
| 404 |
+
|
| 405 |
+
- Total elements in base: ( ~1B) 1050677312
|
| 406 |
+
- Percentage of total elements: 13.08%
|
| 407 |
+
|
| 408 |
+
|
| 409 |
+
### <a name="blk_0">Block 0 Tensor Group : ~218M Elements</a>
|
| 410 |
+
|
| 411 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 412 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 413 |
+
| 4 | blk.0.attn_k.weight | Block 0 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 414 |
+
| 5 | blk.0.attn_norm.weight | Block 0 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 415 |
+
| 6 | blk.0.attn_output.weight | Block 0 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 416 |
+
| 7 | blk.0.attn_q.weight | Block 0 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 417 |
+
| 8 | blk.0.attn_v.weight | Block 0 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 418 |
+
| 9 | blk.0.ffn_down.weight | Block 0 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 419 |
+
| 10 | blk.0.ffn_gate.weight | Block 0 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 420 |
+
| 11 | blk.0.ffn_norm.weight | Block 0 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 421 |
+
| 12 | blk.0.ffn_up.weight | Block 0 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 422 |
+
|
| 423 |
+
- Total elements in blk.0: (~218M) 218112000
|
| 424 |
+
- Percentage of total elements: 2.72%
|
| 425 |
+
|
| 426 |
+
|
| 427 |
+
### <a name="blk_1">Block 1 Tensor Group : ~218M Elements</a>
|
| 428 |
+
|
| 429 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 430 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 431 |
+
| 13 | blk.1.attn_k.weight | Block 1 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 432 |
+
| 14 | blk.1.attn_norm.weight | Block 1 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 433 |
+
| 15 | blk.1.attn_output.weight | Block 1 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 434 |
+
| 16 | blk.1.attn_q.weight | Block 1 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 435 |
+
| 17 | blk.1.attn_v.weight | Block 1 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 436 |
+
| 18 | blk.1.ffn_down.weight | Block 1 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 437 |
+
| 19 | blk.1.ffn_gate.weight | Block 1 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 438 |
+
| 20 | blk.1.ffn_norm.weight | Block 1 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 439 |
+
| 21 | blk.1.ffn_up.weight | Block 1 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 440 |
+
|
| 441 |
+
- Total elements in blk.1: (~218M) 218112000
|
| 442 |
+
- Percentage of total elements: 2.72%
|
| 443 |
+
|
| 444 |
+
|
| 445 |
+
### <a name="blk_2">Block 2 Tensor Group : ~218M Elements</a>
|
| 446 |
+
|
| 447 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 448 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 449 |
+
| 22 | blk.2.attn_k.weight | Block 2 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 450 |
+
| 23 | blk.2.attn_norm.weight | Block 2 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 451 |
+
| 24 | blk.2.attn_output.weight | Block 2 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 452 |
+
| 25 | blk.2.attn_q.weight | Block 2 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 453 |
+
| 26 | blk.2.attn_v.weight | Block 2 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 454 |
+
| 27 | blk.2.ffn_down.weight | Block 2 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 455 |
+
| 28 | blk.2.ffn_gate.weight | Block 2 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 456 |
+
| 29 | blk.2.ffn_norm.weight | Block 2 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 457 |
+
| 30 | blk.2.ffn_up.weight | Block 2 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 458 |
+
|
| 459 |
+
- Total elements in blk.2: (~218M) 218112000
|
| 460 |
+
- Percentage of total elements: 2.72%
|
| 461 |
+
|
| 462 |
+
|
| 463 |
+
### <a name="blk_3">Block 3 Tensor Group : ~218M Elements</a>
|
| 464 |
+
|
| 465 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 466 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 467 |
+
| 31 | blk.3.attn_k.weight | Block 3 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 468 |
+
| 32 | blk.3.attn_norm.weight | Block 3 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 469 |
+
| 33 | blk.3.attn_output.weight | Block 3 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 470 |
+
| 34 | blk.3.attn_q.weight | Block 3 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 471 |
+
| 35 | blk.3.attn_v.weight | Block 3 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 472 |
+
| 36 | blk.3.ffn_down.weight | Block 3 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 473 |
+
| 37 | blk.3.ffn_gate.weight | Block 3 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 474 |
+
| 38 | blk.3.ffn_norm.weight | Block 3 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 475 |
+
| 39 | blk.3.ffn_up.weight | Block 3 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 476 |
+
|
| 477 |
+
- Total elements in blk.3: (~218M) 218112000
|
| 478 |
+
- Percentage of total elements: 2.72%
|
| 479 |
+
|
| 480 |
+
|
| 481 |
+
### <a name="blk_4">Block 4 Tensor Group : ~218M Elements</a>
|
| 482 |
+
|
| 483 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 484 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 485 |
+
| 40 | blk.4.attn_k.weight | Block 4 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 486 |
+
| 41 | blk.4.attn_norm.weight | Block 4 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 487 |
+
| 42 | blk.4.attn_output.weight | Block 4 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 488 |
+
| 43 | blk.4.attn_q.weight | Block 4 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 489 |
+
| 44 | blk.4.attn_v.weight | Block 4 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 490 |
+
| 45 | blk.4.ffn_down.weight | Block 4 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 491 |
+
| 46 | blk.4.ffn_gate.weight | Block 4 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 492 |
+
| 47 | blk.4.ffn_norm.weight | Block 4 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 493 |
+
| 48 | blk.4.ffn_up.weight | Block 4 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 494 |
+
|
| 495 |
+
- Total elements in blk.4: (~218M) 218112000
|
| 496 |
+
- Percentage of total elements: 2.72%
|
| 497 |
+
|
| 498 |
+
|
| 499 |
+
### <a name="blk_5">Block 5 Tensor Group : ~218M Elements</a>
|
| 500 |
+
|
| 501 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 502 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 503 |
+
| 49 | blk.5.attn_k.weight | Block 5 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 504 |
+
| 50 | blk.5.attn_norm.weight | Block 5 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 505 |
+
| 51 | blk.5.attn_output.weight | Block 5 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 506 |
+
| 52 | blk.5.attn_q.weight | Block 5 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 507 |
+
| 53 | blk.5.attn_v.weight | Block 5 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 508 |
+
| 54 | blk.5.ffn_down.weight | Block 5 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 509 |
+
| 55 | blk.5.ffn_gate.weight | Block 5 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 510 |
+
| 56 | blk.5.ffn_norm.weight | Block 5 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 511 |
+
| 57 | blk.5.ffn_up.weight | Block 5 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 512 |
+
|
| 513 |
+
- Total elements in blk.5: (~218M) 218112000
|
| 514 |
+
- Percentage of total elements: 2.72%
|
| 515 |
+
|
| 516 |
+
|
| 517 |
+
### <a name="blk_6">Block 6 Tensor Group : ~218M Elements</a>
|
| 518 |
+
|
| 519 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 520 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 521 |
+
| 58 | blk.6.attn_k.weight | Block 6 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 522 |
+
| 59 | blk.6.attn_norm.weight | Block 6 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 523 |
+
| 60 | blk.6.attn_output.weight | Block 6 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 524 |
+
| 61 | blk.6.attn_q.weight | Block 6 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 525 |
+
| 62 | blk.6.attn_v.weight | Block 6 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 526 |
+
| 63 | blk.6.ffn_down.weight | Block 6 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 527 |
+
| 64 | blk.6.ffn_gate.weight | Block 6 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 528 |
+
| 65 | blk.6.ffn_norm.weight | Block 6 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 529 |
+
| 66 | blk.6.ffn_up.weight | Block 6 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 530 |
+
|
| 531 |
+
- Total elements in blk.6: (~218M) 218112000
|
| 532 |
+
- Percentage of total elements: 2.72%
|
| 533 |
+
|
| 534 |
+
|
| 535 |
+
### <a name="blk_7">Block 7 Tensor Group : ~218M Elements</a>
|
| 536 |
+
|
| 537 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 538 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 539 |
+
| 67 | blk.7.attn_k.weight | Block 7 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 540 |
+
| 68 | blk.7.attn_norm.weight | Block 7 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 541 |
+
| 69 | blk.7.attn_output.weight | Block 7 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 542 |
+
| 70 | blk.7.attn_q.weight | Block 7 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 543 |
+
| 71 | blk.7.attn_v.weight | Block 7 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 544 |
+
| 72 | blk.7.ffn_down.weight | Block 7 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 545 |
+
| 73 | blk.7.ffn_gate.weight | Block 7 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 546 |
+
| 74 | blk.7.ffn_norm.weight | Block 7 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 547 |
+
| 75 | blk.7.ffn_up.weight | Block 7 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 548 |
+
|
| 549 |
+
- Total elements in blk.7: (~218M) 218112000
|
| 550 |
+
- Percentage of total elements: 2.72%
|
| 551 |
+
|
| 552 |
+
|
| 553 |
+
### <a name="blk_8">Block 8 Tensor Group : ~218M Elements</a>
|
| 554 |
+
|
| 555 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 556 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 557 |
+
| 76 | blk.8.attn_k.weight | Block 8 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 558 |
+
| 77 | blk.8.attn_norm.weight | Block 8 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 559 |
+
| 78 | blk.8.attn_output.weight | Block 8 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 560 |
+
| 79 | blk.8.attn_q.weight | Block 8 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 561 |
+
| 80 | blk.8.attn_v.weight | Block 8 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 562 |
+
| 81 | blk.8.ffn_down.weight | Block 8 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 563 |
+
| 82 | blk.8.ffn_gate.weight | Block 8 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 564 |
+
| 83 | blk.8.ffn_norm.weight | Block 8 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 565 |
+
| 84 | blk.8.ffn_up.weight | Block 8 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 566 |
+
|
| 567 |
+
- Total elements in blk.8: (~218M) 218112000
|
| 568 |
+
- Percentage of total elements: 2.72%
|
| 569 |
+
|
| 570 |
+
|
| 571 |
+
### <a name="blk_9">Block 9 Tensor Group : ~218M Elements</a>
|
| 572 |
+
|
| 573 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 574 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 575 |
+
| 85 | blk.9.attn_k.weight | Block 9 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 576 |
+
| 86 | blk.9.attn_norm.weight | Block 9 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 577 |
+
| 87 | blk.9.attn_output.weight | Block 9 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 578 |
+
| 88 | blk.9.attn_q.weight | Block 9 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 579 |
+
| 89 | blk.9.attn_v.weight | Block 9 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 580 |
+
| 90 | blk.9.ffn_down.weight | Block 9 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 581 |
+
| 91 | blk.9.ffn_gate.weight | Block 9 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 582 |
+
| 92 | blk.9.ffn_norm.weight | Block 9 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 583 |
+
| 93 | blk.9.ffn_up.weight | Block 9 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 584 |
+
|
| 585 |
+
- Total elements in blk.9: (~218M) 218112000
|
| 586 |
+
- Percentage of total elements: 2.72%
|
| 587 |
+
|
| 588 |
+
|
| 589 |
+
### <a name="blk_10">Block 10 Tensor Group : ~218M Elements</a>
|
| 590 |
+
|
| 591 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 592 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 593 |
+
| 94 | blk.10.attn_k.weight | Block 10 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 594 |
+
| 95 | blk.10.attn_norm.weight | Block 10 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 595 |
+
| 96 | blk.10.attn_output.weight | Block 10 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 596 |
+
| 97 | blk.10.attn_q.weight | Block 10 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 597 |
+
| 98 | blk.10.attn_v.weight | Block 10 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 598 |
+
| 99 | blk.10.ffn_down.weight | Block 10 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 599 |
+
| 100 | blk.10.ffn_gate.weight | Block 10 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 600 |
+
| 101 | blk.10.ffn_norm.weight | Block 10 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 601 |
+
| 102 | blk.10.ffn_up.weight | Block 10 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 602 |
+
|
| 603 |
+
- Total elements in blk.10: (~218M) 218112000
|
| 604 |
+
- Percentage of total elements: 2.72%
|
| 605 |
+
|
| 606 |
+
|
| 607 |
+
### <a name="blk_11">Block 11 Tensor Group : ~218M Elements</a>
|
| 608 |
+
|
| 609 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 610 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 611 |
+
| 103 | blk.11.attn_k.weight | Block 11 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 612 |
+
| 104 | blk.11.attn_norm.weight | Block 11 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 613 |
+
| 105 | blk.11.attn_output.weight | Block 11 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 614 |
+
| 106 | blk.11.attn_q.weight | Block 11 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 615 |
+
| 107 | blk.11.attn_v.weight | Block 11 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 616 |
+
| 108 | blk.11.ffn_down.weight | Block 11 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 617 |
+
| 109 | blk.11.ffn_gate.weight | Block 11 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 618 |
+
| 110 | blk.11.ffn_norm.weight | Block 11 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 619 |
+
| 111 | blk.11.ffn_up.weight | Block 11 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 620 |
+
|
| 621 |
+
- Total elements in blk.11: (~218M) 218112000
|
| 622 |
+
- Percentage of total elements: 2.72%
|
| 623 |
+
|
| 624 |
+
|
| 625 |
+
### <a name="blk_12">Block 12 Tensor Group : ~218M Elements</a>
|
| 626 |
+
|
| 627 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 628 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 629 |
+
| 112 | blk.12.attn_k.weight | Block 12 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 630 |
+
| 113 | blk.12.attn_norm.weight | Block 12 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 631 |
+
| 114 | blk.12.attn_output.weight | Block 12 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 632 |
+
| 115 | blk.12.attn_q.weight | Block 12 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 633 |
+
| 116 | blk.12.attn_v.weight | Block 12 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 634 |
+
| 117 | blk.12.ffn_down.weight | Block 12 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 635 |
+
| 118 | blk.12.ffn_gate.weight | Block 12 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 636 |
+
| 119 | blk.12.ffn_norm.weight | Block 12 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 637 |
+
| 120 | blk.12.ffn_up.weight | Block 12 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 638 |
+
|
| 639 |
+
- Total elements in blk.12: (~218M) 218112000
|
| 640 |
+
- Percentage of total elements: 2.72%
|
| 641 |
+
|
| 642 |
+
|
| 643 |
+
### <a name="blk_13">Block 13 Tensor Group : ~218M Elements</a>
|
| 644 |
+
|
| 645 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 646 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 647 |
+
| 121 | blk.13.attn_k.weight | Block 13 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 648 |
+
| 122 | blk.13.attn_norm.weight | Block 13 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 649 |
+
| 123 | blk.13.attn_output.weight | Block 13 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 650 |
+
| 124 | blk.13.attn_q.weight | Block 13 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 651 |
+
| 125 | blk.13.attn_v.weight | Block 13 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 652 |
+
| 126 | blk.13.ffn_down.weight | Block 13 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 653 |
+
| 127 | blk.13.ffn_gate.weight | Block 13 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 654 |
+
| 128 | blk.13.ffn_norm.weight | Block 13 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 655 |
+
| 129 | blk.13.ffn_up.weight | Block 13 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 656 |
+
|
| 657 |
+
- Total elements in blk.13: (~218M) 218112000
|
| 658 |
+
- Percentage of total elements: 2.72%
|
| 659 |
+
|
| 660 |
+
|
| 661 |
+
### <a name="blk_14">Block 14 Tensor Group : ~218M Elements</a>
|
| 662 |
+
|
| 663 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 664 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 665 |
+
| 130 | blk.14.attn_k.weight | Block 14 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 666 |
+
| 131 | blk.14.attn_norm.weight | Block 14 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 667 |
+
| 132 | blk.14.attn_output.weight | Block 14 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 668 |
+
| 133 | blk.14.attn_q.weight | Block 14 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 669 |
+
| 134 | blk.14.attn_v.weight | Block 14 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 670 |
+
| 135 | blk.14.ffn_down.weight | Block 14 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 671 |
+
| 136 | blk.14.ffn_gate.weight | Block 14 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 672 |
+
| 137 | blk.14.ffn_norm.weight | Block 14 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 673 |
+
| 138 | blk.14.ffn_up.weight | Block 14 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 674 |
+
|
| 675 |
+
- Total elements in blk.14: (~218M) 218112000
|
| 676 |
+
- Percentage of total elements: 2.72%
|
| 677 |
+
|
| 678 |
+
|
| 679 |
+
### <a name="blk_15">Block 15 Tensor Group : ~218M Elements</a>
|
| 680 |
+
|
| 681 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 682 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 683 |
+
| 139 | blk.15.attn_k.weight | Block 15 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 684 |
+
| 140 | blk.15.attn_norm.weight | Block 15 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 685 |
+
| 141 | blk.15.attn_output.weight | Block 15 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 686 |
+
| 142 | blk.15.attn_q.weight | Block 15 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 687 |
+
| 143 | blk.15.attn_v.weight | Block 15 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 688 |
+
| 144 | blk.15.ffn_down.weight | Block 15 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 689 |
+
| 145 | blk.15.ffn_gate.weight | Block 15 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 690 |
+
| 146 | blk.15.ffn_norm.weight | Block 15 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 691 |
+
| 147 | blk.15.ffn_up.weight | Block 15 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q3_K |
|
| 692 |
+
|
| 693 |
+
- Total elements in blk.15: (~218M) 218112000
|
| 694 |
+
- Percentage of total elements: 2.72%
|
| 695 |
+
|
| 696 |
+
|
| 697 |
+
### <a name="blk_16">Block 16 Tensor Group : ~218M Elements</a>
|
| 698 |
+
|
| 699 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 700 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 701 |
+
| 148 | blk.16.attn_k.weight | Block 16 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 702 |
+
| 149 | blk.16.attn_norm.weight | Block 16 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 703 |
+
| 150 | blk.16.attn_output.weight | Block 16 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 704 |
+
| 151 | blk.16.attn_q.weight | Block 16 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 705 |
+
| 152 | blk.16.attn_v.weight | Block 16 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 706 |
+
| 153 | blk.16.ffn_down.weight | Block 16 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q4_K |
|
| 707 |
+
| 154 | blk.16.ffn_gate.weight | Block 16 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 708 |
+
| 155 | blk.16.ffn_norm.weight | Block 16 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 709 |
+
| 156 | blk.16.ffn_up.weight | Block 16 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 710 |
+
|
| 711 |
+
- Total elements in blk.16: (~218M) 218112000
|
| 712 |
+
- Percentage of total elements: 2.72%
|
| 713 |
+
|
| 714 |
+
|
| 715 |
+
### <a name="blk_17">Block 17 Tensor Group : ~218M Elements</a>
|
| 716 |
+
|
| 717 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 718 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 719 |
+
| 157 | blk.17.attn_k.weight | Block 17 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 720 |
+
| 158 | blk.17.attn_norm.weight | Block 17 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 721 |
+
| 159 | blk.17.attn_output.weight | Block 17 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 722 |
+
| 160 | blk.17.attn_q.weight | Block 17 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 723 |
+
| 161 | blk.17.attn_v.weight | Block 17 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 724 |
+
| 162 | blk.17.ffn_down.weight | Block 17 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 725 |
+
| 163 | blk.17.ffn_gate.weight | Block 17 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 726 |
+
| 164 | blk.17.ffn_norm.weight | Block 17 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 727 |
+
| 165 | blk.17.ffn_up.weight | Block 17 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 728 |
+
|
| 729 |
+
- Total elements in blk.17: (~218M) 218112000
|
| 730 |
+
- Percentage of total elements: 2.72%
|
| 731 |
+
|
| 732 |
+
|
| 733 |
+
### <a name="blk_18">Block 18 Tensor Group : ~218M Elements</a>
|
| 734 |
+
|
| 735 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 736 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 737 |
+
| 166 | blk.18.attn_k.weight | Block 18 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 738 |
+
| 167 | blk.18.attn_norm.weight | Block 18 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 739 |
+
| 168 | blk.18.attn_output.weight | Block 18 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 740 |
+
| 169 | blk.18.attn_q.weight | Block 18 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 741 |
+
| 170 | blk.18.attn_v.weight | Block 18 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 742 |
+
| 171 | blk.18.ffn_down.weight | Block 18 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 743 |
+
| 172 | blk.18.ffn_gate.weight | Block 18 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 744 |
+
| 173 | blk.18.ffn_norm.weight | Block 18 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 745 |
+
| 174 | blk.18.ffn_up.weight | Block 18 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 746 |
+
|
| 747 |
+
- Total elements in blk.18: (~218M) 218112000
|
| 748 |
+
- Percentage of total elements: 2.72%
|
| 749 |
+
|
| 750 |
+
|
| 751 |
+
### <a name="blk_19">Block 19 Tensor Group : ~218M Elements</a>
|
| 752 |
+
|
| 753 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 754 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 755 |
+
| 175 | blk.19.attn_k.weight | Block 19 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 756 |
+
| 176 | blk.19.attn_norm.weight | Block 19 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 757 |
+
| 177 | blk.19.attn_output.weight | Block 19 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 758 |
+
| 178 | blk.19.attn_q.weight | Block 19 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 759 |
+
| 179 | blk.19.attn_v.weight | Block 19 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 760 |
+
| 180 | blk.19.ffn_down.weight | Block 19 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 761 |
+
| 181 | blk.19.ffn_gate.weight | Block 19 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 762 |
+
| 182 | blk.19.ffn_norm.weight | Block 19 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 763 |
+
| 183 | blk.19.ffn_up.weight | Block 19 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 764 |
+
|
| 765 |
+
- Total elements in blk.19: (~218M) 218112000
|
| 766 |
+
- Percentage of total elements: 2.72%
|
| 767 |
+
|
| 768 |
+
|
| 769 |
+
### <a name="blk_20">Block 20 Tensor Group : ~218M Elements</a>
|
| 770 |
+
|
| 771 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 772 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 773 |
+
| 184 | blk.20.attn_k.weight | Block 20 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 774 |
+
| 185 | blk.20.attn_norm.weight | Block 20 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 775 |
+
| 186 | blk.20.attn_output.weight | Block 20 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 776 |
+
| 187 | blk.20.attn_q.weight | Block 20 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 777 |
+
| 188 | blk.20.attn_v.weight | Block 20 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 778 |
+
| 189 | blk.20.ffn_down.weight | Block 20 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 779 |
+
| 190 | blk.20.ffn_gate.weight | Block 20 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 780 |
+
| 191 | blk.20.ffn_norm.weight | Block 20 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 781 |
+
| 192 | blk.20.ffn_up.weight | Block 20 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 782 |
+
|
| 783 |
+
- Total elements in blk.20: (~218M) 218112000
|
| 784 |
+
- Percentage of total elements: 2.72%
|
| 785 |
+
|
| 786 |
+
|
| 787 |
+
### <a name="blk_21">Block 21 Tensor Group : ~218M Elements</a>
|
| 788 |
+
|
| 789 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 790 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 791 |
+
| 193 | blk.21.attn_k.weight | Block 21 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 792 |
+
| 194 | blk.21.attn_norm.weight | Block 21 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 793 |
+
| 195 | blk.21.attn_output.weight | Block 21 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 794 |
+
| 196 | blk.21.attn_q.weight | Block 21 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 795 |
+
| 197 | blk.21.attn_v.weight | Block 21 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 796 |
+
| 198 | blk.21.ffn_down.weight | Block 21 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 797 |
+
| 199 | blk.21.ffn_gate.weight | Block 21 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 798 |
+
| 200 | blk.21.ffn_norm.weight | Block 21 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 799 |
+
| 201 | blk.21.ffn_up.weight | Block 21 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 800 |
+
|
| 801 |
+
- Total elements in blk.21: (~218M) 218112000
|
| 802 |
+
- Percentage of total elements: 2.72%
|
| 803 |
+
|
| 804 |
+
|
| 805 |
+
### <a name="blk_22">Block 22 Tensor Group : ~218M Elements</a>
|
| 806 |
+
|
| 807 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 808 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 809 |
+
| 202 | blk.22.attn_k.weight | Block 22 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 810 |
+
| 203 | blk.22.attn_norm.weight | Block 22 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 811 |
+
| 204 | blk.22.attn_output.weight | Block 22 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 812 |
+
| 205 | blk.22.attn_q.weight | Block 22 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 813 |
+
| 206 | blk.22.attn_v.weight | Block 22 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 814 |
+
| 207 | blk.22.ffn_down.weight | Block 22 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 815 |
+
| 208 | blk.22.ffn_gate.weight | Block 22 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 816 |
+
| 209 | blk.22.ffn_norm.weight | Block 22 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 817 |
+
| 210 | blk.22.ffn_up.weight | Block 22 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 818 |
+
|
| 819 |
+
- Total elements in blk.22: (~218M) 218112000
|
| 820 |
+
- Percentage of total elements: 2.72%
|
| 821 |
+
|
| 822 |
+
|
| 823 |
+
### <a name="blk_23">Block 23 Tensor Group : ~218M Elements</a>
|
| 824 |
+
|
| 825 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 826 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 827 |
+
| 211 | blk.23.attn_k.weight | Block 23 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 828 |
+
| 212 | blk.23.attn_norm.weight | Block 23 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 829 |
+
| 213 | blk.23.attn_output.weight | Block 23 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 830 |
+
| 214 | blk.23.attn_q.weight | Block 23 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 831 |
+
| 215 | blk.23.attn_v.weight | Block 23 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 832 |
+
| 216 | blk.23.ffn_down.weight | Block 23 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 833 |
+
| 217 | blk.23.ffn_gate.weight | Block 23 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 834 |
+
| 218 | blk.23.ffn_norm.weight | Block 23 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 835 |
+
| 219 | blk.23.ffn_up.weight | Block 23 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 836 |
+
|
| 837 |
+
- Total elements in blk.23: (~218M) 218112000
|
| 838 |
+
- Percentage of total elements: 2.72%
|
| 839 |
+
|
| 840 |
+
|
| 841 |
+
### <a name="blk_24">Block 24 Tensor Group : ~218M Elements</a>
|
| 842 |
+
|
| 843 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 844 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 845 |
+
| 220 | blk.24.attn_k.weight | Block 24 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 846 |
+
| 221 | blk.24.attn_norm.weight | Block 24 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 847 |
+
| 222 | blk.24.attn_output.weight | Block 24 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 848 |
+
| 223 | blk.24.attn_q.weight | Block 24 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 849 |
+
| 224 | blk.24.attn_v.weight | Block 24 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 850 |
+
| 225 | blk.24.ffn_down.weight | Block 24 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 851 |
+
| 226 | blk.24.ffn_gate.weight | Block 24 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 852 |
+
| 227 | blk.24.ffn_norm.weight | Block 24 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 853 |
+
| 228 | blk.24.ffn_up.weight | Block 24 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 854 |
+
|
| 855 |
+
- Total elements in blk.24: (~218M) 218112000
|
| 856 |
+
- Percentage of total elements: 2.72%
|
| 857 |
+
|
| 858 |
+
|
| 859 |
+
### <a name="blk_25">Block 25 Tensor Group : ~218M Elements</a>
|
| 860 |
+
|
| 861 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 862 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 863 |
+
| 229 | blk.25.attn_k.weight | Block 25 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 864 |
+
| 230 | blk.25.attn_norm.weight | Block 25 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 865 |
+
| 231 | blk.25.attn_output.weight | Block 25 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 866 |
+
| 232 | blk.25.attn_q.weight | Block 25 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 867 |
+
| 233 | blk.25.attn_v.weight | Block 25 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 868 |
+
| 234 | blk.25.ffn_down.weight | Block 25 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 869 |
+
| 235 | blk.25.ffn_gate.weight | Block 25 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 870 |
+
| 236 | blk.25.ffn_norm.weight | Block 25 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 871 |
+
| 237 | blk.25.ffn_up.weight | Block 25 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 872 |
+
|
| 873 |
+
- Total elements in blk.25: (~218M) 218112000
|
| 874 |
+
- Percentage of total elements: 2.72%
|
| 875 |
+
|
| 876 |
+
|
| 877 |
+
### <a name="blk_26">Block 26 Tensor Group : ~218M Elements</a>
|
| 878 |
+
|
| 879 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 880 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 881 |
+
| 238 | blk.26.attn_k.weight | Block 26 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 882 |
+
| 239 | blk.26.attn_norm.weight | Block 26 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 883 |
+
| 240 | blk.26.attn_output.weight | Block 26 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 884 |
+
| 241 | blk.26.attn_q.weight | Block 26 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 885 |
+
| 242 | blk.26.attn_v.weight | Block 26 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 886 |
+
| 243 | blk.26.ffn_down.weight | Block 26 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 887 |
+
| 244 | blk.26.ffn_gate.weight | Block 26 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 888 |
+
| 245 | blk.26.ffn_norm.weight | Block 26 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 889 |
+
| 246 | blk.26.ffn_up.weight | Block 26 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 890 |
+
|
| 891 |
+
- Total elements in blk.26: (~218M) 218112000
|
| 892 |
+
- Percentage of total elements: 2.72%
|
| 893 |
+
|
| 894 |
+
|
| 895 |
+
### <a name="blk_27">Block 27 Tensor Group : ~218M Elements</a>
|
| 896 |
+
|
| 897 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 898 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 899 |
+
| 247 | blk.27.attn_k.weight | Block 27 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 900 |
+
| 248 | blk.27.attn_norm.weight | Block 27 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 901 |
+
| 249 | blk.27.attn_output.weight | Block 27 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 902 |
+
| 250 | blk.27.attn_q.weight | Block 27 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 903 |
+
| 251 | blk.27.attn_v.weight | Block 27 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 904 |
+
| 252 | blk.27.ffn_down.weight | Block 27 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 905 |
+
| 253 | blk.27.ffn_gate.weight | Block 27 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 906 |
+
| 254 | blk.27.ffn_norm.weight | Block 27 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 907 |
+
| 255 | blk.27.ffn_up.weight | Block 27 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 908 |
+
|
| 909 |
+
- Total elements in blk.27: (~218M) 218112000
|
| 910 |
+
- Percentage of total elements: 2.72%
|
| 911 |
+
|
| 912 |
+
|
| 913 |
+
### <a name="blk_28">Block 28 Tensor Group : ~218M Elements</a>
|
| 914 |
+
|
| 915 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 916 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 917 |
+
| 256 | blk.28.attn_k.weight | Block 28 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 918 |
+
| 257 | blk.28.attn_norm.weight | Block 28 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 919 |
+
| 258 | blk.28.attn_output.weight | Block 28 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 920 |
+
| 259 | blk.28.attn_q.weight | Block 28 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 921 |
+
| 260 | blk.28.attn_v.weight | Block 28 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 922 |
+
| 261 | blk.28.ffn_down.weight | Block 28 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 923 |
+
| 262 | blk.28.ffn_gate.weight | Block 28 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 924 |
+
| 263 | blk.28.ffn_norm.weight | Block 28 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 925 |
+
| 264 | blk.28.ffn_up.weight | Block 28 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 926 |
+
|
| 927 |
+
- Total elements in blk.28: (~218M) 218112000
|
| 928 |
+
- Percentage of total elements: 2.72%
|
| 929 |
+
|
| 930 |
+
|
| 931 |
+
### <a name="blk_29">Block 29 Tensor Group : ~218M Elements</a>
|
| 932 |
+
|
| 933 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 934 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 935 |
+
| 265 | blk.29.attn_k.weight | Block 29 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 936 |
+
| 266 | blk.29.attn_norm.weight | Block 29 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 937 |
+
| 267 | blk.29.attn_output.weight | Block 29 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 938 |
+
| 268 | blk.29.attn_q.weight | Block 29 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 939 |
+
| 269 | blk.29.attn_v.weight | Block 29 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 940 |
+
| 270 | blk.29.ffn_down.weight | Block 29 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 941 |
+
| 271 | blk.29.ffn_gate.weight | Block 29 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 942 |
+
| 272 | blk.29.ffn_norm.weight | Block 29 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 943 |
+
| 273 | blk.29.ffn_up.weight | Block 29 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 944 |
+
|
| 945 |
+
- Total elements in blk.29: (~218M) 218112000
|
| 946 |
+
- Percentage of total elements: 2.72%
|
| 947 |
+
|
| 948 |
+
|
| 949 |
+
### <a name="blk_30">Block 30 Tensor Group : ~218M Elements</a>
|
| 950 |
+
|
| 951 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 952 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 953 |
+
| 274 | blk.30.attn_k.weight | Block 30 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 954 |
+
| 275 | blk.30.attn_norm.weight | Block 30 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 955 |
+
| 276 | blk.30.attn_output.weight | Block 30 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 956 |
+
| 277 | blk.30.attn_q.weight | Block 30 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 957 |
+
| 278 | blk.30.attn_v.weight | Block 30 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 958 |
+
| 279 | blk.30.ffn_down.weight | Block 30 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 959 |
+
| 280 | blk.30.ffn_gate.weight | Block 30 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 960 |
+
| 281 | blk.30.ffn_norm.weight | Block 30 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 961 |
+
| 282 | blk.30.ffn_up.weight | Block 30 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 962 |
+
|
| 963 |
+
- Total elements in blk.30: (~218M) 218112000
|
| 964 |
+
- Percentage of total elements: 2.72%
|
| 965 |
+
|
| 966 |
+
|
| 967 |
+
### <a name="blk_31">Block 31 Tensor Group : ~218M Elements</a>
|
| 968 |
+
|
| 969 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 970 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 971 |
+
| 283 | blk.31.attn_k.weight | Block 31 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q3_K |
|
| 972 |
+
| 284 | blk.31.attn_norm.weight | Block 31 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 973 |
+
| 285 | blk.31.attn_output.weight | Block 31 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 974 |
+
| 286 | blk.31.attn_q.weight | Block 31 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q3_K |
|
| 975 |
+
| 287 | blk.31.attn_v.weight | Block 31 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 976 |
+
| 288 | blk.31.ffn_down.weight | Block 31 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 977 |
+
| 289 | blk.31.ffn_gate.weight | Block 31 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 978 |
+
| 290 | blk.31.ffn_norm.weight | Block 31 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 979 |
+
| 291 | blk.31.ffn_up.weight | Block 31 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 980 |
+
|
| 981 |
+
- Total elements in blk.31: (~218M) 218112000
|
| 982 |
+
- Percentage of total elements: 2.72%
|
scores/Watt-Tool-8B-Q5_K_M.md
ADDED
|
@@ -0,0 +1,982 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Watt-Tool-8B-Q5_K_M.gguf - GGUF Internal File Dump
|
| 2 |
+
|
| 3 |
+
- Endian: LITTLE endian
|
| 4 |
+
|
| 5 |
+
## Key Value Metadata Store
|
| 6 |
+
|
| 7 |
+
There are 43 key-value pairs in this file
|
| 8 |
+
|
| 9 |
+
| POS | TYPE | Count | Key | Value |
|
| 10 |
+
|----:|:---------|-------:|:---------------------------------------|:--------------------------------------------------------------------|
|
| 11 |
+
| 1 | UINT32 | 1 | GGUF.version | 3 |
|
| 12 |
+
| 2 | UINT64 | 1 | GGUF.tensor_count | 292 |
|
| 13 |
+
| 3 | UINT64 | 1 | GGUF.kv_count | 40 |
|
| 14 |
+
| 4 | STRING | 1 | general.architecture | `llama` |
|
| 15 |
+
| 5 | STRING | 1 | general.type | `model` |
|
| 16 |
+
| 6 | STRING | 1 | general.name | `Watt Tool 8B GGUF` |
|
| 17 |
+
| 7 | STRING | 1 | general.finetune | `GGUF` |
|
| 18 |
+
| 8 | STRING | 1 | general.basename | `Watt-Tool` |
|
| 19 |
+
| 9 | STRING | 1 | general.size_label | `8B` |
|
| 20 |
+
| 10 | STRING | 1 | general.license | `apache-2.0` |
|
| 21 |
+
| 11 | UINT32 | 1 | general.base_model.count | 1 |
|
| 22 |
+
| 12 | STRING | 1 | general.base_model.0.name | `Llama 3.1 8B Instruct` |
|
| 23 |
+
| 13 | STRING | 1 | general.base_model.0.organization | `Meta Llama` |
|
| 24 |
+
| 14 | STRING | 1 | general.base_model.0.repo_url | `https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct` |
|
| 25 |
+
| 15 | [STRING] | 4 | general.tags | [ `function-calling`, `tool-use`, `llama`, `bfcl` ] |
|
| 26 |
+
| 16 | [STRING] | 1 | general.languages | [ `en` ] |
|
| 27 |
+
| 17 | UINT32 | 1 | llama.block_count | 32 |
|
| 28 |
+
| 18 | UINT32 | 1 | llama.context_length | 131072 |
|
| 29 |
+
| 19 | UINT32 | 1 | llama.embedding_length | 4096 |
|
| 30 |
+
| 20 | UINT32 | 1 | llama.feed_forward_length | 14336 |
|
| 31 |
+
| 21 | UINT32 | 1 | llama.attention.head_count | 32 |
|
| 32 |
+
| 22 | UINT32 | 1 | llama.attention.head_count_kv | 8 |
|
| 33 |
+
| 23 | FLOAT32 | 1 | llama.rope.freq_base | 500000.0 |
|
| 34 |
+
| 24 | FLOAT32 | 1 | llama.attention.layer_norm_rms_epsilon | 1e-05 |
|
| 35 |
+
| 25 | UINT32 | 1 | llama.attention.key_length | 128 |
|
| 36 |
+
| 26 | UINT32 | 1 | llama.attention.value_length | 128 |
|
| 37 |
+
| 27 | UINT32 | 1 | llama.vocab_size | 128256 |
|
| 38 |
+
| 28 | UINT32 | 1 | llama.rope.dimension_count | 128 |
|
| 39 |
+
| 29 | STRING | 1 | tokenizer.ggml.model | `gpt2` |
|
| 40 |
+
| 30 | STRING | 1 | tokenizer.ggml.pre | `llama-bpe` |
|
| 41 |
+
| 31 | [STRING] | 128256 | tokenizer.ggml.tokens | [ `!`, `"`, `#`, `$`, `%`, ... ] |
|
| 42 |
+
| 32 | [INT32] | 128256 | tokenizer.ggml.token_type | [ 1, 1, 1, 1, 1, 1, 1, ... ] |
|
| 43 |
+
| 33 | [STRING] | 280147 | tokenizer.ggml.merges | [ `Ġ Ġ`, `Ġ ĠĠĠ`, `ĠĠ ĠĠ`, `ĠĠĠ Ġ`, `i n`, ... ] |
|
| 44 |
+
| 34 | UINT32 | 1 | tokenizer.ggml.bos_token_id | 128000 |
|
| 45 |
+
| 35 | UINT32 | 1 | tokenizer.ggml.eos_token_id | 128009 |
|
| 46 |
+
| 36 | UINT32 | 1 | tokenizer.ggml.padding_token_id | 128009 |
|
| 47 |
+
| 37 | STRING | 1 | tokenizer.chat_template | `{{ '<|begin_of_text|>' }}{% if`...`d|>' }}{% endif %}{% endfor %}` |
|
| 48 |
+
| 38 | UINT32 | 1 | general.quantization_version | 2 |
|
| 49 |
+
| 39 | UINT32 | 1 | general.file_type | 17 |
|
| 50 |
+
| 40 | STRING | 1 | quantize.imatrix.file | `./imatrix/imatrix-Watt-Tool-8B-small.dat` |
|
| 51 |
+
| 41 | STRING | 1 | quantize.imatrix.dataset | `../../datasets/imatrix/calibration_eur_small.txt` |
|
| 52 |
+
| 42 | INT32 | 1 | quantize.imatrix.entries_count | 225 |
|
| 53 |
+
| 43 | INT32 | 1 | quantize.imatrix.chunks_count | 962 |
|
| 54 |
+
|
| 55 |
+
## Tensors Overview ~8B Elements
|
| 56 |
+
|
| 57 |
+
Total number of elements in all tensors: 8030261312 Elements
|
| 58 |
+
|
| 59 |
+
- [Watt-Tool-8B-Q5\_K\_M.gguf - GGUF Internal File Dump](#watt-tool-8b-q5_k_mgguf---gguf-internal-file-dump)
|
| 60 |
+
- [Key Value Metadata Store](#key-value-metadata-store)
|
| 61 |
+
- [Tensors Overview ~8B Elements](#tensors-overview-8b-elements)
|
| 62 |
+
- [Tensor Data Offset](#tensor-data-offset)
|
| 63 |
+
- [Base Tensor Group : ~1B Elements](#base-tensor-group--1b-elements)
|
| 64 |
+
- [Block 0 Tensor Group : ~218M Elements](#block-0-tensor-group--218m-elements)
|
| 65 |
+
- [Block 1 Tensor Group : ~218M Elements](#block-1-tensor-group--218m-elements)
|
| 66 |
+
- [Block 2 Tensor Group : ~218M Elements](#block-2-tensor-group--218m-elements)
|
| 67 |
+
- [Block 3 Tensor Group : ~218M Elements](#block-3-tensor-group--218m-elements)
|
| 68 |
+
- [Block 4 Tensor Group : ~218M Elements](#block-4-tensor-group--218m-elements)
|
| 69 |
+
- [Block 5 Tensor Group : ~218M Elements](#block-5-tensor-group--218m-elements)
|
| 70 |
+
- [Block 6 Tensor Group : ~218M Elements](#block-6-tensor-group--218m-elements)
|
| 71 |
+
- [Block 7 Tensor Group : ~218M Elements](#block-7-tensor-group--218m-elements)
|
| 72 |
+
- [Block 8 Tensor Group : ~218M Elements](#block-8-tensor-group--218m-elements)
|
| 73 |
+
- [Block 9 Tensor Group : ~218M Elements](#block-9-tensor-group--218m-elements)
|
| 74 |
+
- [Block 10 Tensor Group : ~218M Elements](#block-10-tensor-group--218m-elements)
|
| 75 |
+
- [Block 11 Tensor Group : ~218M Elements](#block-11-tensor-group--218m-elements)
|
| 76 |
+
- [Block 12 Tensor Group : ~218M Elements](#block-12-tensor-group--218m-elements)
|
| 77 |
+
- [Block 13 Tensor Group : ~218M Elements](#block-13-tensor-group--218m-elements)
|
| 78 |
+
- [Block 14 Tensor Group : ~218M Elements](#block-14-tensor-group--218m-elements)
|
| 79 |
+
- [Block 15 Tensor Group : ~218M Elements](#block-15-tensor-group--218m-elements)
|
| 80 |
+
- [Block 16 Tensor Group : ~218M Elements](#block-16-tensor-group--218m-elements)
|
| 81 |
+
- [Block 17 Tensor Group : ~218M Elements](#block-17-tensor-group--218m-elements)
|
| 82 |
+
- [Block 18 Tensor Group : ~218M Elements](#block-18-tensor-group--218m-elements)
|
| 83 |
+
- [Block 19 Tensor Group : ~218M Elements](#block-19-tensor-group--218m-elements)
|
| 84 |
+
- [Block 20 Tensor Group : ~218M Elements](#block-20-tensor-group--218m-elements)
|
| 85 |
+
- [Block 21 Tensor Group : ~218M Elements](#block-21-tensor-group--218m-elements)
|
| 86 |
+
- [Block 22 Tensor Group : ~218M Elements](#block-22-tensor-group--218m-elements)
|
| 87 |
+
- [Block 23 Tensor Group : ~218M Elements](#block-23-tensor-group--218m-elements)
|
| 88 |
+
- [Block 24 Tensor Group : ~218M Elements](#block-24-tensor-group--218m-elements)
|
| 89 |
+
- [Block 25 Tensor Group : ~218M Elements](#block-25-tensor-group--218m-elements)
|
| 90 |
+
- [Block 26 Tensor Group : ~218M Elements](#block-26-tensor-group--218m-elements)
|
| 91 |
+
- [Block 27 Tensor Group : ~218M Elements](#block-27-tensor-group--218m-elements)
|
| 92 |
+
- [Block 28 Tensor Group : ~218M Elements](#block-28-tensor-group--218m-elements)
|
| 93 |
+
- [Block 29 Tensor Group : ~218M Elements](#block-29-tensor-group--218m-elements)
|
| 94 |
+
- [Block 30 Tensor Group : ~218M Elements](#block-30-tensor-group--218m-elements)
|
| 95 |
+
- [Block 31 Tensor Group : ~218M Elements](#block-31-tensor-group--218m-elements)
|
| 96 |
+
|
| 97 |
+
### Tensor Data Offset
|
| 98 |
+
|
| 99 |
+
This table contains the offset and data segment relative to start of file
|
| 100 |
+
|
| 101 |
+
| T_ID | Tensor Layer Name | Data Offset (B) | Data Size (B) |
|
| 102 |
+
|-----:|:--------------------------|-----------------:|-----------------:|
|
| 103 |
+
| 0 | output.weight | 0x779620 | 0x15870000 |
|
| 104 |
+
| 1 | output_norm.weight | 0x15fe9620 | 0x4000 |
|
| 105 |
+
| 2 | rope_freqs.weight | 0x15fed620 | 0x100 |
|
| 106 |
+
| 3 | token_embd.weight | 0x15fed720 | 0xd746000 |
|
| 107 |
+
| 4 | blk.0.attn_k.weight | 0x23733720 | 0x240000 |
|
| 108 |
+
| 5 | blk.0.attn_norm.weight | 0x23973720 | 0x4000 |
|
| 109 |
+
| 6 | blk.0.attn_output.weight | 0x23977720 | 0xb00000 |
|
| 110 |
+
| 7 | blk.0.attn_q.weight | 0x24477720 | 0x900000 |
|
| 111 |
+
| 8 | blk.0.attn_v.weight | 0x24d77720 | 0x2c0000 |
|
| 112 |
+
| 9 | blk.0.ffn_down.weight | 0x25037720 | 0x2df0000 |
|
| 113 |
+
| 10 | blk.0.ffn_gate.weight | 0x27e27720 | 0x1f80000 |
|
| 114 |
+
| 11 | blk.0.ffn_norm.weight | 0x29da7720 | 0x4000 |
|
| 115 |
+
| 12 | blk.0.ffn_up.weight | 0x29dab720 | 0x1f80000 |
|
| 116 |
+
| 13 | blk.1.attn_k.weight | 0x2bd2b720 | 0x240000 |
|
| 117 |
+
| 14 | blk.1.attn_norm.weight | 0x2bf6b720 | 0x4000 |
|
| 118 |
+
| 15 | blk.1.attn_output.weight | 0x2bf6f720 | 0xb00000 |
|
| 119 |
+
| 16 | blk.1.attn_q.weight | 0x2ca6f720 | 0x900000 |
|
| 120 |
+
| 17 | blk.1.attn_v.weight | 0x2d36f720 | 0x2c0000 |
|
| 121 |
+
| 18 | blk.1.ffn_down.weight | 0x2d62f720 | 0x2df0000 |
|
| 122 |
+
| 19 | blk.1.ffn_gate.weight | 0x3041f720 | 0x1f80000 |
|
| 123 |
+
| 20 | blk.1.ffn_norm.weight | 0x3239f720 | 0x4000 |
|
| 124 |
+
| 21 | blk.1.ffn_up.weight | 0x323a3720 | 0x1f80000 |
|
| 125 |
+
| 22 | blk.2.attn_k.weight | 0x34323720 | 0x240000 |
|
| 126 |
+
| 23 | blk.2.attn_norm.weight | 0x34563720 | 0x4000 |
|
| 127 |
+
| 24 | blk.2.attn_output.weight | 0x34567720 | 0xb00000 |
|
| 128 |
+
| 25 | blk.2.attn_q.weight | 0x35067720 | 0x900000 |
|
| 129 |
+
| 26 | blk.2.attn_v.weight | 0x35967720 | 0x2c0000 |
|
| 130 |
+
| 27 | blk.2.ffn_down.weight | 0x35c27720 | 0x2df0000 |
|
| 131 |
+
| 28 | blk.2.ffn_gate.weight | 0x38a17720 | 0x1f80000 |
|
| 132 |
+
| 29 | blk.2.ffn_norm.weight | 0x3a997720 | 0x4000 |
|
| 133 |
+
| 30 | blk.2.ffn_up.weight | 0x3a99b720 | 0x1f80000 |
|
| 134 |
+
| 31 | blk.3.attn_k.weight | 0x3c91b720 | 0x240000 |
|
| 135 |
+
| 32 | blk.3.attn_norm.weight | 0x3cb5b720 | 0x4000 |
|
| 136 |
+
| 33 | blk.3.attn_output.weight | 0x3cb5f720 | 0xb00000 |
|
| 137 |
+
| 34 | blk.3.attn_q.weight | 0x3d65f720 | 0x900000 |
|
| 138 |
+
| 35 | blk.3.attn_v.weight | 0x3df5f720 | 0x2c0000 |
|
| 139 |
+
| 36 | blk.3.ffn_down.weight | 0x3e21f720 | 0x2df0000 |
|
| 140 |
+
| 37 | blk.3.ffn_gate.weight | 0x4100f720 | 0x1f80000 |
|
| 141 |
+
| 38 | blk.3.ffn_norm.weight | 0x42f8f720 | 0x4000 |
|
| 142 |
+
| 39 | blk.3.ffn_up.weight | 0x42f93720 | 0x1f80000 |
|
| 143 |
+
| 40 | blk.4.attn_k.weight | 0x44f13720 | 0x240000 |
|
| 144 |
+
| 41 | blk.4.attn_norm.weight | 0x45153720 | 0x4000 |
|
| 145 |
+
| 42 | blk.4.attn_output.weight | 0x45157720 | 0xb00000 |
|
| 146 |
+
| 43 | blk.4.attn_q.weight | 0x45c57720 | 0x900000 |
|
| 147 |
+
| 44 | blk.4.attn_v.weight | 0x46557720 | 0x2c0000 |
|
| 148 |
+
| 45 | blk.4.ffn_down.weight | 0x46817720 | 0x2df0000 |
|
| 149 |
+
| 46 | blk.4.ffn_gate.weight | 0x49607720 | 0x1f80000 |
|
| 150 |
+
| 47 | blk.4.ffn_norm.weight | 0x4b587720 | 0x4000 |
|
| 151 |
+
| 48 | blk.4.ffn_up.weight | 0x4b58b720 | 0x1f80000 |
|
| 152 |
+
| 49 | blk.5.attn_k.weight | 0x4d50b720 | 0x240000 |
|
| 153 |
+
| 50 | blk.5.attn_norm.weight | 0x4d74b720 | 0x4000 |
|
| 154 |
+
| 51 | blk.5.attn_output.weight | 0x4d74f720 | 0xb00000 |
|
| 155 |
+
| 52 | blk.5.attn_q.weight | 0x4e24f720 | 0x900000 |
|
| 156 |
+
| 53 | blk.5.attn_v.weight | 0x4eb4f720 | 0x2c0000 |
|
| 157 |
+
| 54 | blk.5.ffn_down.weight | 0x4ee0f720 | 0x2df0000 |
|
| 158 |
+
| 55 | blk.5.ffn_gate.weight | 0x51bff720 | 0x1f80000 |
|
| 159 |
+
| 56 | blk.5.ffn_norm.weight | 0x53b7f720 | 0x4000 |
|
| 160 |
+
| 57 | blk.5.ffn_up.weight | 0x53b83720 | 0x1f80000 |
|
| 161 |
+
| 58 | blk.6.attn_k.weight | 0x55b03720 | 0x240000 |
|
| 162 |
+
| 59 | blk.6.attn_norm.weight | 0x55d43720 | 0x4000 |
|
| 163 |
+
| 60 | blk.6.attn_output.weight | 0x55d47720 | 0xb00000 |
|
| 164 |
+
| 61 | blk.6.attn_q.weight | 0x56847720 | 0x900000 |
|
| 165 |
+
| 62 | blk.6.attn_v.weight | 0x57147720 | 0x2c0000 |
|
| 166 |
+
| 63 | blk.6.ffn_down.weight | 0x57407720 | 0x2df0000 |
|
| 167 |
+
| 64 | blk.6.ffn_gate.weight | 0x5a1f7720 | 0x1f80000 |
|
| 168 |
+
| 65 | blk.6.ffn_norm.weight | 0x5c177720 | 0x4000 |
|
| 169 |
+
| 66 | blk.6.ffn_up.weight | 0x5c17b720 | 0x1f80000 |
|
| 170 |
+
| 67 | blk.7.attn_k.weight | 0x5e0fb720 | 0x240000 |
|
| 171 |
+
| 68 | blk.7.attn_norm.weight | 0x5e33b720 | 0x4000 |
|
| 172 |
+
| 69 | blk.7.attn_output.weight | 0x5e33f720 | 0xb00000 |
|
| 173 |
+
| 70 | blk.7.attn_q.weight | 0x5ee3f720 | 0x900000 |
|
| 174 |
+
| 71 | blk.7.attn_v.weight | 0x5f73f720 | 0x2c0000 |
|
| 175 |
+
| 72 | blk.7.ffn_down.weight | 0x5f9ff720 | 0x2df0000 |
|
| 176 |
+
| 73 | blk.7.ffn_gate.weight | 0x627ef720 | 0x1f80000 |
|
| 177 |
+
| 74 | blk.7.ffn_norm.weight | 0x6476f720 | 0x4000 |
|
| 178 |
+
| 75 | blk.7.ffn_up.weight | 0x64773720 | 0x1f80000 |
|
| 179 |
+
| 76 | blk.8.attn_k.weight | 0x666f3720 | 0x240000 |
|
| 180 |
+
| 77 | blk.8.attn_norm.weight | 0x66933720 | 0x4000 |
|
| 181 |
+
| 78 | blk.8.attn_output.weight | 0x66937720 | 0xb00000 |
|
| 182 |
+
| 79 | blk.8.attn_q.weight | 0x67437720 | 0x900000 |
|
| 183 |
+
| 80 | blk.8.attn_v.weight | 0x67d37720 | 0x2c0000 |
|
| 184 |
+
| 81 | blk.8.ffn_down.weight | 0x67ff7720 | 0x2df0000 |
|
| 185 |
+
| 82 | blk.8.ffn_gate.weight | 0x6ade7720 | 0x1f80000 |
|
| 186 |
+
| 83 | blk.8.ffn_norm.weight | 0x6cd67720 | 0x4000 |
|
| 187 |
+
| 84 | blk.8.ffn_up.weight | 0x6cd6b720 | 0x1f80000 |
|
| 188 |
+
| 85 | blk.9.attn_k.weight | 0x6eceb720 | 0x240000 |
|
| 189 |
+
| 86 | blk.9.attn_norm.weight | 0x6ef2b720 | 0x4000 |
|
| 190 |
+
| 87 | blk.9.attn_output.weight | 0x6ef2f720 | 0xb00000 |
|
| 191 |
+
| 88 | blk.9.attn_q.weight | 0x6fa2f720 | 0x900000 |
|
| 192 |
+
| 89 | blk.9.attn_v.weight | 0x7032f720 | 0x2c0000 |
|
| 193 |
+
| 90 | blk.9.ffn_down.weight | 0x705ef720 | 0x2df0000 |
|
| 194 |
+
| 91 | blk.9.ffn_gate.weight | 0x733df720 | 0x1f80000 |
|
| 195 |
+
| 92 | blk.9.ffn_norm.weight | 0x7535f720 | 0x4000 |
|
| 196 |
+
| 93 | blk.9.ffn_up.weight | 0x75363720 | 0x1f80000 |
|
| 197 |
+
| 94 | blk.10.attn_k.weight | 0x772e3720 | 0x240000 |
|
| 198 |
+
| 95 | blk.10.attn_norm.weight | 0x77523720 | 0x4000 |
|
| 199 |
+
| 96 | blk.10.attn_output.weight | 0x77527720 | 0xb00000 |
|
| 200 |
+
| 97 | blk.10.attn_q.weight | 0x78027720 | 0x900000 |
|
| 201 |
+
| 98 | blk.10.attn_v.weight | 0x78927720 | 0x2c0000 |
|
| 202 |
+
| 99 | blk.10.ffn_down.weight | 0x78be7720 | 0x2df0000 |
|
| 203 |
+
| 100 | blk.10.ffn_gate.weight | 0x7b9d7720 | 0x1f80000 |
|
| 204 |
+
| 101 | blk.10.ffn_norm.weight | 0x7d957720 | 0x4000 |
|
| 205 |
+
| 102 | blk.10.ffn_up.weight | 0x7d95b720 | 0x1f80000 |
|
| 206 |
+
| 103 | blk.11.attn_k.weight | 0x7f8db720 | 0x240000 |
|
| 207 |
+
| 104 | blk.11.attn_norm.weight | 0x7fb1b720 | 0x4000 |
|
| 208 |
+
| 105 | blk.11.attn_output.weight | 0x7fb1f720 | 0xb00000 |
|
| 209 |
+
| 106 | blk.11.attn_q.weight | 0x8061f720 | 0x900000 |
|
| 210 |
+
| 107 | blk.11.attn_v.weight | 0x80f1f720 | 0x2c0000 |
|
| 211 |
+
| 108 | blk.11.ffn_down.weight | 0x811df720 | 0x2df0000 |
|
| 212 |
+
| 109 | blk.11.ffn_gate.weight | 0x83fcf720 | 0x1f80000 |
|
| 213 |
+
| 110 | blk.11.ffn_norm.weight | 0x85f4f720 | 0x4000 |
|
| 214 |
+
| 111 | blk.11.ffn_up.weight | 0x85f53720 | 0x1f80000 |
|
| 215 |
+
| 112 | blk.12.attn_k.weight | 0x87ed3720 | 0x240000 |
|
| 216 |
+
| 113 | blk.12.attn_norm.weight | 0x88113720 | 0x4000 |
|
| 217 |
+
| 114 | blk.12.attn_output.weight | 0x88117720 | 0xb00000 |
|
| 218 |
+
| 115 | blk.12.attn_q.weight | 0x88c17720 | 0x900000 |
|
| 219 |
+
| 116 | blk.12.attn_v.weight | 0x89517720 | 0x2c0000 |
|
| 220 |
+
| 117 | blk.12.ffn_down.weight | 0x897d7720 | 0x2df0000 |
|
| 221 |
+
| 118 | blk.12.ffn_gate.weight | 0x8c5c7720 | 0x1f80000 |
|
| 222 |
+
| 119 | blk.12.ffn_norm.weight | 0x8e547720 | 0x4000 |
|
| 223 |
+
| 120 | blk.12.ffn_up.weight | 0x8e54b720 | 0x1f80000 |
|
| 224 |
+
| 121 | blk.13.attn_k.weight | 0x904cb720 | 0x2c0000 |
|
| 225 |
+
| 122 | blk.13.attn_norm.weight | 0x9078b720 | 0x4000 |
|
| 226 |
+
| 123 | blk.13.attn_output.weight | 0x9078f720 | 0xb00000 |
|
| 227 |
+
| 124 | blk.13.attn_q.weight | 0x9128f720 | 0xb00000 |
|
| 228 |
+
| 125 | blk.13.attn_v.weight | 0x91d8f720 | 0x348000 |
|
| 229 |
+
| 126 | blk.13.ffn_down.weight | 0x920d7720 | 0x2df0000 |
|
| 230 |
+
| 127 | blk.13.ffn_gate.weight | 0x94ec7720 | 0x1f80000 |
|
| 231 |
+
| 128 | blk.13.ffn_norm.weight | 0x96e47720 | 0x4000 |
|
| 232 |
+
| 129 | blk.13.ffn_up.weight | 0x96e4b720 | 0x1f80000 |
|
| 233 |
+
| 130 | blk.14.attn_k.weight | 0x98dcb720 | 0x2c0000 |
|
| 234 |
+
| 131 | blk.14.attn_norm.weight | 0x9908b720 | 0x4000 |
|
| 235 |
+
| 132 | blk.14.attn_output.weight | 0x9908f720 | 0xb00000 |
|
| 236 |
+
| 133 | blk.14.attn_q.weight | 0x99b8f720 | 0xb00000 |
|
| 237 |
+
| 134 | blk.14.attn_v.weight | 0x9a68f720 | 0x348000 |
|
| 238 |
+
| 135 | blk.14.ffn_down.weight | 0x9a9d7720 | 0x2df0000 |
|
| 239 |
+
| 136 | blk.14.ffn_gate.weight | 0x9d7c7720 | 0x1f80000 |
|
| 240 |
+
| 137 | blk.14.ffn_norm.weight | 0x9f747720 | 0x4000 |
|
| 241 |
+
| 138 | blk.14.ffn_up.weight | 0x9f74b720 | 0x1f80000 |
|
| 242 |
+
| 139 | blk.15.attn_k.weight | 0xa16cb720 | 0x240000 |
|
| 243 |
+
| 140 | blk.15.attn_norm.weight | 0xa190b720 | 0x4000 |
|
| 244 |
+
| 141 | blk.15.attn_output.weight | 0xa190f720 | 0xb00000 |
|
| 245 |
+
| 142 | blk.15.attn_q.weight | 0xa240f720 | 0x900000 |
|
| 246 |
+
| 143 | blk.15.attn_v.weight | 0xa2d0f720 | 0x2c0000 |
|
| 247 |
+
| 144 | blk.15.ffn_down.weight | 0xa2fcf720 | 0x2df0000 |
|
| 248 |
+
| 145 | blk.15.ffn_gate.weight | 0xa5dbf720 | 0x1f80000 |
|
| 249 |
+
| 146 | blk.15.ffn_norm.weight | 0xa7d3f720 | 0x4000 |
|
| 250 |
+
| 147 | blk.15.ffn_up.weight | 0xa7d43720 | 0x1f80000 |
|
| 251 |
+
| 148 | blk.16.attn_k.weight | 0xa9cc3720 | 0x2c0000 |
|
| 252 |
+
| 149 | blk.16.attn_norm.weight | 0xa9f83720 | 0x4000 |
|
| 253 |
+
| 150 | blk.16.attn_output.weight | 0xa9f87720 | 0xb00000 |
|
| 254 |
+
| 151 | blk.16.attn_q.weight | 0xaaa87720 | 0xb00000 |
|
| 255 |
+
| 152 | blk.16.attn_v.weight | 0xab587720 | 0x348000 |
|
| 256 |
+
| 153 | blk.16.ffn_down.weight | 0xab8cf720 | 0x2df0000 |
|
| 257 |
+
| 154 | blk.16.ffn_gate.weight | 0xae6bf720 | 0x2680000 |
|
| 258 |
+
| 155 | blk.16.ffn_norm.weight | 0xb0d3f720 | 0x4000 |
|
| 259 |
+
| 156 | blk.16.ffn_up.weight | 0xb0d43720 | 0x2680000 |
|
| 260 |
+
| 157 | blk.17.attn_k.weight | 0xb33c3720 | 0x240000 |
|
| 261 |
+
| 158 | blk.17.attn_norm.weight | 0xb3603720 | 0x4000 |
|
| 262 |
+
| 159 | blk.17.attn_output.weight | 0xb3607720 | 0xb00000 |
|
| 263 |
+
| 160 | blk.17.attn_q.weight | 0xb4107720 | 0x900000 |
|
| 264 |
+
| 161 | blk.17.attn_v.weight | 0xb4a07720 | 0x2c0000 |
|
| 265 |
+
| 162 | blk.17.ffn_down.weight | 0xb4cc7720 | 0x2df0000 |
|
| 266 |
+
| 163 | blk.17.ffn_gate.weight | 0xb7ab7720 | 0x2680000 |
|
| 267 |
+
| 164 | blk.17.ffn_norm.weight | 0xba137720 | 0x4000 |
|
| 268 |
+
| 165 | blk.17.ffn_up.weight | 0xba13b720 | 0x2680000 |
|
| 269 |
+
| 166 | blk.18.attn_k.weight | 0xbc7bb720 | 0x2c0000 |
|
| 270 |
+
| 167 | blk.18.attn_norm.weight | 0xbca7b720 | 0x4000 |
|
| 271 |
+
| 168 | blk.18.attn_output.weight | 0xbca7f720 | 0xb00000 |
|
| 272 |
+
| 169 | blk.18.attn_q.weight | 0xbd57f720 | 0xb00000 |
|
| 273 |
+
| 170 | blk.18.attn_v.weight | 0xbe07f720 | 0x348000 |
|
| 274 |
+
| 171 | blk.18.ffn_down.weight | 0xbe3c7720 | 0x2df0000 |
|
| 275 |
+
| 172 | blk.18.ffn_gate.weight | 0xc11b7720 | 0x2680000 |
|
| 276 |
+
| 173 | blk.18.ffn_norm.weight | 0xc3837720 | 0x4000 |
|
| 277 |
+
| 174 | blk.18.ffn_up.weight | 0xc383b720 | 0x2680000 |
|
| 278 |
+
| 175 | blk.19.attn_k.weight | 0xc5ebb720 | 0x2c0000 |
|
| 279 |
+
| 176 | blk.19.attn_norm.weight | 0xc617b720 | 0x4000 |
|
| 280 |
+
| 177 | blk.19.attn_output.weight | 0xc617f720 | 0xb00000 |
|
| 281 |
+
| 178 | blk.19.attn_q.weight | 0xc6c7f720 | 0xb00000 |
|
| 282 |
+
| 179 | blk.19.attn_v.weight | 0xc777f720 | 0x348000 |
|
| 283 |
+
| 180 | blk.19.ffn_down.weight | 0xc7ac7720 | 0x2df0000 |
|
| 284 |
+
| 181 | blk.19.ffn_gate.weight | 0xca8b7720 | 0x2680000 |
|
| 285 |
+
| 182 | blk.19.ffn_norm.weight | 0xccf37720 | 0x4000 |
|
| 286 |
+
| 183 | blk.19.ffn_up.weight | 0xccf3b720 | 0x2680000 |
|
| 287 |
+
| 184 | blk.20.attn_k.weight | 0xcf5bb720 | 0x2c0000 |
|
| 288 |
+
| 185 | blk.20.attn_norm.weight | 0xcf87b720 | 0x4000 |
|
| 289 |
+
| 186 | blk.20.attn_output.weight | 0xcf87f720 | 0xb00000 |
|
| 290 |
+
| 187 | blk.20.attn_q.weight | 0xd037f720 | 0xb00000 |
|
| 291 |
+
| 188 | blk.20.attn_v.weight | 0xd0e7f720 | 0x348000 |
|
| 292 |
+
| 189 | blk.20.ffn_down.weight | 0xd11c7720 | 0x2df0000 |
|
| 293 |
+
| 190 | blk.20.ffn_gate.weight | 0xd3fb7720 | 0x2680000 |
|
| 294 |
+
| 191 | blk.20.ffn_norm.weight | 0xd6637720 | 0x4000 |
|
| 295 |
+
| 192 | blk.20.ffn_up.weight | 0xd663b720 | 0x2680000 |
|
| 296 |
+
| 193 | blk.21.attn_k.weight | 0xd8cbb720 | 0x2c0000 |
|
| 297 |
+
| 194 | blk.21.attn_norm.weight | 0xd8f7b720 | 0x4000 |
|
| 298 |
+
| 195 | blk.21.attn_output.weight | 0xd8f7f720 | 0xb00000 |
|
| 299 |
+
| 196 | blk.21.attn_q.weight | 0xd9a7f720 | 0xb00000 |
|
| 300 |
+
| 197 | blk.21.attn_v.weight | 0xda57f720 | 0x348000 |
|
| 301 |
+
| 198 | blk.21.ffn_down.weight | 0xda8c7720 | 0x2df0000 |
|
| 302 |
+
| 199 | blk.21.ffn_gate.weight | 0xdd6b7720 | 0x2680000 |
|
| 303 |
+
| 200 | blk.21.ffn_norm.weight | 0xdfd37720 | 0x4000 |
|
| 304 |
+
| 201 | blk.21.ffn_up.weight | 0xdfd3b720 | 0x2680000 |
|
| 305 |
+
| 202 | blk.22.attn_k.weight | 0xe23bb720 | 0x2c0000 |
|
| 306 |
+
| 203 | blk.22.attn_norm.weight | 0xe267b720 | 0x4000 |
|
| 307 |
+
| 204 | blk.22.attn_output.weight | 0xe267f720 | 0xb00000 |
|
| 308 |
+
| 205 | blk.22.attn_q.weight | 0xe317f720 | 0xb00000 |
|
| 309 |
+
| 206 | blk.22.attn_v.weight | 0xe3c7f720 | 0x348000 |
|
| 310 |
+
| 207 | blk.22.ffn_down.weight | 0xe3fc7720 | 0x2df0000 |
|
| 311 |
+
| 208 | blk.22.ffn_gate.weight | 0xe6db7720 | 0x2680000 |
|
| 312 |
+
| 209 | blk.22.ffn_norm.weight | 0xe9437720 | 0x4000 |
|
| 313 |
+
| 210 | blk.22.ffn_up.weight | 0xe943b720 | 0x2680000 |
|
| 314 |
+
| 211 | blk.23.attn_k.weight | 0xebabb720 | 0x2c0000 |
|
| 315 |
+
| 212 | blk.23.attn_norm.weight | 0xebd7b720 | 0x4000 |
|
| 316 |
+
| 213 | blk.23.attn_output.weight | 0xebd7f720 | 0xb00000 |
|
| 317 |
+
| 214 | blk.23.attn_q.weight | 0xec87f720 | 0xb00000 |
|
| 318 |
+
| 215 | blk.23.attn_v.weight | 0xed37f720 | 0x348000 |
|
| 319 |
+
| 216 | blk.23.ffn_down.weight | 0xed6c7720 | 0x2df0000 |
|
| 320 |
+
| 217 | blk.23.ffn_gate.weight | 0xf04b7720 | 0x2680000 |
|
| 321 |
+
| 218 | blk.23.ffn_norm.weight | 0xf2b37720 | 0x4000 |
|
| 322 |
+
| 219 | blk.23.ffn_up.weight | 0xf2b3b720 | 0x2680000 |
|
| 323 |
+
| 220 | blk.24.attn_k.weight | 0xf51bb720 | 0x2c0000 |
|
| 324 |
+
| 221 | blk.24.attn_norm.weight | 0xf547b720 | 0x4000 |
|
| 325 |
+
| 222 | blk.24.attn_output.weight | 0xf547f720 | 0xb00000 |
|
| 326 |
+
| 223 | blk.24.attn_q.weight | 0xf5f7f720 | 0xb00000 |
|
| 327 |
+
| 224 | blk.24.attn_v.weight | 0xf6a7f720 | 0x348000 |
|
| 328 |
+
| 225 | blk.24.ffn_down.weight | 0xf6dc7720 | 0x2df0000 |
|
| 329 |
+
| 226 | blk.24.ffn_gate.weight | 0xf9bb7720 | 0x2680000 |
|
| 330 |
+
| 227 | blk.24.ffn_norm.weight | 0xfc237720 | 0x4000 |
|
| 331 |
+
| 228 | blk.24.ffn_up.weight | 0xfc23b720 | 0x2680000 |
|
| 332 |
+
| 229 | blk.25.attn_k.weight | 0xfe8bb720 | 0x2c0000 |
|
| 333 |
+
| 230 | blk.25.attn_norm.weight | 0xfeb7b720 | 0x4000 |
|
| 334 |
+
| 231 | blk.25.attn_output.weight | 0xfeb7f720 | 0xb00000 |
|
| 335 |
+
| 232 | blk.25.attn_q.weight | 0xff67f720 | 0xb00000 |
|
| 336 |
+
| 233 | blk.25.attn_v.weight | 0x10017f720 | 0x348000 |
|
| 337 |
+
| 234 | blk.25.ffn_down.weight | 0x1004c7720 | 0x2df0000 |
|
| 338 |
+
| 235 | blk.25.ffn_gate.weight | 0x1032b7720 | 0x2680000 |
|
| 339 |
+
| 236 | blk.25.ffn_norm.weight | 0x105937720 | 0x4000 |
|
| 340 |
+
| 237 | blk.25.ffn_up.weight | 0x10593b720 | 0x2680000 |
|
| 341 |
+
| 238 | blk.26.attn_k.weight | 0x107fbb720 | 0x2c0000 |
|
| 342 |
+
| 239 | blk.26.attn_norm.weight | 0x10827b720 | 0x4000 |
|
| 343 |
+
| 240 | blk.26.attn_output.weight | 0x10827f720 | 0xb00000 |
|
| 344 |
+
| 241 | blk.26.attn_q.weight | 0x108d7f720 | 0xb00000 |
|
| 345 |
+
| 242 | blk.26.attn_v.weight | 0x10987f720 | 0x348000 |
|
| 346 |
+
| 243 | blk.26.ffn_down.weight | 0x109bc7720 | 0x2df0000 |
|
| 347 |
+
| 244 | blk.26.ffn_gate.weight | 0x10c9b7720 | 0x2680000 |
|
| 348 |
+
| 245 | blk.26.ffn_norm.weight | 0x10f037720 | 0x4000 |
|
| 349 |
+
| 246 | blk.26.ffn_up.weight | 0x10f03b720 | 0x2680000 |
|
| 350 |
+
| 247 | blk.27.attn_k.weight | 0x1116bb720 | 0x2c0000 |
|
| 351 |
+
| 248 | blk.27.attn_norm.weight | 0x11197b720 | 0x4000 |
|
| 352 |
+
| 249 | blk.27.attn_output.weight | 0x11197f720 | 0xb00000 |
|
| 353 |
+
| 250 | blk.27.attn_q.weight | 0x11247f720 | 0xb00000 |
|
| 354 |
+
| 251 | blk.27.attn_v.weight | 0x112f7f720 | 0x348000 |
|
| 355 |
+
| 252 | blk.27.ffn_down.weight | 0x1132c7720 | 0x2df0000 |
|
| 356 |
+
| 253 | blk.27.ffn_gate.weight | 0x1160b7720 | 0x2680000 |
|
| 357 |
+
| 254 | blk.27.ffn_norm.weight | 0x118737720 | 0x4000 |
|
| 358 |
+
| 255 | blk.27.ffn_up.weight | 0x11873b720 | 0x2680000 |
|
| 359 |
+
| 256 | blk.28.attn_k.weight | 0x11adbb720 | 0x2c0000 |
|
| 360 |
+
| 257 | blk.28.attn_norm.weight | 0x11b07b720 | 0x4000 |
|
| 361 |
+
| 258 | blk.28.attn_output.weight | 0x11b07f720 | 0xb00000 |
|
| 362 |
+
| 259 | blk.28.attn_q.weight | 0x11bb7f720 | 0xb00000 |
|
| 363 |
+
| 260 | blk.28.attn_v.weight | 0x11c67f720 | 0x348000 |
|
| 364 |
+
| 261 | blk.28.ffn_down.weight | 0x11c9c7720 | 0x2df0000 |
|
| 365 |
+
| 262 | blk.28.ffn_gate.weight | 0x11f7b7720 | 0x2680000 |
|
| 366 |
+
| 263 | blk.28.ffn_norm.weight | 0x121e37720 | 0x4000 |
|
| 367 |
+
| 264 | blk.28.ffn_up.weight | 0x121e3b720 | 0x2680000 |
|
| 368 |
+
| 265 | blk.29.attn_k.weight | 0x1244bb720 | 0x2c0000 |
|
| 369 |
+
| 266 | blk.29.attn_norm.weight | 0x12477b720 | 0x4000 |
|
| 370 |
+
| 267 | blk.29.attn_output.weight | 0x12477f720 | 0xb00000 |
|
| 371 |
+
| 268 | blk.29.attn_q.weight | 0x12527f720 | 0xb00000 |
|
| 372 |
+
| 269 | blk.29.attn_v.weight | 0x125d7f720 | 0x348000 |
|
| 373 |
+
| 270 | blk.29.ffn_down.weight | 0x1260c7720 | 0x2df0000 |
|
| 374 |
+
| 271 | blk.29.ffn_gate.weight | 0x128eb7720 | 0x2680000 |
|
| 375 |
+
| 272 | blk.29.ffn_norm.weight | 0x12b537720 | 0x4000 |
|
| 376 |
+
| 273 | blk.29.ffn_up.weight | 0x12b53b720 | 0x2680000 |
|
| 377 |
+
| 274 | blk.30.attn_k.weight | 0x12dbbb720 | 0x2c0000 |
|
| 378 |
+
| 275 | blk.30.attn_norm.weight | 0x12de7b720 | 0x4000 |
|
| 379 |
+
| 276 | blk.30.attn_output.weight | 0x12de7f720 | 0xb00000 |
|
| 380 |
+
| 277 | blk.30.attn_q.weight | 0x12e97f720 | 0xb00000 |
|
| 381 |
+
| 278 | blk.30.attn_v.weight | 0x12f47f720 | 0x348000 |
|
| 382 |
+
| 279 | blk.30.ffn_down.weight | 0x12f7c7720 | 0x2df0000 |
|
| 383 |
+
| 280 | blk.30.ffn_gate.weight | 0x1325b7720 | 0x2680000 |
|
| 384 |
+
| 281 | blk.30.ffn_norm.weight | 0x134c37720 | 0x4000 |
|
| 385 |
+
| 282 | blk.30.ffn_up.weight | 0x134c3b720 | 0x2680000 |
|
| 386 |
+
| 283 | blk.31.attn_k.weight | 0x1372bb720 | 0x240000 |
|
| 387 |
+
| 284 | blk.31.attn_norm.weight | 0x1374fb720 | 0x4000 |
|
| 388 |
+
| 285 | blk.31.attn_output.weight | 0x1374ff720 | 0xb00000 |
|
| 389 |
+
| 286 | blk.31.attn_q.weight | 0x137fff720 | 0x900000 |
|
| 390 |
+
| 287 | blk.31.attn_v.weight | 0x1388ff720 | 0x2c0000 |
|
| 391 |
+
| 288 | blk.31.ffn_down.weight | 0x138bbf720 | 0x2df0000 |
|
| 392 |
+
| 289 | blk.31.ffn_gate.weight | 0x13b9af720 | 0x2680000 |
|
| 393 |
+
| 290 | blk.31.ffn_norm.weight | 0x13e02f720 | 0x4000 |
|
| 394 |
+
| 291 | blk.31.ffn_up.weight | 0x13e033720 | 0x2680000 |
|
| 395 |
+
|
| 396 |
+
### <a name="base">Base Tensor Group : ~1B Elements</a>
|
| 397 |
+
|
| 398 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 399 |
+
|-----:|:-------------------|:---------------------------------|:------------------|:----------------------|:-----|
|
| 400 |
+
| 0 | output.weight | Output (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | Q5_K |
|
| 401 |
+
| 1 | output_norm.weight | Output Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 402 |
+
| 2 | rope_freqs.weight | Rope_Freqs (W) | ( 64) 64 | 64 x 1 x 1 x 1 | F32 |
|
| 403 |
+
| 3 | token_embd.weight | Token Embedding (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | Q3_K |
|
| 404 |
+
|
| 405 |
+
- Total elements in base: ( ~1B) 1050677312
|
| 406 |
+
- Percentage of total elements: 13.08%
|
| 407 |
+
|
| 408 |
+
|
| 409 |
+
### <a name="blk_0">Block 0 Tensor Group : ~218M Elements</a>
|
| 410 |
+
|
| 411 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 412 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 413 |
+
| 4 | blk.0.attn_k.weight | Block 0 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 414 |
+
| 5 | blk.0.attn_norm.weight | Block 0 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 415 |
+
| 6 | blk.0.attn_output.weight | Block 0 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 416 |
+
| 7 | blk.0.attn_q.weight | Block 0 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 417 |
+
| 8 | blk.0.attn_v.weight | Block 0 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 418 |
+
| 9 | blk.0.ffn_down.weight | Block 0 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 419 |
+
| 10 | blk.0.ffn_gate.weight | Block 0 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 420 |
+
| 11 | blk.0.ffn_norm.weight | Block 0 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 421 |
+
| 12 | blk.0.ffn_up.weight | Block 0 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 422 |
+
|
| 423 |
+
- Total elements in blk.0: (~218M) 218112000
|
| 424 |
+
- Percentage of total elements: 2.72%
|
| 425 |
+
|
| 426 |
+
|
| 427 |
+
### <a name="blk_1">Block 1 Tensor Group : ~218M Elements</a>
|
| 428 |
+
|
| 429 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 430 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 431 |
+
| 13 | blk.1.attn_k.weight | Block 1 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 432 |
+
| 14 | blk.1.attn_norm.weight | Block 1 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 433 |
+
| 15 | blk.1.attn_output.weight | Block 1 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 434 |
+
| 16 | blk.1.attn_q.weight | Block 1 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 435 |
+
| 17 | blk.1.attn_v.weight | Block 1 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 436 |
+
| 18 | blk.1.ffn_down.weight | Block 1 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 437 |
+
| 19 | blk.1.ffn_gate.weight | Block 1 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 438 |
+
| 20 | blk.1.ffn_norm.weight | Block 1 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 439 |
+
| 21 | blk.1.ffn_up.weight | Block 1 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 440 |
+
|
| 441 |
+
- Total elements in blk.1: (~218M) 218112000
|
| 442 |
+
- Percentage of total elements: 2.72%
|
| 443 |
+
|
| 444 |
+
|
| 445 |
+
### <a name="blk_2">Block 2 Tensor Group : ~218M Elements</a>
|
| 446 |
+
|
| 447 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 448 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 449 |
+
| 22 | blk.2.attn_k.weight | Block 2 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 450 |
+
| 23 | blk.2.attn_norm.weight | Block 2 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 451 |
+
| 24 | blk.2.attn_output.weight | Block 2 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 452 |
+
| 25 | blk.2.attn_q.weight | Block 2 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 453 |
+
| 26 | blk.2.attn_v.weight | Block 2 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 454 |
+
| 27 | blk.2.ffn_down.weight | Block 2 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 455 |
+
| 28 | blk.2.ffn_gate.weight | Block 2 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 456 |
+
| 29 | blk.2.ffn_norm.weight | Block 2 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 457 |
+
| 30 | blk.2.ffn_up.weight | Block 2 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 458 |
+
|
| 459 |
+
- Total elements in blk.2: (~218M) 218112000
|
| 460 |
+
- Percentage of total elements: 2.72%
|
| 461 |
+
|
| 462 |
+
|
| 463 |
+
### <a name="blk_3">Block 3 Tensor Group : ~218M Elements</a>
|
| 464 |
+
|
| 465 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 466 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 467 |
+
| 31 | blk.3.attn_k.weight | Block 3 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 468 |
+
| 32 | blk.3.attn_norm.weight | Block 3 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 469 |
+
| 33 | blk.3.attn_output.weight | Block 3 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 470 |
+
| 34 | blk.3.attn_q.weight | Block 3 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 471 |
+
| 35 | blk.3.attn_v.weight | Block 3 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 472 |
+
| 36 | blk.3.ffn_down.weight | Block 3 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 473 |
+
| 37 | blk.3.ffn_gate.weight | Block 3 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 474 |
+
| 38 | blk.3.ffn_norm.weight | Block 3 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 475 |
+
| 39 | blk.3.ffn_up.weight | Block 3 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 476 |
+
|
| 477 |
+
- Total elements in blk.3: (~218M) 218112000
|
| 478 |
+
- Percentage of total elements: 2.72%
|
| 479 |
+
|
| 480 |
+
|
| 481 |
+
### <a name="blk_4">Block 4 Tensor Group : ~218M Elements</a>
|
| 482 |
+
|
| 483 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 484 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 485 |
+
| 40 | blk.4.attn_k.weight | Block 4 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 486 |
+
| 41 | blk.4.attn_norm.weight | Block 4 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 487 |
+
| 42 | blk.4.attn_output.weight | Block 4 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 488 |
+
| 43 | blk.4.attn_q.weight | Block 4 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 489 |
+
| 44 | blk.4.attn_v.weight | Block 4 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 490 |
+
| 45 | blk.4.ffn_down.weight | Block 4 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 491 |
+
| 46 | blk.4.ffn_gate.weight | Block 4 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 492 |
+
| 47 | blk.4.ffn_norm.weight | Block 4 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 493 |
+
| 48 | blk.4.ffn_up.weight | Block 4 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 494 |
+
|
| 495 |
+
- Total elements in blk.4: (~218M) 218112000
|
| 496 |
+
- Percentage of total elements: 2.72%
|
| 497 |
+
|
| 498 |
+
|
| 499 |
+
### <a name="blk_5">Block 5 Tensor Group : ~218M Elements</a>
|
| 500 |
+
|
| 501 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 502 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 503 |
+
| 49 | blk.5.attn_k.weight | Block 5 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 504 |
+
| 50 | blk.5.attn_norm.weight | Block 5 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 505 |
+
| 51 | blk.5.attn_output.weight | Block 5 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 506 |
+
| 52 | blk.5.attn_q.weight | Block 5 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 507 |
+
| 53 | blk.5.attn_v.weight | Block 5 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 508 |
+
| 54 | blk.5.ffn_down.weight | Block 5 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 509 |
+
| 55 | blk.5.ffn_gate.weight | Block 5 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 510 |
+
| 56 | blk.5.ffn_norm.weight | Block 5 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 511 |
+
| 57 | blk.5.ffn_up.weight | Block 5 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 512 |
+
|
| 513 |
+
- Total elements in blk.5: (~218M) 218112000
|
| 514 |
+
- Percentage of total elements: 2.72%
|
| 515 |
+
|
| 516 |
+
|
| 517 |
+
### <a name="blk_6">Block 6 Tensor Group : ~218M Elements</a>
|
| 518 |
+
|
| 519 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 520 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 521 |
+
| 58 | blk.6.attn_k.weight | Block 6 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 522 |
+
| 59 | blk.6.attn_norm.weight | Block 6 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 523 |
+
| 60 | blk.6.attn_output.weight | Block 6 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 524 |
+
| 61 | blk.6.attn_q.weight | Block 6 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 525 |
+
| 62 | blk.6.attn_v.weight | Block 6 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 526 |
+
| 63 | blk.6.ffn_down.weight | Block 6 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 527 |
+
| 64 | blk.6.ffn_gate.weight | Block 6 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 528 |
+
| 65 | blk.6.ffn_norm.weight | Block 6 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 529 |
+
| 66 | blk.6.ffn_up.weight | Block 6 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 530 |
+
|
| 531 |
+
- Total elements in blk.6: (~218M) 218112000
|
| 532 |
+
- Percentage of total elements: 2.72%
|
| 533 |
+
|
| 534 |
+
|
| 535 |
+
### <a name="blk_7">Block 7 Tensor Group : ~218M Elements</a>
|
| 536 |
+
|
| 537 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 538 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 539 |
+
| 67 | blk.7.attn_k.weight | Block 7 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 540 |
+
| 68 | blk.7.attn_norm.weight | Block 7 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 541 |
+
| 69 | blk.7.attn_output.weight | Block 7 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 542 |
+
| 70 | blk.7.attn_q.weight | Block 7 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 543 |
+
| 71 | blk.7.attn_v.weight | Block 7 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 544 |
+
| 72 | blk.7.ffn_down.weight | Block 7 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 545 |
+
| 73 | blk.7.ffn_gate.weight | Block 7 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 546 |
+
| 74 | blk.7.ffn_norm.weight | Block 7 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 547 |
+
| 75 | blk.7.ffn_up.weight | Block 7 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 548 |
+
|
| 549 |
+
- Total elements in blk.7: (~218M) 218112000
|
| 550 |
+
- Percentage of total elements: 2.72%
|
| 551 |
+
|
| 552 |
+
|
| 553 |
+
### <a name="blk_8">Block 8 Tensor Group : ~218M Elements</a>
|
| 554 |
+
|
| 555 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 556 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 557 |
+
| 76 | blk.8.attn_k.weight | Block 8 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 558 |
+
| 77 | blk.8.attn_norm.weight | Block 8 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 559 |
+
| 78 | blk.8.attn_output.weight | Block 8 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 560 |
+
| 79 | blk.8.attn_q.weight | Block 8 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 561 |
+
| 80 | blk.8.attn_v.weight | Block 8 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 562 |
+
| 81 | blk.8.ffn_down.weight | Block 8 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 563 |
+
| 82 | blk.8.ffn_gate.weight | Block 8 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 564 |
+
| 83 | blk.8.ffn_norm.weight | Block 8 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 565 |
+
| 84 | blk.8.ffn_up.weight | Block 8 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 566 |
+
|
| 567 |
+
- Total elements in blk.8: (~218M) 218112000
|
| 568 |
+
- Percentage of total elements: 2.72%
|
| 569 |
+
|
| 570 |
+
|
| 571 |
+
### <a name="blk_9">Block 9 Tensor Group : ~218M Elements</a>
|
| 572 |
+
|
| 573 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 574 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 575 |
+
| 85 | blk.9.attn_k.weight | Block 9 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 576 |
+
| 86 | blk.9.attn_norm.weight | Block 9 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 577 |
+
| 87 | blk.9.attn_output.weight | Block 9 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 578 |
+
| 88 | blk.9.attn_q.weight | Block 9 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 579 |
+
| 89 | blk.9.attn_v.weight | Block 9 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 580 |
+
| 90 | blk.9.ffn_down.weight | Block 9 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 581 |
+
| 91 | blk.9.ffn_gate.weight | Block 9 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 582 |
+
| 92 | blk.9.ffn_norm.weight | Block 9 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 583 |
+
| 93 | blk.9.ffn_up.weight | Block 9 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 584 |
+
|
| 585 |
+
- Total elements in blk.9: (~218M) 218112000
|
| 586 |
+
- Percentage of total elements: 2.72%
|
| 587 |
+
|
| 588 |
+
|
| 589 |
+
### <a name="blk_10">Block 10 Tensor Group : ~218M Elements</a>
|
| 590 |
+
|
| 591 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 592 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 593 |
+
| 94 | blk.10.attn_k.weight | Block 10 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 594 |
+
| 95 | blk.10.attn_norm.weight | Block 10 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 595 |
+
| 96 | blk.10.attn_output.weight | Block 10 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 596 |
+
| 97 | blk.10.attn_q.weight | Block 10 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 597 |
+
| 98 | blk.10.attn_v.weight | Block 10 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 598 |
+
| 99 | blk.10.ffn_down.weight | Block 10 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 599 |
+
| 100 | blk.10.ffn_gate.weight | Block 10 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 600 |
+
| 101 | blk.10.ffn_norm.weight | Block 10 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 601 |
+
| 102 | blk.10.ffn_up.weight | Block 10 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 602 |
+
|
| 603 |
+
- Total elements in blk.10: (~218M) 218112000
|
| 604 |
+
- Percentage of total elements: 2.72%
|
| 605 |
+
|
| 606 |
+
|
| 607 |
+
### <a name="blk_11">Block 11 Tensor Group : ~218M Elements</a>
|
| 608 |
+
|
| 609 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 610 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 611 |
+
| 103 | blk.11.attn_k.weight | Block 11 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 612 |
+
| 104 | blk.11.attn_norm.weight | Block 11 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 613 |
+
| 105 | blk.11.attn_output.weight | Block 11 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 614 |
+
| 106 | blk.11.attn_q.weight | Block 11 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 615 |
+
| 107 | blk.11.attn_v.weight | Block 11 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 616 |
+
| 108 | blk.11.ffn_down.weight | Block 11 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 617 |
+
| 109 | blk.11.ffn_gate.weight | Block 11 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 618 |
+
| 110 | blk.11.ffn_norm.weight | Block 11 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 619 |
+
| 111 | blk.11.ffn_up.weight | Block 11 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 620 |
+
|
| 621 |
+
- Total elements in blk.11: (~218M) 218112000
|
| 622 |
+
- Percentage of total elements: 2.72%
|
| 623 |
+
|
| 624 |
+
|
| 625 |
+
### <a name="blk_12">Block 12 Tensor Group : ~218M Elements</a>
|
| 626 |
+
|
| 627 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 628 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 629 |
+
| 112 | blk.12.attn_k.weight | Block 12 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 630 |
+
| 113 | blk.12.attn_norm.weight | Block 12 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 631 |
+
| 114 | blk.12.attn_output.weight | Block 12 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 632 |
+
| 115 | blk.12.attn_q.weight | Block 12 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 633 |
+
| 116 | blk.12.attn_v.weight | Block 12 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 634 |
+
| 117 | blk.12.ffn_down.weight | Block 12 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 635 |
+
| 118 | blk.12.ffn_gate.weight | Block 12 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 636 |
+
| 119 | blk.12.ffn_norm.weight | Block 12 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 637 |
+
| 120 | blk.12.ffn_up.weight | Block 12 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 638 |
+
|
| 639 |
+
- Total elements in blk.12: (~218M) 218112000
|
| 640 |
+
- Percentage of total elements: 2.72%
|
| 641 |
+
|
| 642 |
+
|
| 643 |
+
### <a name="blk_13">Block 13 Tensor Group : ~218M Elements</a>
|
| 644 |
+
|
| 645 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 646 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 647 |
+
| 121 | blk.13.attn_k.weight | Block 13 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 648 |
+
| 122 | blk.13.attn_norm.weight | Block 13 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 649 |
+
| 123 | blk.13.attn_output.weight | Block 13 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 650 |
+
| 124 | blk.13.attn_q.weight | Block 13 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 651 |
+
| 125 | blk.13.attn_v.weight | Block 13 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 652 |
+
| 126 | blk.13.ffn_down.weight | Block 13 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 653 |
+
| 127 | blk.13.ffn_gate.weight | Block 13 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 654 |
+
| 128 | blk.13.ffn_norm.weight | Block 13 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 655 |
+
| 129 | blk.13.ffn_up.weight | Block 13 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 656 |
+
|
| 657 |
+
- Total elements in blk.13: (~218M) 218112000
|
| 658 |
+
- Percentage of total elements: 2.72%
|
| 659 |
+
|
| 660 |
+
|
| 661 |
+
### <a name="blk_14">Block 14 Tensor Group : ~218M Elements</a>
|
| 662 |
+
|
| 663 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 664 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 665 |
+
| 130 | blk.14.attn_k.weight | Block 14 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 666 |
+
| 131 | blk.14.attn_norm.weight | Block 14 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 667 |
+
| 132 | blk.14.attn_output.weight | Block 14 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 668 |
+
| 133 | blk.14.attn_q.weight | Block 14 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 669 |
+
| 134 | blk.14.attn_v.weight | Block 14 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 670 |
+
| 135 | blk.14.ffn_down.weight | Block 14 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 671 |
+
| 136 | blk.14.ffn_gate.weight | Block 14 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 672 |
+
| 137 | blk.14.ffn_norm.weight | Block 14 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 673 |
+
| 138 | blk.14.ffn_up.weight | Block 14 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 674 |
+
|
| 675 |
+
- Total elements in blk.14: (~218M) 218112000
|
| 676 |
+
- Percentage of total elements: 2.72%
|
| 677 |
+
|
| 678 |
+
|
| 679 |
+
### <a name="blk_15">Block 15 Tensor Group : ~218M Elements</a>
|
| 680 |
+
|
| 681 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 682 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 683 |
+
| 139 | blk.15.attn_k.weight | Block 15 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 684 |
+
| 140 | blk.15.attn_norm.weight | Block 15 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 685 |
+
| 141 | blk.15.attn_output.weight | Block 15 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 686 |
+
| 142 | blk.15.attn_q.weight | Block 15 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 687 |
+
| 143 | blk.15.attn_v.weight | Block 15 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 688 |
+
| 144 | blk.15.ffn_down.weight | Block 15 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 689 |
+
| 145 | blk.15.ffn_gate.weight | Block 15 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 690 |
+
| 146 | blk.15.ffn_norm.weight | Block 15 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 691 |
+
| 147 | blk.15.ffn_up.weight | Block 15 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 692 |
+
|
| 693 |
+
- Total elements in blk.15: (~218M) 218112000
|
| 694 |
+
- Percentage of total elements: 2.72%
|
| 695 |
+
|
| 696 |
+
|
| 697 |
+
### <a name="blk_16">Block 16 Tensor Group : ~218M Elements</a>
|
| 698 |
+
|
| 699 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 700 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 701 |
+
| 148 | blk.16.attn_k.weight | Block 16 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 702 |
+
| 149 | blk.16.attn_norm.weight | Block 16 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 703 |
+
| 150 | blk.16.attn_output.weight | Block 16 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 704 |
+
| 151 | blk.16.attn_q.weight | Block 16 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 705 |
+
| 152 | blk.16.attn_v.weight | Block 16 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 706 |
+
| 153 | blk.16.ffn_down.weight | Block 16 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 707 |
+
| 154 | blk.16.ffn_gate.weight | Block 16 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 708 |
+
| 155 | blk.16.ffn_norm.weight | Block 16 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 709 |
+
| 156 | blk.16.ffn_up.weight | Block 16 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 710 |
+
|
| 711 |
+
- Total elements in blk.16: (~218M) 218112000
|
| 712 |
+
- Percentage of total elements: 2.72%
|
| 713 |
+
|
| 714 |
+
|
| 715 |
+
### <a name="blk_17">Block 17 Tensor Group : ~218M Elements</a>
|
| 716 |
+
|
| 717 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 718 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 719 |
+
| 157 | blk.17.attn_k.weight | Block 17 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 720 |
+
| 158 | blk.17.attn_norm.weight | Block 17 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 721 |
+
| 159 | blk.17.attn_output.weight | Block 17 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 722 |
+
| 160 | blk.17.attn_q.weight | Block 17 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 723 |
+
| 161 | blk.17.attn_v.weight | Block 17 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 724 |
+
| 162 | blk.17.ffn_down.weight | Block 17 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 725 |
+
| 163 | blk.17.ffn_gate.weight | Block 17 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 726 |
+
| 164 | blk.17.ffn_norm.weight | Block 17 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 727 |
+
| 165 | blk.17.ffn_up.weight | Block 17 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 728 |
+
|
| 729 |
+
- Total elements in blk.17: (~218M) 218112000
|
| 730 |
+
- Percentage of total elements: 2.72%
|
| 731 |
+
|
| 732 |
+
|
| 733 |
+
### <a name="blk_18">Block 18 Tensor Group : ~218M Elements</a>
|
| 734 |
+
|
| 735 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 736 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 737 |
+
| 166 | blk.18.attn_k.weight | Block 18 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 738 |
+
| 167 | blk.18.attn_norm.weight | Block 18 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 739 |
+
| 168 | blk.18.attn_output.weight | Block 18 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 740 |
+
| 169 | blk.18.attn_q.weight | Block 18 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 741 |
+
| 170 | blk.18.attn_v.weight | Block 18 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 742 |
+
| 171 | blk.18.ffn_down.weight | Block 18 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 743 |
+
| 172 | blk.18.ffn_gate.weight | Block 18 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 744 |
+
| 173 | blk.18.ffn_norm.weight | Block 18 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 745 |
+
| 174 | blk.18.ffn_up.weight | Block 18 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 746 |
+
|
| 747 |
+
- Total elements in blk.18: (~218M) 218112000
|
| 748 |
+
- Percentage of total elements: 2.72%
|
| 749 |
+
|
| 750 |
+
|
| 751 |
+
### <a name="blk_19">Block 19 Tensor Group : ~218M Elements</a>
|
| 752 |
+
|
| 753 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 754 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 755 |
+
| 175 | blk.19.attn_k.weight | Block 19 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 756 |
+
| 176 | blk.19.attn_norm.weight | Block 19 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 757 |
+
| 177 | blk.19.attn_output.weight | Block 19 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 758 |
+
| 178 | blk.19.attn_q.weight | Block 19 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 759 |
+
| 179 | blk.19.attn_v.weight | Block 19 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 760 |
+
| 180 | blk.19.ffn_down.weight | Block 19 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 761 |
+
| 181 | blk.19.ffn_gate.weight | Block 19 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 762 |
+
| 182 | blk.19.ffn_norm.weight | Block 19 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 763 |
+
| 183 | blk.19.ffn_up.weight | Block 19 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 764 |
+
|
| 765 |
+
- Total elements in blk.19: (~218M) 218112000
|
| 766 |
+
- Percentage of total elements: 2.72%
|
| 767 |
+
|
| 768 |
+
|
| 769 |
+
### <a name="blk_20">Block 20 Tensor Group : ~218M Elements</a>
|
| 770 |
+
|
| 771 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 772 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 773 |
+
| 184 | blk.20.attn_k.weight | Block 20 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 774 |
+
| 185 | blk.20.attn_norm.weight | Block 20 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 775 |
+
| 186 | blk.20.attn_output.weight | Block 20 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 776 |
+
| 187 | blk.20.attn_q.weight | Block 20 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 777 |
+
| 188 | blk.20.attn_v.weight | Block 20 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 778 |
+
| 189 | blk.20.ffn_down.weight | Block 20 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 779 |
+
| 190 | blk.20.ffn_gate.weight | Block 20 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 780 |
+
| 191 | blk.20.ffn_norm.weight | Block 20 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 781 |
+
| 192 | blk.20.ffn_up.weight | Block 20 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 782 |
+
|
| 783 |
+
- Total elements in blk.20: (~218M) 218112000
|
| 784 |
+
- Percentage of total elements: 2.72%
|
| 785 |
+
|
| 786 |
+
|
| 787 |
+
### <a name="blk_21">Block 21 Tensor Group : ~218M Elements</a>
|
| 788 |
+
|
| 789 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 790 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 791 |
+
| 193 | blk.21.attn_k.weight | Block 21 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 792 |
+
| 194 | blk.21.attn_norm.weight | Block 21 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 793 |
+
| 195 | blk.21.attn_output.weight | Block 21 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 794 |
+
| 196 | blk.21.attn_q.weight | Block 21 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 795 |
+
| 197 | blk.21.attn_v.weight | Block 21 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 796 |
+
| 198 | blk.21.ffn_down.weight | Block 21 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 797 |
+
| 199 | blk.21.ffn_gate.weight | Block 21 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 798 |
+
| 200 | blk.21.ffn_norm.weight | Block 21 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 799 |
+
| 201 | blk.21.ffn_up.weight | Block 21 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 800 |
+
|
| 801 |
+
- Total elements in blk.21: (~218M) 218112000
|
| 802 |
+
- Percentage of total elements: 2.72%
|
| 803 |
+
|
| 804 |
+
|
| 805 |
+
### <a name="blk_22">Block 22 Tensor Group : ~218M Elements</a>
|
| 806 |
+
|
| 807 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 808 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 809 |
+
| 202 | blk.22.attn_k.weight | Block 22 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 810 |
+
| 203 | blk.22.attn_norm.weight | Block 22 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 811 |
+
| 204 | blk.22.attn_output.weight | Block 22 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 812 |
+
| 205 | blk.22.attn_q.weight | Block 22 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 813 |
+
| 206 | blk.22.attn_v.weight | Block 22 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 814 |
+
| 207 | blk.22.ffn_down.weight | Block 22 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 815 |
+
| 208 | blk.22.ffn_gate.weight | Block 22 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 816 |
+
| 209 | blk.22.ffn_norm.weight | Block 22 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 817 |
+
| 210 | blk.22.ffn_up.weight | Block 22 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 818 |
+
|
| 819 |
+
- Total elements in blk.22: (~218M) 218112000
|
| 820 |
+
- Percentage of total elements: 2.72%
|
| 821 |
+
|
| 822 |
+
|
| 823 |
+
### <a name="blk_23">Block 23 Tensor Group : ~218M Elements</a>
|
| 824 |
+
|
| 825 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 826 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 827 |
+
| 211 | blk.23.attn_k.weight | Block 23 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 828 |
+
| 212 | blk.23.attn_norm.weight | Block 23 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 829 |
+
| 213 | blk.23.attn_output.weight | Block 23 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 830 |
+
| 214 | blk.23.attn_q.weight | Block 23 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 831 |
+
| 215 | blk.23.attn_v.weight | Block 23 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 832 |
+
| 216 | blk.23.ffn_down.weight | Block 23 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 833 |
+
| 217 | blk.23.ffn_gate.weight | Block 23 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 834 |
+
| 218 | blk.23.ffn_norm.weight | Block 23 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 835 |
+
| 219 | blk.23.ffn_up.weight | Block 23 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 836 |
+
|
| 837 |
+
- Total elements in blk.23: (~218M) 218112000
|
| 838 |
+
- Percentage of total elements: 2.72%
|
| 839 |
+
|
| 840 |
+
|
| 841 |
+
### <a name="blk_24">Block 24 Tensor Group : ~218M Elements</a>
|
| 842 |
+
|
| 843 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 844 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 845 |
+
| 220 | blk.24.attn_k.weight | Block 24 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 846 |
+
| 221 | blk.24.attn_norm.weight | Block 24 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 847 |
+
| 222 | blk.24.attn_output.weight | Block 24 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 848 |
+
| 223 | blk.24.attn_q.weight | Block 24 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 849 |
+
| 224 | blk.24.attn_v.weight | Block 24 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 850 |
+
| 225 | blk.24.ffn_down.weight | Block 24 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 851 |
+
| 226 | blk.24.ffn_gate.weight | Block 24 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 852 |
+
| 227 | blk.24.ffn_norm.weight | Block 24 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 853 |
+
| 228 | blk.24.ffn_up.weight | Block 24 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 854 |
+
|
| 855 |
+
- Total elements in blk.24: (~218M) 218112000
|
| 856 |
+
- Percentage of total elements: 2.72%
|
| 857 |
+
|
| 858 |
+
|
| 859 |
+
### <a name="blk_25">Block 25 Tensor Group : ~218M Elements</a>
|
| 860 |
+
|
| 861 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 862 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 863 |
+
| 229 | blk.25.attn_k.weight | Block 25 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 864 |
+
| 230 | blk.25.attn_norm.weight | Block 25 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 865 |
+
| 231 | blk.25.attn_output.weight | Block 25 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 866 |
+
| 232 | blk.25.attn_q.weight | Block 25 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 867 |
+
| 233 | blk.25.attn_v.weight | Block 25 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 868 |
+
| 234 | blk.25.ffn_down.weight | Block 25 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 869 |
+
| 235 | blk.25.ffn_gate.weight | Block 25 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 870 |
+
| 236 | blk.25.ffn_norm.weight | Block 25 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 871 |
+
| 237 | blk.25.ffn_up.weight | Block 25 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 872 |
+
|
| 873 |
+
- Total elements in blk.25: (~218M) 218112000
|
| 874 |
+
- Percentage of total elements: 2.72%
|
| 875 |
+
|
| 876 |
+
|
| 877 |
+
### <a name="blk_26">Block 26 Tensor Group : ~218M Elements</a>
|
| 878 |
+
|
| 879 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 880 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 881 |
+
| 238 | blk.26.attn_k.weight | Block 26 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 882 |
+
| 239 | blk.26.attn_norm.weight | Block 26 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 883 |
+
| 240 | blk.26.attn_output.weight | Block 26 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 884 |
+
| 241 | blk.26.attn_q.weight | Block 26 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 885 |
+
| 242 | blk.26.attn_v.weight | Block 26 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 886 |
+
| 243 | blk.26.ffn_down.weight | Block 26 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 887 |
+
| 244 | blk.26.ffn_gate.weight | Block 26 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 888 |
+
| 245 | blk.26.ffn_norm.weight | Block 26 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 889 |
+
| 246 | blk.26.ffn_up.weight | Block 26 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 890 |
+
|
| 891 |
+
- Total elements in blk.26: (~218M) 218112000
|
| 892 |
+
- Percentage of total elements: 2.72%
|
| 893 |
+
|
| 894 |
+
|
| 895 |
+
### <a name="blk_27">Block 27 Tensor Group : ~218M Elements</a>
|
| 896 |
+
|
| 897 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 898 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 899 |
+
| 247 | blk.27.attn_k.weight | Block 27 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 900 |
+
| 248 | blk.27.attn_norm.weight | Block 27 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 901 |
+
| 249 | blk.27.attn_output.weight | Block 27 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 902 |
+
| 250 | blk.27.attn_q.weight | Block 27 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 903 |
+
| 251 | blk.27.attn_v.weight | Block 27 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 904 |
+
| 252 | blk.27.ffn_down.weight | Block 27 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 905 |
+
| 253 | blk.27.ffn_gate.weight | Block 27 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 906 |
+
| 254 | blk.27.ffn_norm.weight | Block 27 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 907 |
+
| 255 | blk.27.ffn_up.weight | Block 27 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 908 |
+
|
| 909 |
+
- Total elements in blk.27: (~218M) 218112000
|
| 910 |
+
- Percentage of total elements: 2.72%
|
| 911 |
+
|
| 912 |
+
|
| 913 |
+
### <a name="blk_28">Block 28 Tensor Group : ~218M Elements</a>
|
| 914 |
+
|
| 915 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 916 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 917 |
+
| 256 | blk.28.attn_k.weight | Block 28 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 918 |
+
| 257 | blk.28.attn_norm.weight | Block 28 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 919 |
+
| 258 | blk.28.attn_output.weight | Block 28 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 920 |
+
| 259 | blk.28.attn_q.weight | Block 28 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 921 |
+
| 260 | blk.28.attn_v.weight | Block 28 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 922 |
+
| 261 | blk.28.ffn_down.weight | Block 28 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 923 |
+
| 262 | blk.28.ffn_gate.weight | Block 28 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 924 |
+
| 263 | blk.28.ffn_norm.weight | Block 28 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 925 |
+
| 264 | blk.28.ffn_up.weight | Block 28 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 926 |
+
|
| 927 |
+
- Total elements in blk.28: (~218M) 218112000
|
| 928 |
+
- Percentage of total elements: 2.72%
|
| 929 |
+
|
| 930 |
+
|
| 931 |
+
### <a name="blk_29">Block 29 Tensor Group : ~218M Elements</a>
|
| 932 |
+
|
| 933 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 934 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 935 |
+
| 265 | blk.29.attn_k.weight | Block 29 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 936 |
+
| 266 | blk.29.attn_norm.weight | Block 29 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 937 |
+
| 267 | blk.29.attn_output.weight | Block 29 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 938 |
+
| 268 | blk.29.attn_q.weight | Block 29 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 939 |
+
| 269 | blk.29.attn_v.weight | Block 29 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 940 |
+
| 270 | blk.29.ffn_down.weight | Block 29 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 941 |
+
| 271 | blk.29.ffn_gate.weight | Block 29 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 942 |
+
| 272 | blk.29.ffn_norm.weight | Block 29 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 943 |
+
| 273 | blk.29.ffn_up.weight | Block 29 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 944 |
+
|
| 945 |
+
- Total elements in blk.29: (~218M) 218112000
|
| 946 |
+
- Percentage of total elements: 2.72%
|
| 947 |
+
|
| 948 |
+
|
| 949 |
+
### <a name="blk_30">Block 30 Tensor Group : ~218M Elements</a>
|
| 950 |
+
|
| 951 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 952 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 953 |
+
| 274 | blk.30.attn_k.weight | Block 30 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 954 |
+
| 275 | blk.30.attn_norm.weight | Block 30 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 955 |
+
| 276 | blk.30.attn_output.weight | Block 30 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 956 |
+
| 277 | blk.30.attn_q.weight | Block 30 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 957 |
+
| 278 | blk.30.attn_v.weight | Block 30 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 958 |
+
| 279 | blk.30.ffn_down.weight | Block 30 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 959 |
+
| 280 | blk.30.ffn_gate.weight | Block 30 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 960 |
+
| 281 | blk.30.ffn_norm.weight | Block 30 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 961 |
+
| 282 | blk.30.ffn_up.weight | Block 30 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 962 |
+
|
| 963 |
+
- Total elements in blk.30: (~218M) 218112000
|
| 964 |
+
- Percentage of total elements: 2.72%
|
| 965 |
+
|
| 966 |
+
|
| 967 |
+
### <a name="blk_31">Block 31 Tensor Group : ~218M Elements</a>
|
| 968 |
+
|
| 969 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 970 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 971 |
+
| 283 | blk.31.attn_k.weight | Block 31 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 972 |
+
| 284 | blk.31.attn_norm.weight | Block 31 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 973 |
+
| 285 | blk.31.attn_output.weight | Block 31 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 974 |
+
| 286 | blk.31.attn_q.weight | Block 31 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 975 |
+
| 287 | blk.31.attn_v.weight | Block 31 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 976 |
+
| 288 | blk.31.ffn_down.weight | Block 31 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 977 |
+
| 289 | blk.31.ffn_gate.weight | Block 31 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 978 |
+
| 290 | blk.31.ffn_norm.weight | Block 31 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 979 |
+
| 291 | blk.31.ffn_up.weight | Block 31 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 980 |
+
|
| 981 |
+
- Total elements in blk.31: (~218M) 218112000
|
| 982 |
+
- Percentage of total elements: 2.72%
|
scores/Watt-Tool-8B-Q5_K_S.md
ADDED
|
@@ -0,0 +1,982 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Watt-Tool-8B-Q5_K_S.gguf - GGUF Internal File Dump
|
| 2 |
+
|
| 3 |
+
- Endian: LITTLE endian
|
| 4 |
+
|
| 5 |
+
## Key Value Metadata Store
|
| 6 |
+
|
| 7 |
+
There are 43 key-value pairs in this file
|
| 8 |
+
|
| 9 |
+
| POS | TYPE | Count | Key | Value |
|
| 10 |
+
|----:|:---------|-------:|:---------------------------------------|:--------------------------------------------------------------------|
|
| 11 |
+
| 1 | UINT32 | 1 | GGUF.version | 3 |
|
| 12 |
+
| 2 | UINT64 | 1 | GGUF.tensor_count | 292 |
|
| 13 |
+
| 3 | UINT64 | 1 | GGUF.kv_count | 40 |
|
| 14 |
+
| 4 | STRING | 1 | general.architecture | `llama` |
|
| 15 |
+
| 5 | STRING | 1 | general.type | `model` |
|
| 16 |
+
| 6 | STRING | 1 | general.name | `Watt Tool 8B GGUF` |
|
| 17 |
+
| 7 | STRING | 1 | general.finetune | `GGUF` |
|
| 18 |
+
| 8 | STRING | 1 | general.basename | `Watt-Tool` |
|
| 19 |
+
| 9 | STRING | 1 | general.size_label | `8B` |
|
| 20 |
+
| 10 | STRING | 1 | general.license | `apache-2.0` |
|
| 21 |
+
| 11 | UINT32 | 1 | general.base_model.count | 1 |
|
| 22 |
+
| 12 | STRING | 1 | general.base_model.0.name | `Llama 3.1 8B Instruct` |
|
| 23 |
+
| 13 | STRING | 1 | general.base_model.0.organization | `Meta Llama` |
|
| 24 |
+
| 14 | STRING | 1 | general.base_model.0.repo_url | `https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct` |
|
| 25 |
+
| 15 | [STRING] | 4 | general.tags | [ `function-calling`, `tool-use`, `llama`, `bfcl` ] |
|
| 26 |
+
| 16 | [STRING] | 1 | general.languages | [ `en` ] |
|
| 27 |
+
| 17 | UINT32 | 1 | llama.block_count | 32 |
|
| 28 |
+
| 18 | UINT32 | 1 | llama.context_length | 131072 |
|
| 29 |
+
| 19 | UINT32 | 1 | llama.embedding_length | 4096 |
|
| 30 |
+
| 20 | UINT32 | 1 | llama.feed_forward_length | 14336 |
|
| 31 |
+
| 21 | UINT32 | 1 | llama.attention.head_count | 32 |
|
| 32 |
+
| 22 | UINT32 | 1 | llama.attention.head_count_kv | 8 |
|
| 33 |
+
| 23 | FLOAT32 | 1 | llama.rope.freq_base | 500000.0 |
|
| 34 |
+
| 24 | FLOAT32 | 1 | llama.attention.layer_norm_rms_epsilon | 1e-05 |
|
| 35 |
+
| 25 | UINT32 | 1 | llama.attention.key_length | 128 |
|
| 36 |
+
| 26 | UINT32 | 1 | llama.attention.value_length | 128 |
|
| 37 |
+
| 27 | UINT32 | 1 | llama.vocab_size | 128256 |
|
| 38 |
+
| 28 | UINT32 | 1 | llama.rope.dimension_count | 128 |
|
| 39 |
+
| 29 | STRING | 1 | tokenizer.ggml.model | `gpt2` |
|
| 40 |
+
| 30 | STRING | 1 | tokenizer.ggml.pre | `llama-bpe` |
|
| 41 |
+
| 31 | [STRING] | 128256 | tokenizer.ggml.tokens | [ `!`, `"`, `#`, `$`, `%`, ... ] |
|
| 42 |
+
| 32 | [INT32] | 128256 | tokenizer.ggml.token_type | [ 1, 1, 1, 1, 1, 1, 1, ... ] |
|
| 43 |
+
| 33 | [STRING] | 280147 | tokenizer.ggml.merges | [ `Ġ Ġ`, `Ġ ĠĠĠ`, `ĠĠ ĠĠ`, `ĠĠĠ Ġ`, `i n`, ... ] |
|
| 44 |
+
| 34 | UINT32 | 1 | tokenizer.ggml.bos_token_id | 128000 |
|
| 45 |
+
| 35 | UINT32 | 1 | tokenizer.ggml.eos_token_id | 128009 |
|
| 46 |
+
| 36 | UINT32 | 1 | tokenizer.ggml.padding_token_id | 128009 |
|
| 47 |
+
| 37 | STRING | 1 | tokenizer.chat_template | `{{ '<|begin_of_text|>' }}{% if`...`d|>' }}{% endif %}{% endfor %}` |
|
| 48 |
+
| 38 | UINT32 | 1 | general.quantization_version | 2 |
|
| 49 |
+
| 39 | UINT32 | 1 | general.file_type | 16 |
|
| 50 |
+
| 40 | STRING | 1 | quantize.imatrix.file | `./imatrix/imatrix-Watt-Tool-8B-small.dat` |
|
| 51 |
+
| 41 | STRING | 1 | quantize.imatrix.dataset | `../../datasets/imatrix/calibration_eur_small.txt` |
|
| 52 |
+
| 42 | INT32 | 1 | quantize.imatrix.entries_count | 225 |
|
| 53 |
+
| 43 | INT32 | 1 | quantize.imatrix.chunks_count | 962 |
|
| 54 |
+
|
| 55 |
+
## Tensors Overview ~8B Elements
|
| 56 |
+
|
| 57 |
+
Total number of elements in all tensors: 8030261312 Elements
|
| 58 |
+
|
| 59 |
+
- [Watt-Tool-8B-Q5\_K\_S.gguf - GGUF Internal File Dump](#watt-tool-8b-q5_k_sgguf---gguf-internal-file-dump)
|
| 60 |
+
- [Key Value Metadata Store](#key-value-metadata-store)
|
| 61 |
+
- [Tensors Overview ~8B Elements](#tensors-overview-8b-elements)
|
| 62 |
+
- [Tensor Data Offset](#tensor-data-offset)
|
| 63 |
+
- [Base Tensor Group : ~1B Elements](#base-tensor-group--1b-elements)
|
| 64 |
+
- [Block 0 Tensor Group : ~218M Elements](#block-0-tensor-group--218m-elements)
|
| 65 |
+
- [Block 1 Tensor Group : ~218M Elements](#block-1-tensor-group--218m-elements)
|
| 66 |
+
- [Block 2 Tensor Group : ~218M Elements](#block-2-tensor-group--218m-elements)
|
| 67 |
+
- [Block 3 Tensor Group : ~218M Elements](#block-3-tensor-group--218m-elements)
|
| 68 |
+
- [Block 4 Tensor Group : ~218M Elements](#block-4-tensor-group--218m-elements)
|
| 69 |
+
- [Block 5 Tensor Group : ~218M Elements](#block-5-tensor-group--218m-elements)
|
| 70 |
+
- [Block 6 Tensor Group : ~218M Elements](#block-6-tensor-group--218m-elements)
|
| 71 |
+
- [Block 7 Tensor Group : ~218M Elements](#block-7-tensor-group--218m-elements)
|
| 72 |
+
- [Block 8 Tensor Group : ~218M Elements](#block-8-tensor-group--218m-elements)
|
| 73 |
+
- [Block 9 Tensor Group : ~218M Elements](#block-9-tensor-group--218m-elements)
|
| 74 |
+
- [Block 10 Tensor Group : ~218M Elements](#block-10-tensor-group--218m-elements)
|
| 75 |
+
- [Block 11 Tensor Group : ~218M Elements](#block-11-tensor-group--218m-elements)
|
| 76 |
+
- [Block 12 Tensor Group : ~218M Elements](#block-12-tensor-group--218m-elements)
|
| 77 |
+
- [Block 13 Tensor Group : ~218M Elements](#block-13-tensor-group--218m-elements)
|
| 78 |
+
- [Block 14 Tensor Group : ~218M Elements](#block-14-tensor-group--218m-elements)
|
| 79 |
+
- [Block 15 Tensor Group : ~218M Elements](#block-15-tensor-group--218m-elements)
|
| 80 |
+
- [Block 16 Tensor Group : ~218M Elements](#block-16-tensor-group--218m-elements)
|
| 81 |
+
- [Block 17 Tensor Group : ~218M Elements](#block-17-tensor-group--218m-elements)
|
| 82 |
+
- [Block 18 Tensor Group : ~218M Elements](#block-18-tensor-group--218m-elements)
|
| 83 |
+
- [Block 19 Tensor Group : ~218M Elements](#block-19-tensor-group--218m-elements)
|
| 84 |
+
- [Block 20 Tensor Group : ~218M Elements](#block-20-tensor-group--218m-elements)
|
| 85 |
+
- [Block 21 Tensor Group : ~218M Elements](#block-21-tensor-group--218m-elements)
|
| 86 |
+
- [Block 22 Tensor Group : ~218M Elements](#block-22-tensor-group--218m-elements)
|
| 87 |
+
- [Block 23 Tensor Group : ~218M Elements](#block-23-tensor-group--218m-elements)
|
| 88 |
+
- [Block 24 Tensor Group : ~218M Elements](#block-24-tensor-group--218m-elements)
|
| 89 |
+
- [Block 25 Tensor Group : ~218M Elements](#block-25-tensor-group--218m-elements)
|
| 90 |
+
- [Block 26 Tensor Group : ~218M Elements](#block-26-tensor-group--218m-elements)
|
| 91 |
+
- [Block 27 Tensor Group : ~218M Elements](#block-27-tensor-group--218m-elements)
|
| 92 |
+
- [Block 28 Tensor Group : ~218M Elements](#block-28-tensor-group--218m-elements)
|
| 93 |
+
- [Block 29 Tensor Group : ~218M Elements](#block-29-tensor-group--218m-elements)
|
| 94 |
+
- [Block 30 Tensor Group : ~218M Elements](#block-30-tensor-group--218m-elements)
|
| 95 |
+
- [Block 31 Tensor Group : ~218M Elements](#block-31-tensor-group--218m-elements)
|
| 96 |
+
|
| 97 |
+
### Tensor Data Offset
|
| 98 |
+
|
| 99 |
+
This table contains the offset and data segment relative to start of file
|
| 100 |
+
|
| 101 |
+
| T_ID | Tensor Layer Name | Data Offset (B) | Data Size (B) |
|
| 102 |
+
|-----:|:--------------------------|-----------------:|-----------------:|
|
| 103 |
+
| 0 | output.weight | 0x779620 | 0x15870000 |
|
| 104 |
+
| 1 | output_norm.weight | 0x15fe9620 | 0x4000 |
|
| 105 |
+
| 2 | rope_freqs.weight | 0x15fed620 | 0x100 |
|
| 106 |
+
| 3 | token_embd.weight | 0x15fed720 | 0xd746000 |
|
| 107 |
+
| 4 | blk.0.attn_k.weight | 0x23733720 | 0x240000 |
|
| 108 |
+
| 5 | blk.0.attn_norm.weight | 0x23973720 | 0x4000 |
|
| 109 |
+
| 6 | blk.0.attn_output.weight | 0x23977720 | 0xb00000 |
|
| 110 |
+
| 7 | blk.0.attn_q.weight | 0x24477720 | 0x900000 |
|
| 111 |
+
| 8 | blk.0.attn_v.weight | 0x24d77720 | 0x2c0000 |
|
| 112 |
+
| 9 | blk.0.ffn_down.weight | 0x25037720 | 0x2680000 |
|
| 113 |
+
| 10 | blk.0.ffn_gate.weight | 0x276b7720 | 0x1f80000 |
|
| 114 |
+
| 11 | blk.0.ffn_norm.weight | 0x29637720 | 0x4000 |
|
| 115 |
+
| 12 | blk.0.ffn_up.weight | 0x2963b720 | 0x1f80000 |
|
| 116 |
+
| 13 | blk.1.attn_k.weight | 0x2b5bb720 | 0x240000 |
|
| 117 |
+
| 14 | blk.1.attn_norm.weight | 0x2b7fb720 | 0x4000 |
|
| 118 |
+
| 15 | blk.1.attn_output.weight | 0x2b7ff720 | 0xb00000 |
|
| 119 |
+
| 16 | blk.1.attn_q.weight | 0x2c2ff720 | 0x900000 |
|
| 120 |
+
| 17 | blk.1.attn_v.weight | 0x2cbff720 | 0x2c0000 |
|
| 121 |
+
| 18 | blk.1.ffn_down.weight | 0x2cebf720 | 0x2df0000 |
|
| 122 |
+
| 19 | blk.1.ffn_gate.weight | 0x2fcaf720 | 0x1f80000 |
|
| 123 |
+
| 20 | blk.1.ffn_norm.weight | 0x31c2f720 | 0x4000 |
|
| 124 |
+
| 21 | blk.1.ffn_up.weight | 0x31c33720 | 0x1f80000 |
|
| 125 |
+
| 22 | blk.2.attn_k.weight | 0x33bb3720 | 0x240000 |
|
| 126 |
+
| 23 | blk.2.attn_norm.weight | 0x33df3720 | 0x4000 |
|
| 127 |
+
| 24 | blk.2.attn_output.weight | 0x33df7720 | 0xb00000 |
|
| 128 |
+
| 25 | blk.2.attn_q.weight | 0x348f7720 | 0x900000 |
|
| 129 |
+
| 26 | blk.2.attn_v.weight | 0x351f7720 | 0x2c0000 |
|
| 130 |
+
| 27 | blk.2.ffn_down.weight | 0x354b7720 | 0x2680000 |
|
| 131 |
+
| 28 | blk.2.ffn_gate.weight | 0x37b37720 | 0x1f80000 |
|
| 132 |
+
| 29 | blk.2.ffn_norm.weight | 0x39ab7720 | 0x4000 |
|
| 133 |
+
| 30 | blk.2.ffn_up.weight | 0x39abb720 | 0x1f80000 |
|
| 134 |
+
| 31 | blk.3.attn_k.weight | 0x3ba3b720 | 0x240000 |
|
| 135 |
+
| 32 | blk.3.attn_norm.weight | 0x3bc7b720 | 0x4000 |
|
| 136 |
+
| 33 | blk.3.attn_output.weight | 0x3bc7f720 | 0xb00000 |
|
| 137 |
+
| 34 | blk.3.attn_q.weight | 0x3c77f720 | 0x900000 |
|
| 138 |
+
| 35 | blk.3.attn_v.weight | 0x3d07f720 | 0x2c0000 |
|
| 139 |
+
| 36 | blk.3.ffn_down.weight | 0x3d33f720 | 0x2680000 |
|
| 140 |
+
| 37 | blk.3.ffn_gate.weight | 0x3f9bf720 | 0x1f80000 |
|
| 141 |
+
| 38 | blk.3.ffn_norm.weight | 0x4193f720 | 0x4000 |
|
| 142 |
+
| 39 | blk.3.ffn_up.weight | 0x41943720 | 0x1f80000 |
|
| 143 |
+
| 40 | blk.4.attn_k.weight | 0x438c3720 | 0x240000 |
|
| 144 |
+
| 41 | blk.4.attn_norm.weight | 0x43b03720 | 0x4000 |
|
| 145 |
+
| 42 | blk.4.attn_output.weight | 0x43b07720 | 0xb00000 |
|
| 146 |
+
| 43 | blk.4.attn_q.weight | 0x44607720 | 0x900000 |
|
| 147 |
+
| 44 | blk.4.attn_v.weight | 0x44f07720 | 0x2c0000 |
|
| 148 |
+
| 45 | blk.4.ffn_down.weight | 0x451c7720 | 0x2680000 |
|
| 149 |
+
| 46 | blk.4.ffn_gate.weight | 0x47847720 | 0x1f80000 |
|
| 150 |
+
| 47 | blk.4.ffn_norm.weight | 0x497c7720 | 0x4000 |
|
| 151 |
+
| 48 | blk.4.ffn_up.weight | 0x497cb720 | 0x1f80000 |
|
| 152 |
+
| 49 | blk.5.attn_k.weight | 0x4b74b720 | 0x240000 |
|
| 153 |
+
| 50 | blk.5.attn_norm.weight | 0x4b98b720 | 0x4000 |
|
| 154 |
+
| 51 | blk.5.attn_output.weight | 0x4b98f720 | 0xb00000 |
|
| 155 |
+
| 52 | blk.5.attn_q.weight | 0x4c48f720 | 0x900000 |
|
| 156 |
+
| 53 | blk.5.attn_v.weight | 0x4cd8f720 | 0x2c0000 |
|
| 157 |
+
| 54 | blk.5.ffn_down.weight | 0x4d04f720 | 0x2680000 |
|
| 158 |
+
| 55 | blk.5.ffn_gate.weight | 0x4f6cf720 | 0x1f80000 |
|
| 159 |
+
| 56 | blk.5.ffn_norm.weight | 0x5164f720 | 0x4000 |
|
| 160 |
+
| 57 | blk.5.ffn_up.weight | 0x51653720 | 0x1f80000 |
|
| 161 |
+
| 58 | blk.6.attn_k.weight | 0x535d3720 | 0x240000 |
|
| 162 |
+
| 59 | blk.6.attn_norm.weight | 0x53813720 | 0x4000 |
|
| 163 |
+
| 60 | blk.6.attn_output.weight | 0x53817720 | 0xb00000 |
|
| 164 |
+
| 61 | blk.6.attn_q.weight | 0x54317720 | 0x900000 |
|
| 165 |
+
| 62 | blk.6.attn_v.weight | 0x54c17720 | 0x2c0000 |
|
| 166 |
+
| 63 | blk.6.ffn_down.weight | 0x54ed7720 | 0x2680000 |
|
| 167 |
+
| 64 | blk.6.ffn_gate.weight | 0x57557720 | 0x1f80000 |
|
| 168 |
+
| 65 | blk.6.ffn_norm.weight | 0x594d7720 | 0x4000 |
|
| 169 |
+
| 66 | blk.6.ffn_up.weight | 0x594db720 | 0x1f80000 |
|
| 170 |
+
| 67 | blk.7.attn_k.weight | 0x5b45b720 | 0x240000 |
|
| 171 |
+
| 68 | blk.7.attn_norm.weight | 0x5b69b720 | 0x4000 |
|
| 172 |
+
| 69 | blk.7.attn_output.weight | 0x5b69f720 | 0xb00000 |
|
| 173 |
+
| 70 | blk.7.attn_q.weight | 0x5c19f720 | 0x900000 |
|
| 174 |
+
| 71 | blk.7.attn_v.weight | 0x5ca9f720 | 0x2c0000 |
|
| 175 |
+
| 72 | blk.7.ffn_down.weight | 0x5cd5f720 | 0x2680000 |
|
| 176 |
+
| 73 | blk.7.ffn_gate.weight | 0x5f3df720 | 0x1f80000 |
|
| 177 |
+
| 74 | blk.7.ffn_norm.weight | 0x6135f720 | 0x4000 |
|
| 178 |
+
| 75 | blk.7.ffn_up.weight | 0x61363720 | 0x1f80000 |
|
| 179 |
+
| 76 | blk.8.attn_k.weight | 0x632e3720 | 0x240000 |
|
| 180 |
+
| 77 | blk.8.attn_norm.weight | 0x63523720 | 0x4000 |
|
| 181 |
+
| 78 | blk.8.attn_output.weight | 0x63527720 | 0xb00000 |
|
| 182 |
+
| 79 | blk.8.attn_q.weight | 0x64027720 | 0x900000 |
|
| 183 |
+
| 80 | blk.8.attn_v.weight | 0x64927720 | 0x2c0000 |
|
| 184 |
+
| 81 | blk.8.ffn_down.weight | 0x64be7720 | 0x2680000 |
|
| 185 |
+
| 82 | blk.8.ffn_gate.weight | 0x67267720 | 0x1f80000 |
|
| 186 |
+
| 83 | blk.8.ffn_norm.weight | 0x691e7720 | 0x4000 |
|
| 187 |
+
| 84 | blk.8.ffn_up.weight | 0x691eb720 | 0x1f80000 |
|
| 188 |
+
| 85 | blk.9.attn_k.weight | 0x6b16b720 | 0x240000 |
|
| 189 |
+
| 86 | blk.9.attn_norm.weight | 0x6b3ab720 | 0x4000 |
|
| 190 |
+
| 87 | blk.9.attn_output.weight | 0x6b3af720 | 0xb00000 |
|
| 191 |
+
| 88 | blk.9.attn_q.weight | 0x6beaf720 | 0x900000 |
|
| 192 |
+
| 89 | blk.9.attn_v.weight | 0x6c7af720 | 0x2c0000 |
|
| 193 |
+
| 90 | blk.9.ffn_down.weight | 0x6ca6f720 | 0x2680000 |
|
| 194 |
+
| 91 | blk.9.ffn_gate.weight | 0x6f0ef720 | 0x1f80000 |
|
| 195 |
+
| 92 | blk.9.ffn_norm.weight | 0x7106f720 | 0x4000 |
|
| 196 |
+
| 93 | blk.9.ffn_up.weight | 0x71073720 | 0x1f80000 |
|
| 197 |
+
| 94 | blk.10.attn_k.weight | 0x72ff3720 | 0x240000 |
|
| 198 |
+
| 95 | blk.10.attn_norm.weight | 0x73233720 | 0x4000 |
|
| 199 |
+
| 96 | blk.10.attn_output.weight | 0x73237720 | 0xb00000 |
|
| 200 |
+
| 97 | blk.10.attn_q.weight | 0x73d37720 | 0x900000 |
|
| 201 |
+
| 98 | blk.10.attn_v.weight | 0x74637720 | 0x2c0000 |
|
| 202 |
+
| 99 | blk.10.ffn_down.weight | 0x748f7720 | 0x2680000 |
|
| 203 |
+
| 100 | blk.10.ffn_gate.weight | 0x76f77720 | 0x1f80000 |
|
| 204 |
+
| 101 | blk.10.ffn_norm.weight | 0x78ef7720 | 0x4000 |
|
| 205 |
+
| 102 | blk.10.ffn_up.weight | 0x78efb720 | 0x1f80000 |
|
| 206 |
+
| 103 | blk.11.attn_k.weight | 0x7ae7b720 | 0x240000 |
|
| 207 |
+
| 104 | blk.11.attn_norm.weight | 0x7b0bb720 | 0x4000 |
|
| 208 |
+
| 105 | blk.11.attn_output.weight | 0x7b0bf720 | 0xb00000 |
|
| 209 |
+
| 106 | blk.11.attn_q.weight | 0x7bbbf720 | 0x900000 |
|
| 210 |
+
| 107 | blk.11.attn_v.weight | 0x7c4bf720 | 0x2c0000 |
|
| 211 |
+
| 108 | blk.11.ffn_down.weight | 0x7c77f720 | 0x2680000 |
|
| 212 |
+
| 109 | blk.11.ffn_gate.weight | 0x7edff720 | 0x1f80000 |
|
| 213 |
+
| 110 | blk.11.ffn_norm.weight | 0x80d7f720 | 0x4000 |
|
| 214 |
+
| 111 | blk.11.ffn_up.weight | 0x80d83720 | 0x1f80000 |
|
| 215 |
+
| 112 | blk.12.attn_k.weight | 0x82d03720 | 0x240000 |
|
| 216 |
+
| 113 | blk.12.attn_norm.weight | 0x82f43720 | 0x4000 |
|
| 217 |
+
| 114 | blk.12.attn_output.weight | 0x82f47720 | 0xb00000 |
|
| 218 |
+
| 115 | blk.12.attn_q.weight | 0x83a47720 | 0x900000 |
|
| 219 |
+
| 116 | blk.12.attn_v.weight | 0x84347720 | 0x2c0000 |
|
| 220 |
+
| 117 | blk.12.ffn_down.weight | 0x84607720 | 0x2680000 |
|
| 221 |
+
| 118 | blk.12.ffn_gate.weight | 0x86c87720 | 0x1f80000 |
|
| 222 |
+
| 119 | blk.12.ffn_norm.weight | 0x88c07720 | 0x4000 |
|
| 223 |
+
| 120 | blk.12.ffn_up.weight | 0x88c0b720 | 0x1f80000 |
|
| 224 |
+
| 121 | blk.13.attn_k.weight | 0x8ab8b720 | 0x2c0000 |
|
| 225 |
+
| 122 | blk.13.attn_norm.weight | 0x8ae4b720 | 0x4000 |
|
| 226 |
+
| 123 | blk.13.attn_output.weight | 0x8ae4f720 | 0xb00000 |
|
| 227 |
+
| 124 | blk.13.attn_q.weight | 0x8b94f720 | 0xb00000 |
|
| 228 |
+
| 125 | blk.13.attn_v.weight | 0x8c44f720 | 0x2c0000 |
|
| 229 |
+
| 126 | blk.13.ffn_down.weight | 0x8c70f720 | 0x2680000 |
|
| 230 |
+
| 127 | blk.13.ffn_gate.weight | 0x8ed8f720 | 0x1f80000 |
|
| 231 |
+
| 128 | blk.13.ffn_norm.weight | 0x90d0f720 | 0x4000 |
|
| 232 |
+
| 129 | blk.13.ffn_up.weight | 0x90d13720 | 0x1f80000 |
|
| 233 |
+
| 130 | blk.14.attn_k.weight | 0x92c93720 | 0x2c0000 |
|
| 234 |
+
| 131 | blk.14.attn_norm.weight | 0x92f53720 | 0x4000 |
|
| 235 |
+
| 132 | blk.14.attn_output.weight | 0x92f57720 | 0xb00000 |
|
| 236 |
+
| 133 | blk.14.attn_q.weight | 0x93a57720 | 0xb00000 |
|
| 237 |
+
| 134 | blk.14.attn_v.weight | 0x94557720 | 0x2c0000 |
|
| 238 |
+
| 135 | blk.14.ffn_down.weight | 0x94817720 | 0x2680000 |
|
| 239 |
+
| 136 | blk.14.ffn_gate.weight | 0x96e97720 | 0x1f80000 |
|
| 240 |
+
| 137 | blk.14.ffn_norm.weight | 0x98e17720 | 0x4000 |
|
| 241 |
+
| 138 | blk.14.ffn_up.weight | 0x98e1b720 | 0x1f80000 |
|
| 242 |
+
| 139 | blk.15.attn_k.weight | 0x9ad9b720 | 0x240000 |
|
| 243 |
+
| 140 | blk.15.attn_norm.weight | 0x9afdb720 | 0x4000 |
|
| 244 |
+
| 141 | blk.15.attn_output.weight | 0x9afdf720 | 0xb00000 |
|
| 245 |
+
| 142 | blk.15.attn_q.weight | 0x9badf720 | 0x900000 |
|
| 246 |
+
| 143 | blk.15.attn_v.weight | 0x9c3df720 | 0x2c0000 |
|
| 247 |
+
| 144 | blk.15.ffn_down.weight | 0x9c69f720 | 0x2680000 |
|
| 248 |
+
| 145 | blk.15.ffn_gate.weight | 0x9ed1f720 | 0x1f80000 |
|
| 249 |
+
| 146 | blk.15.ffn_norm.weight | 0xa0c9f720 | 0x4000 |
|
| 250 |
+
| 147 | blk.15.ffn_up.weight | 0xa0ca3720 | 0x1f80000 |
|
| 251 |
+
| 148 | blk.16.attn_k.weight | 0xa2c23720 | 0x2c0000 |
|
| 252 |
+
| 149 | blk.16.attn_norm.weight | 0xa2ee3720 | 0x4000 |
|
| 253 |
+
| 150 | blk.16.attn_output.weight | 0xa2ee7720 | 0xb00000 |
|
| 254 |
+
| 151 | blk.16.attn_q.weight | 0xa39e7720 | 0xb00000 |
|
| 255 |
+
| 152 | blk.16.attn_v.weight | 0xa44e7720 | 0x2c0000 |
|
| 256 |
+
| 153 | blk.16.ffn_down.weight | 0xa47a7720 | 0x2680000 |
|
| 257 |
+
| 154 | blk.16.ffn_gate.weight | 0xa6e27720 | 0x2680000 |
|
| 258 |
+
| 155 | blk.16.ffn_norm.weight | 0xa94a7720 | 0x4000 |
|
| 259 |
+
| 156 | blk.16.ffn_up.weight | 0xa94ab720 | 0x2680000 |
|
| 260 |
+
| 157 | blk.17.attn_k.weight | 0xabb2b720 | 0x240000 |
|
| 261 |
+
| 158 | blk.17.attn_norm.weight | 0xabd6b720 | 0x4000 |
|
| 262 |
+
| 159 | blk.17.attn_output.weight | 0xabd6f720 | 0xb00000 |
|
| 263 |
+
| 160 | blk.17.attn_q.weight | 0xac86f720 | 0x900000 |
|
| 264 |
+
| 161 | blk.17.attn_v.weight | 0xad16f720 | 0x2c0000 |
|
| 265 |
+
| 162 | blk.17.ffn_down.weight | 0xad42f720 | 0x2df0000 |
|
| 266 |
+
| 163 | blk.17.ffn_gate.weight | 0xb021f720 | 0x2680000 |
|
| 267 |
+
| 164 | blk.17.ffn_norm.weight | 0xb289f720 | 0x4000 |
|
| 268 |
+
| 165 | blk.17.ffn_up.weight | 0xb28a3720 | 0x2680000 |
|
| 269 |
+
| 166 | blk.18.attn_k.weight | 0xb4f23720 | 0x2c0000 |
|
| 270 |
+
| 167 | blk.18.attn_norm.weight | 0xb51e3720 | 0x4000 |
|
| 271 |
+
| 168 | blk.18.attn_output.weight | 0xb51e7720 | 0xb00000 |
|
| 272 |
+
| 169 | blk.18.attn_q.weight | 0xb5ce7720 | 0xb00000 |
|
| 273 |
+
| 170 | blk.18.attn_v.weight | 0xb67e7720 | 0x2c0000 |
|
| 274 |
+
| 171 | blk.18.ffn_down.weight | 0xb6aa7720 | 0x2df0000 |
|
| 275 |
+
| 172 | blk.18.ffn_gate.weight | 0xb9897720 | 0x2680000 |
|
| 276 |
+
| 173 | blk.18.ffn_norm.weight | 0xbbf17720 | 0x4000 |
|
| 277 |
+
| 174 | blk.18.ffn_up.weight | 0xbbf1b720 | 0x2680000 |
|
| 278 |
+
| 175 | blk.19.attn_k.weight | 0xbe59b720 | 0x2c0000 |
|
| 279 |
+
| 176 | blk.19.attn_norm.weight | 0xbe85b720 | 0x4000 |
|
| 280 |
+
| 177 | blk.19.attn_output.weight | 0xbe85f720 | 0xb00000 |
|
| 281 |
+
| 178 | blk.19.attn_q.weight | 0xbf35f720 | 0xb00000 |
|
| 282 |
+
| 179 | blk.19.attn_v.weight | 0xbfe5f720 | 0x2c0000 |
|
| 283 |
+
| 180 | blk.19.ffn_down.weight | 0xc011f720 | 0x2df0000 |
|
| 284 |
+
| 181 | blk.19.ffn_gate.weight | 0xc2f0f720 | 0x2680000 |
|
| 285 |
+
| 182 | blk.19.ffn_norm.weight | 0xc558f720 | 0x4000 |
|
| 286 |
+
| 183 | blk.19.ffn_up.weight | 0xc5593720 | 0x2680000 |
|
| 287 |
+
| 184 | blk.20.attn_k.weight | 0xc7c13720 | 0x2c0000 |
|
| 288 |
+
| 185 | blk.20.attn_norm.weight | 0xc7ed3720 | 0x4000 |
|
| 289 |
+
| 186 | blk.20.attn_output.weight | 0xc7ed7720 | 0xb00000 |
|
| 290 |
+
| 187 | blk.20.attn_q.weight | 0xc89d7720 | 0xb00000 |
|
| 291 |
+
| 188 | blk.20.attn_v.weight | 0xc94d7720 | 0x2c0000 |
|
| 292 |
+
| 189 | blk.20.ffn_down.weight | 0xc9797720 | 0x2df0000 |
|
| 293 |
+
| 190 | blk.20.ffn_gate.weight | 0xcc587720 | 0x2680000 |
|
| 294 |
+
| 191 | blk.20.ffn_norm.weight | 0xcec07720 | 0x4000 |
|
| 295 |
+
| 192 | blk.20.ffn_up.weight | 0xcec0b720 | 0x2680000 |
|
| 296 |
+
| 193 | blk.21.attn_k.weight | 0xd128b720 | 0x2c0000 |
|
| 297 |
+
| 194 | blk.21.attn_norm.weight | 0xd154b720 | 0x4000 |
|
| 298 |
+
| 195 | blk.21.attn_output.weight | 0xd154f720 | 0xb00000 |
|
| 299 |
+
| 196 | blk.21.attn_q.weight | 0xd204f720 | 0xb00000 |
|
| 300 |
+
| 197 | blk.21.attn_v.weight | 0xd2b4f720 | 0x2c0000 |
|
| 301 |
+
| 198 | blk.21.ffn_down.weight | 0xd2e0f720 | 0x2df0000 |
|
| 302 |
+
| 199 | blk.21.ffn_gate.weight | 0xd5bff720 | 0x2680000 |
|
| 303 |
+
| 200 | blk.21.ffn_norm.weight | 0xd827f720 | 0x4000 |
|
| 304 |
+
| 201 | blk.21.ffn_up.weight | 0xd8283720 | 0x2680000 |
|
| 305 |
+
| 202 | blk.22.attn_k.weight | 0xda903720 | 0x2c0000 |
|
| 306 |
+
| 203 | blk.22.attn_norm.weight | 0xdabc3720 | 0x4000 |
|
| 307 |
+
| 204 | blk.22.attn_output.weight | 0xdabc7720 | 0xb00000 |
|
| 308 |
+
| 205 | blk.22.attn_q.weight | 0xdb6c7720 | 0xb00000 |
|
| 309 |
+
| 206 | blk.22.attn_v.weight | 0xdc1c7720 | 0x2c0000 |
|
| 310 |
+
| 207 | blk.22.ffn_down.weight | 0xdc487720 | 0x2df0000 |
|
| 311 |
+
| 208 | blk.22.ffn_gate.weight | 0xdf277720 | 0x2680000 |
|
| 312 |
+
| 209 | blk.22.ffn_norm.weight | 0xe18f7720 | 0x4000 |
|
| 313 |
+
| 210 | blk.22.ffn_up.weight | 0xe18fb720 | 0x2680000 |
|
| 314 |
+
| 211 | blk.23.attn_k.weight | 0xe3f7b720 | 0x2c0000 |
|
| 315 |
+
| 212 | blk.23.attn_norm.weight | 0xe423b720 | 0x4000 |
|
| 316 |
+
| 213 | blk.23.attn_output.weight | 0xe423f720 | 0xb00000 |
|
| 317 |
+
| 214 | blk.23.attn_q.weight | 0xe4d3f720 | 0xb00000 |
|
| 318 |
+
| 215 | blk.23.attn_v.weight | 0xe583f720 | 0x2c0000 |
|
| 319 |
+
| 216 | blk.23.ffn_down.weight | 0xe5aff720 | 0x2df0000 |
|
| 320 |
+
| 217 | blk.23.ffn_gate.weight | 0xe88ef720 | 0x2680000 |
|
| 321 |
+
| 218 | blk.23.ffn_norm.weight | 0xeaf6f720 | 0x4000 |
|
| 322 |
+
| 219 | blk.23.ffn_up.weight | 0xeaf73720 | 0x2680000 |
|
| 323 |
+
| 220 | blk.24.attn_k.weight | 0xed5f3720 | 0x2c0000 |
|
| 324 |
+
| 221 | blk.24.attn_norm.weight | 0xed8b3720 | 0x4000 |
|
| 325 |
+
| 222 | blk.24.attn_output.weight | 0xed8b7720 | 0xb00000 |
|
| 326 |
+
| 223 | blk.24.attn_q.weight | 0xee3b7720 | 0xb00000 |
|
| 327 |
+
| 224 | blk.24.attn_v.weight | 0xeeeb7720 | 0x2c0000 |
|
| 328 |
+
| 225 | blk.24.ffn_down.weight | 0xef177720 | 0x2df0000 |
|
| 329 |
+
| 226 | blk.24.ffn_gate.weight | 0xf1f67720 | 0x2680000 |
|
| 330 |
+
| 227 | blk.24.ffn_norm.weight | 0xf45e7720 | 0x4000 |
|
| 331 |
+
| 228 | blk.24.ffn_up.weight | 0xf45eb720 | 0x2680000 |
|
| 332 |
+
| 229 | blk.25.attn_k.weight | 0xf6c6b720 | 0x2c0000 |
|
| 333 |
+
| 230 | blk.25.attn_norm.weight | 0xf6f2b720 | 0x4000 |
|
| 334 |
+
| 231 | blk.25.attn_output.weight | 0xf6f2f720 | 0xb00000 |
|
| 335 |
+
| 232 | blk.25.attn_q.weight | 0xf7a2f720 | 0xb00000 |
|
| 336 |
+
| 233 | blk.25.attn_v.weight | 0xf852f720 | 0x2c0000 |
|
| 337 |
+
| 234 | blk.25.ffn_down.weight | 0xf87ef720 | 0x2df0000 |
|
| 338 |
+
| 235 | blk.25.ffn_gate.weight | 0xfb5df720 | 0x2680000 |
|
| 339 |
+
| 236 | blk.25.ffn_norm.weight | 0xfdc5f720 | 0x4000 |
|
| 340 |
+
| 237 | blk.25.ffn_up.weight | 0xfdc63720 | 0x2680000 |
|
| 341 |
+
| 238 | blk.26.attn_k.weight | 0x1002e3720 | 0x2c0000 |
|
| 342 |
+
| 239 | blk.26.attn_norm.weight | 0x1005a3720 | 0x4000 |
|
| 343 |
+
| 240 | blk.26.attn_output.weight | 0x1005a7720 | 0xb00000 |
|
| 344 |
+
| 241 | blk.26.attn_q.weight | 0x1010a7720 | 0xb00000 |
|
| 345 |
+
| 242 | blk.26.attn_v.weight | 0x101ba7720 | 0x2c0000 |
|
| 346 |
+
| 243 | blk.26.ffn_down.weight | 0x101e67720 | 0x2df0000 |
|
| 347 |
+
| 244 | blk.26.ffn_gate.weight | 0x104c57720 | 0x2680000 |
|
| 348 |
+
| 245 | blk.26.ffn_norm.weight | 0x1072d7720 | 0x4000 |
|
| 349 |
+
| 246 | blk.26.ffn_up.weight | 0x1072db720 | 0x2680000 |
|
| 350 |
+
| 247 | blk.27.attn_k.weight | 0x10995b720 | 0x2c0000 |
|
| 351 |
+
| 248 | blk.27.attn_norm.weight | 0x109c1b720 | 0x4000 |
|
| 352 |
+
| 249 | blk.27.attn_output.weight | 0x109c1f720 | 0xb00000 |
|
| 353 |
+
| 250 | blk.27.attn_q.weight | 0x10a71f720 | 0xb00000 |
|
| 354 |
+
| 251 | blk.27.attn_v.weight | 0x10b21f720 | 0x2c0000 |
|
| 355 |
+
| 252 | blk.27.ffn_down.weight | 0x10b4df720 | 0x2df0000 |
|
| 356 |
+
| 253 | blk.27.ffn_gate.weight | 0x10e2cf720 | 0x2680000 |
|
| 357 |
+
| 254 | blk.27.ffn_norm.weight | 0x11094f720 | 0x4000 |
|
| 358 |
+
| 255 | blk.27.ffn_up.weight | 0x110953720 | 0x2680000 |
|
| 359 |
+
| 256 | blk.28.attn_k.weight | 0x112fd3720 | 0x2c0000 |
|
| 360 |
+
| 257 | blk.28.attn_norm.weight | 0x113293720 | 0x4000 |
|
| 361 |
+
| 258 | blk.28.attn_output.weight | 0x113297720 | 0xb00000 |
|
| 362 |
+
| 259 | blk.28.attn_q.weight | 0x113d97720 | 0xb00000 |
|
| 363 |
+
| 260 | blk.28.attn_v.weight | 0x114897720 | 0x2c0000 |
|
| 364 |
+
| 261 | blk.28.ffn_down.weight | 0x114b57720 | 0x2df0000 |
|
| 365 |
+
| 262 | blk.28.ffn_gate.weight | 0x117947720 | 0x2680000 |
|
| 366 |
+
| 263 | blk.28.ffn_norm.weight | 0x119fc7720 | 0x4000 |
|
| 367 |
+
| 264 | blk.28.ffn_up.weight | 0x119fcb720 | 0x2680000 |
|
| 368 |
+
| 265 | blk.29.attn_k.weight | 0x11c64b720 | 0x2c0000 |
|
| 369 |
+
| 266 | blk.29.attn_norm.weight | 0x11c90b720 | 0x4000 |
|
| 370 |
+
| 267 | blk.29.attn_output.weight | 0x11c90f720 | 0xb00000 |
|
| 371 |
+
| 268 | blk.29.attn_q.weight | 0x11d40f720 | 0xb00000 |
|
| 372 |
+
| 269 | blk.29.attn_v.weight | 0x11df0f720 | 0x2c0000 |
|
| 373 |
+
| 270 | blk.29.ffn_down.weight | 0x11e1cf720 | 0x2df0000 |
|
| 374 |
+
| 271 | blk.29.ffn_gate.weight | 0x120fbf720 | 0x2680000 |
|
| 375 |
+
| 272 | blk.29.ffn_norm.weight | 0x12363f720 | 0x4000 |
|
| 376 |
+
| 273 | blk.29.ffn_up.weight | 0x123643720 | 0x2680000 |
|
| 377 |
+
| 274 | blk.30.attn_k.weight | 0x125cc3720 | 0x2c0000 |
|
| 378 |
+
| 275 | blk.30.attn_norm.weight | 0x125f83720 | 0x4000 |
|
| 379 |
+
| 276 | blk.30.attn_output.weight | 0x125f87720 | 0xb00000 |
|
| 380 |
+
| 277 | blk.30.attn_q.weight | 0x126a87720 | 0xb00000 |
|
| 381 |
+
| 278 | blk.30.attn_v.weight | 0x127587720 | 0x2c0000 |
|
| 382 |
+
| 279 | blk.30.ffn_down.weight | 0x127847720 | 0x2df0000 |
|
| 383 |
+
| 280 | blk.30.ffn_gate.weight | 0x12a637720 | 0x2680000 |
|
| 384 |
+
| 281 | blk.30.ffn_norm.weight | 0x12ccb7720 | 0x4000 |
|
| 385 |
+
| 282 | blk.30.ffn_up.weight | 0x12ccbb720 | 0x2680000 |
|
| 386 |
+
| 283 | blk.31.attn_k.weight | 0x12f33b720 | 0x240000 |
|
| 387 |
+
| 284 | blk.31.attn_norm.weight | 0x12f57b720 | 0x4000 |
|
| 388 |
+
| 285 | blk.31.attn_output.weight | 0x12f57f720 | 0xb00000 |
|
| 389 |
+
| 286 | blk.31.attn_q.weight | 0x13007f720 | 0x900000 |
|
| 390 |
+
| 287 | blk.31.attn_v.weight | 0x13097f720 | 0x2c0000 |
|
| 391 |
+
| 288 | blk.31.ffn_down.weight | 0x130c3f720 | 0x2df0000 |
|
| 392 |
+
| 289 | blk.31.ffn_gate.weight | 0x133a2f720 | 0x2680000 |
|
| 393 |
+
| 290 | blk.31.ffn_norm.weight | 0x1360af720 | 0x4000 |
|
| 394 |
+
| 291 | blk.31.ffn_up.weight | 0x1360b3720 | 0x2680000 |
|
| 395 |
+
|
| 396 |
+
### <a name="base">Base Tensor Group : ~1B Elements</a>
|
| 397 |
+
|
| 398 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 399 |
+
|-----:|:-------------------|:---------------------------------|:------------------|:----------------------|:-----|
|
| 400 |
+
| 0 | output.weight | Output (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | Q5_K |
|
| 401 |
+
| 1 | output_norm.weight | Output Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 402 |
+
| 2 | rope_freqs.weight | Rope_Freqs (W) | ( 64) 64 | 64 x 1 x 1 x 1 | F32 |
|
| 403 |
+
| 3 | token_embd.weight | Token Embedding (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | Q3_K |
|
| 404 |
+
|
| 405 |
+
- Total elements in base: ( ~1B) 1050677312
|
| 406 |
+
- Percentage of total elements: 13.08%
|
| 407 |
+
|
| 408 |
+
|
| 409 |
+
### <a name="blk_0">Block 0 Tensor Group : ~218M Elements</a>
|
| 410 |
+
|
| 411 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 412 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 413 |
+
| 4 | blk.0.attn_k.weight | Block 0 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 414 |
+
| 5 | blk.0.attn_norm.weight | Block 0 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 415 |
+
| 6 | blk.0.attn_output.weight | Block 0 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 416 |
+
| 7 | blk.0.attn_q.weight | Block 0 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 417 |
+
| 8 | blk.0.attn_v.weight | Block 0 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 418 |
+
| 9 | blk.0.ffn_down.weight | Block 0 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 419 |
+
| 10 | blk.0.ffn_gate.weight | Block 0 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 420 |
+
| 11 | blk.0.ffn_norm.weight | Block 0 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 421 |
+
| 12 | blk.0.ffn_up.weight | Block 0 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 422 |
+
|
| 423 |
+
- Total elements in blk.0: (~218M) 218112000
|
| 424 |
+
- Percentage of total elements: 2.72%
|
| 425 |
+
|
| 426 |
+
|
| 427 |
+
### <a name="blk_1">Block 1 Tensor Group : ~218M Elements</a>
|
| 428 |
+
|
| 429 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 430 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 431 |
+
| 13 | blk.1.attn_k.weight | Block 1 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 432 |
+
| 14 | blk.1.attn_norm.weight | Block 1 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 433 |
+
| 15 | blk.1.attn_output.weight | Block 1 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 434 |
+
| 16 | blk.1.attn_q.weight | Block 1 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 435 |
+
| 17 | blk.1.attn_v.weight | Block 1 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 436 |
+
| 18 | blk.1.ffn_down.weight | Block 1 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 437 |
+
| 19 | blk.1.ffn_gate.weight | Block 1 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 438 |
+
| 20 | blk.1.ffn_norm.weight | Block 1 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 439 |
+
| 21 | blk.1.ffn_up.weight | Block 1 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 440 |
+
|
| 441 |
+
- Total elements in blk.1: (~218M) 218112000
|
| 442 |
+
- Percentage of total elements: 2.72%
|
| 443 |
+
|
| 444 |
+
|
| 445 |
+
### <a name="blk_2">Block 2 Tensor Group : ~218M Elements</a>
|
| 446 |
+
|
| 447 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 448 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 449 |
+
| 22 | blk.2.attn_k.weight | Block 2 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 450 |
+
| 23 | blk.2.attn_norm.weight | Block 2 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 451 |
+
| 24 | blk.2.attn_output.weight | Block 2 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 452 |
+
| 25 | blk.2.attn_q.weight | Block 2 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 453 |
+
| 26 | blk.2.attn_v.weight | Block 2 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 454 |
+
| 27 | blk.2.ffn_down.weight | Block 2 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 455 |
+
| 28 | blk.2.ffn_gate.weight | Block 2 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 456 |
+
| 29 | blk.2.ffn_norm.weight | Block 2 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 457 |
+
| 30 | blk.2.ffn_up.weight | Block 2 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 458 |
+
|
| 459 |
+
- Total elements in blk.2: (~218M) 218112000
|
| 460 |
+
- Percentage of total elements: 2.72%
|
| 461 |
+
|
| 462 |
+
|
| 463 |
+
### <a name="blk_3">Block 3 Tensor Group : ~218M Elements</a>
|
| 464 |
+
|
| 465 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 466 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 467 |
+
| 31 | blk.3.attn_k.weight | Block 3 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 468 |
+
| 32 | blk.3.attn_norm.weight | Block 3 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 469 |
+
| 33 | blk.3.attn_output.weight | Block 3 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 470 |
+
| 34 | blk.3.attn_q.weight | Block 3 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 471 |
+
| 35 | blk.3.attn_v.weight | Block 3 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 472 |
+
| 36 | blk.3.ffn_down.weight | Block 3 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 473 |
+
| 37 | blk.3.ffn_gate.weight | Block 3 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 474 |
+
| 38 | blk.3.ffn_norm.weight | Block 3 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 475 |
+
| 39 | blk.3.ffn_up.weight | Block 3 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 476 |
+
|
| 477 |
+
- Total elements in blk.3: (~218M) 218112000
|
| 478 |
+
- Percentage of total elements: 2.72%
|
| 479 |
+
|
| 480 |
+
|
| 481 |
+
### <a name="blk_4">Block 4 Tensor Group : ~218M Elements</a>
|
| 482 |
+
|
| 483 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 484 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 485 |
+
| 40 | blk.4.attn_k.weight | Block 4 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 486 |
+
| 41 | blk.4.attn_norm.weight | Block 4 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 487 |
+
| 42 | blk.4.attn_output.weight | Block 4 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 488 |
+
| 43 | blk.4.attn_q.weight | Block 4 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 489 |
+
| 44 | blk.4.attn_v.weight | Block 4 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 490 |
+
| 45 | blk.4.ffn_down.weight | Block 4 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 491 |
+
| 46 | blk.4.ffn_gate.weight | Block 4 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 492 |
+
| 47 | blk.4.ffn_norm.weight | Block 4 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 493 |
+
| 48 | blk.4.ffn_up.weight | Block 4 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 494 |
+
|
| 495 |
+
- Total elements in blk.4: (~218M) 218112000
|
| 496 |
+
- Percentage of total elements: 2.72%
|
| 497 |
+
|
| 498 |
+
|
| 499 |
+
### <a name="blk_5">Block 5 Tensor Group : ~218M Elements</a>
|
| 500 |
+
|
| 501 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 502 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 503 |
+
| 49 | blk.5.attn_k.weight | Block 5 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 504 |
+
| 50 | blk.5.attn_norm.weight | Block 5 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 505 |
+
| 51 | blk.5.attn_output.weight | Block 5 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 506 |
+
| 52 | blk.5.attn_q.weight | Block 5 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 507 |
+
| 53 | blk.5.attn_v.weight | Block 5 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 508 |
+
| 54 | blk.5.ffn_down.weight | Block 5 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 509 |
+
| 55 | blk.5.ffn_gate.weight | Block 5 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 510 |
+
| 56 | blk.5.ffn_norm.weight | Block 5 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 511 |
+
| 57 | blk.5.ffn_up.weight | Block 5 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 512 |
+
|
| 513 |
+
- Total elements in blk.5: (~218M) 218112000
|
| 514 |
+
- Percentage of total elements: 2.72%
|
| 515 |
+
|
| 516 |
+
|
| 517 |
+
### <a name="blk_6">Block 6 Tensor Group : ~218M Elements</a>
|
| 518 |
+
|
| 519 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 520 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 521 |
+
| 58 | blk.6.attn_k.weight | Block 6 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 522 |
+
| 59 | blk.6.attn_norm.weight | Block 6 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 523 |
+
| 60 | blk.6.attn_output.weight | Block 6 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 524 |
+
| 61 | blk.6.attn_q.weight | Block 6 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 525 |
+
| 62 | blk.6.attn_v.weight | Block 6 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 526 |
+
| 63 | blk.6.ffn_down.weight | Block 6 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 527 |
+
| 64 | blk.6.ffn_gate.weight | Block 6 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 528 |
+
| 65 | blk.6.ffn_norm.weight | Block 6 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 529 |
+
| 66 | blk.6.ffn_up.weight | Block 6 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 530 |
+
|
| 531 |
+
- Total elements in blk.6: (~218M) 218112000
|
| 532 |
+
- Percentage of total elements: 2.72%
|
| 533 |
+
|
| 534 |
+
|
| 535 |
+
### <a name="blk_7">Block 7 Tensor Group : ~218M Elements</a>
|
| 536 |
+
|
| 537 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 538 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 539 |
+
| 67 | blk.7.attn_k.weight | Block 7 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 540 |
+
| 68 | blk.7.attn_norm.weight | Block 7 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 541 |
+
| 69 | blk.7.attn_output.weight | Block 7 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 542 |
+
| 70 | blk.7.attn_q.weight | Block 7 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 543 |
+
| 71 | blk.7.attn_v.weight | Block 7 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 544 |
+
| 72 | blk.7.ffn_down.weight | Block 7 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 545 |
+
| 73 | blk.7.ffn_gate.weight | Block 7 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 546 |
+
| 74 | blk.7.ffn_norm.weight | Block 7 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 547 |
+
| 75 | blk.7.ffn_up.weight | Block 7 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 548 |
+
|
| 549 |
+
- Total elements in blk.7: (~218M) 218112000
|
| 550 |
+
- Percentage of total elements: 2.72%
|
| 551 |
+
|
| 552 |
+
|
| 553 |
+
### <a name="blk_8">Block 8 Tensor Group : ~218M Elements</a>
|
| 554 |
+
|
| 555 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 556 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 557 |
+
| 76 | blk.8.attn_k.weight | Block 8 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 558 |
+
| 77 | blk.8.attn_norm.weight | Block 8 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 559 |
+
| 78 | blk.8.attn_output.weight | Block 8 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 560 |
+
| 79 | blk.8.attn_q.weight | Block 8 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 561 |
+
| 80 | blk.8.attn_v.weight | Block 8 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 562 |
+
| 81 | blk.8.ffn_down.weight | Block 8 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 563 |
+
| 82 | blk.8.ffn_gate.weight | Block 8 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 564 |
+
| 83 | blk.8.ffn_norm.weight | Block 8 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 565 |
+
| 84 | blk.8.ffn_up.weight | Block 8 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 566 |
+
|
| 567 |
+
- Total elements in blk.8: (~218M) 218112000
|
| 568 |
+
- Percentage of total elements: 2.72%
|
| 569 |
+
|
| 570 |
+
|
| 571 |
+
### <a name="blk_9">Block 9 Tensor Group : ~218M Elements</a>
|
| 572 |
+
|
| 573 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 574 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 575 |
+
| 85 | blk.9.attn_k.weight | Block 9 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 576 |
+
| 86 | blk.9.attn_norm.weight | Block 9 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 577 |
+
| 87 | blk.9.attn_output.weight | Block 9 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 578 |
+
| 88 | blk.9.attn_q.weight | Block 9 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 579 |
+
| 89 | blk.9.attn_v.weight | Block 9 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 580 |
+
| 90 | blk.9.ffn_down.weight | Block 9 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 581 |
+
| 91 | blk.9.ffn_gate.weight | Block 9 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 582 |
+
| 92 | blk.9.ffn_norm.weight | Block 9 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 583 |
+
| 93 | blk.9.ffn_up.weight | Block 9 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 584 |
+
|
| 585 |
+
- Total elements in blk.9: (~218M) 218112000
|
| 586 |
+
- Percentage of total elements: 2.72%
|
| 587 |
+
|
| 588 |
+
|
| 589 |
+
### <a name="blk_10">Block 10 Tensor Group : ~218M Elements</a>
|
| 590 |
+
|
| 591 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 592 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 593 |
+
| 94 | blk.10.attn_k.weight | Block 10 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 594 |
+
| 95 | blk.10.attn_norm.weight | Block 10 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 595 |
+
| 96 | blk.10.attn_output.weight | Block 10 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 596 |
+
| 97 | blk.10.attn_q.weight | Block 10 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 597 |
+
| 98 | blk.10.attn_v.weight | Block 10 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 598 |
+
| 99 | blk.10.ffn_down.weight | Block 10 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 599 |
+
| 100 | blk.10.ffn_gate.weight | Block 10 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 600 |
+
| 101 | blk.10.ffn_norm.weight | Block 10 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 601 |
+
| 102 | blk.10.ffn_up.weight | Block 10 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 602 |
+
|
| 603 |
+
- Total elements in blk.10: (~218M) 218112000
|
| 604 |
+
- Percentage of total elements: 2.72%
|
| 605 |
+
|
| 606 |
+
|
| 607 |
+
### <a name="blk_11">Block 11 Tensor Group : ~218M Elements</a>
|
| 608 |
+
|
| 609 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 610 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 611 |
+
| 103 | blk.11.attn_k.weight | Block 11 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 612 |
+
| 104 | blk.11.attn_norm.weight | Block 11 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 613 |
+
| 105 | blk.11.attn_output.weight | Block 11 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 614 |
+
| 106 | blk.11.attn_q.weight | Block 11 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 615 |
+
| 107 | blk.11.attn_v.weight | Block 11 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 616 |
+
| 108 | blk.11.ffn_down.weight | Block 11 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 617 |
+
| 109 | blk.11.ffn_gate.weight | Block 11 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 618 |
+
| 110 | blk.11.ffn_norm.weight | Block 11 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 619 |
+
| 111 | blk.11.ffn_up.weight | Block 11 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 620 |
+
|
| 621 |
+
- Total elements in blk.11: (~218M) 218112000
|
| 622 |
+
- Percentage of total elements: 2.72%
|
| 623 |
+
|
| 624 |
+
|
| 625 |
+
### <a name="blk_12">Block 12 Tensor Group : ~218M Elements</a>
|
| 626 |
+
|
| 627 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 628 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 629 |
+
| 112 | blk.12.attn_k.weight | Block 12 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 630 |
+
| 113 | blk.12.attn_norm.weight | Block 12 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 631 |
+
| 114 | blk.12.attn_output.weight | Block 12 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 632 |
+
| 115 | blk.12.attn_q.weight | Block 12 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 633 |
+
| 116 | blk.12.attn_v.weight | Block 12 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 634 |
+
| 117 | blk.12.ffn_down.weight | Block 12 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 635 |
+
| 118 | blk.12.ffn_gate.weight | Block 12 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 636 |
+
| 119 | blk.12.ffn_norm.weight | Block 12 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 637 |
+
| 120 | blk.12.ffn_up.weight | Block 12 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 638 |
+
|
| 639 |
+
- Total elements in blk.12: (~218M) 218112000
|
| 640 |
+
- Percentage of total elements: 2.72%
|
| 641 |
+
|
| 642 |
+
|
| 643 |
+
### <a name="blk_13">Block 13 Tensor Group : ~218M Elements</a>
|
| 644 |
+
|
| 645 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 646 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 647 |
+
| 121 | blk.13.attn_k.weight | Block 13 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 648 |
+
| 122 | blk.13.attn_norm.weight | Block 13 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 649 |
+
| 123 | blk.13.attn_output.weight | Block 13 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 650 |
+
| 124 | blk.13.attn_q.weight | Block 13 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 651 |
+
| 125 | blk.13.attn_v.weight | Block 13 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 652 |
+
| 126 | blk.13.ffn_down.weight | Block 13 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 653 |
+
| 127 | blk.13.ffn_gate.weight | Block 13 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 654 |
+
| 128 | blk.13.ffn_norm.weight | Block 13 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 655 |
+
| 129 | blk.13.ffn_up.weight | Block 13 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 656 |
+
|
| 657 |
+
- Total elements in blk.13: (~218M) 218112000
|
| 658 |
+
- Percentage of total elements: 2.72%
|
| 659 |
+
|
| 660 |
+
|
| 661 |
+
### <a name="blk_14">Block 14 Tensor Group : ~218M Elements</a>
|
| 662 |
+
|
| 663 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 664 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 665 |
+
| 130 | blk.14.attn_k.weight | Block 14 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 666 |
+
| 131 | blk.14.attn_norm.weight | Block 14 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 667 |
+
| 132 | blk.14.attn_output.weight | Block 14 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 668 |
+
| 133 | blk.14.attn_q.weight | Block 14 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 669 |
+
| 134 | blk.14.attn_v.weight | Block 14 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 670 |
+
| 135 | blk.14.ffn_down.weight | Block 14 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 671 |
+
| 136 | blk.14.ffn_gate.weight | Block 14 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 672 |
+
| 137 | blk.14.ffn_norm.weight | Block 14 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 673 |
+
| 138 | blk.14.ffn_up.weight | Block 14 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 674 |
+
|
| 675 |
+
- Total elements in blk.14: (~218M) 218112000
|
| 676 |
+
- Percentage of total elements: 2.72%
|
| 677 |
+
|
| 678 |
+
|
| 679 |
+
### <a name="blk_15">Block 15 Tensor Group : ~218M Elements</a>
|
| 680 |
+
|
| 681 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 682 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 683 |
+
| 139 | blk.15.attn_k.weight | Block 15 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 684 |
+
| 140 | blk.15.attn_norm.weight | Block 15 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 685 |
+
| 141 | blk.15.attn_output.weight | Block 15 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 686 |
+
| 142 | blk.15.attn_q.weight | Block 15 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 687 |
+
| 143 | blk.15.attn_v.weight | Block 15 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 688 |
+
| 144 | blk.15.ffn_down.weight | Block 15 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 689 |
+
| 145 | blk.15.ffn_gate.weight | Block 15 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 690 |
+
| 146 | blk.15.ffn_norm.weight | Block 15 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 691 |
+
| 147 | blk.15.ffn_up.weight | Block 15 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q4_K |
|
| 692 |
+
|
| 693 |
+
- Total elements in blk.15: (~218M) 218112000
|
| 694 |
+
- Percentage of total elements: 2.72%
|
| 695 |
+
|
| 696 |
+
|
| 697 |
+
### <a name="blk_16">Block 16 Tensor Group : ~218M Elements</a>
|
| 698 |
+
|
| 699 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 700 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 701 |
+
| 148 | blk.16.attn_k.weight | Block 16 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 702 |
+
| 149 | blk.16.attn_norm.weight | Block 16 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 703 |
+
| 150 | blk.16.attn_output.weight | Block 16 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 704 |
+
| 151 | blk.16.attn_q.weight | Block 16 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 705 |
+
| 152 | blk.16.attn_v.weight | Block 16 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 706 |
+
| 153 | blk.16.ffn_down.weight | Block 16 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q5_K |
|
| 707 |
+
| 154 | blk.16.ffn_gate.weight | Block 16 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 708 |
+
| 155 | blk.16.ffn_norm.weight | Block 16 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 709 |
+
| 156 | blk.16.ffn_up.weight | Block 16 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 710 |
+
|
| 711 |
+
- Total elements in blk.16: (~218M) 218112000
|
| 712 |
+
- Percentage of total elements: 2.72%
|
| 713 |
+
|
| 714 |
+
|
| 715 |
+
### <a name="blk_17">Block 17 Tensor Group : ~218M Elements</a>
|
| 716 |
+
|
| 717 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 718 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 719 |
+
| 157 | blk.17.attn_k.weight | Block 17 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 720 |
+
| 158 | blk.17.attn_norm.weight | Block 17 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 721 |
+
| 159 | blk.17.attn_output.weight | Block 17 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 722 |
+
| 160 | blk.17.attn_q.weight | Block 17 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 723 |
+
| 161 | blk.17.attn_v.weight | Block 17 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 724 |
+
| 162 | blk.17.ffn_down.weight | Block 17 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 725 |
+
| 163 | blk.17.ffn_gate.weight | Block 17 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 726 |
+
| 164 | blk.17.ffn_norm.weight | Block 17 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 727 |
+
| 165 | blk.17.ffn_up.weight | Block 17 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 728 |
+
|
| 729 |
+
- Total elements in blk.17: (~218M) 218112000
|
| 730 |
+
- Percentage of total elements: 2.72%
|
| 731 |
+
|
| 732 |
+
|
| 733 |
+
### <a name="blk_18">Block 18 Tensor Group : ~218M Elements</a>
|
| 734 |
+
|
| 735 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 736 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 737 |
+
| 166 | blk.18.attn_k.weight | Block 18 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 738 |
+
| 167 | blk.18.attn_norm.weight | Block 18 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 739 |
+
| 168 | blk.18.attn_output.weight | Block 18 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 740 |
+
| 169 | blk.18.attn_q.weight | Block 18 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 741 |
+
| 170 | blk.18.attn_v.weight | Block 18 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 742 |
+
| 171 | blk.18.ffn_down.weight | Block 18 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 743 |
+
| 172 | blk.18.ffn_gate.weight | Block 18 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 744 |
+
| 173 | blk.18.ffn_norm.weight | Block 18 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 745 |
+
| 174 | blk.18.ffn_up.weight | Block 18 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 746 |
+
|
| 747 |
+
- Total elements in blk.18: (~218M) 218112000
|
| 748 |
+
- Percentage of total elements: 2.72%
|
| 749 |
+
|
| 750 |
+
|
| 751 |
+
### <a name="blk_19">Block 19 Tensor Group : ~218M Elements</a>
|
| 752 |
+
|
| 753 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 754 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 755 |
+
| 175 | blk.19.attn_k.weight | Block 19 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 756 |
+
| 176 | blk.19.attn_norm.weight | Block 19 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 757 |
+
| 177 | blk.19.attn_output.weight | Block 19 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 758 |
+
| 178 | blk.19.attn_q.weight | Block 19 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 759 |
+
| 179 | blk.19.attn_v.weight | Block 19 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 760 |
+
| 180 | blk.19.ffn_down.weight | Block 19 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 761 |
+
| 181 | blk.19.ffn_gate.weight | Block 19 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 762 |
+
| 182 | blk.19.ffn_norm.weight | Block 19 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 763 |
+
| 183 | blk.19.ffn_up.weight | Block 19 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 764 |
+
|
| 765 |
+
- Total elements in blk.19: (~218M) 218112000
|
| 766 |
+
- Percentage of total elements: 2.72%
|
| 767 |
+
|
| 768 |
+
|
| 769 |
+
### <a name="blk_20">Block 20 Tensor Group : ~218M Elements</a>
|
| 770 |
+
|
| 771 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 772 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 773 |
+
| 184 | blk.20.attn_k.weight | Block 20 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 774 |
+
| 185 | blk.20.attn_norm.weight | Block 20 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 775 |
+
| 186 | blk.20.attn_output.weight | Block 20 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 776 |
+
| 187 | blk.20.attn_q.weight | Block 20 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 777 |
+
| 188 | blk.20.attn_v.weight | Block 20 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 778 |
+
| 189 | blk.20.ffn_down.weight | Block 20 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 779 |
+
| 190 | blk.20.ffn_gate.weight | Block 20 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 780 |
+
| 191 | blk.20.ffn_norm.weight | Block 20 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 781 |
+
| 192 | blk.20.ffn_up.weight | Block 20 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 782 |
+
|
| 783 |
+
- Total elements in blk.20: (~218M) 218112000
|
| 784 |
+
- Percentage of total elements: 2.72%
|
| 785 |
+
|
| 786 |
+
|
| 787 |
+
### <a name="blk_21">Block 21 Tensor Group : ~218M Elements</a>
|
| 788 |
+
|
| 789 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 790 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 791 |
+
| 193 | blk.21.attn_k.weight | Block 21 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 792 |
+
| 194 | blk.21.attn_norm.weight | Block 21 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 793 |
+
| 195 | blk.21.attn_output.weight | Block 21 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 794 |
+
| 196 | blk.21.attn_q.weight | Block 21 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 795 |
+
| 197 | blk.21.attn_v.weight | Block 21 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 796 |
+
| 198 | blk.21.ffn_down.weight | Block 21 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 797 |
+
| 199 | blk.21.ffn_gate.weight | Block 21 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 798 |
+
| 200 | blk.21.ffn_norm.weight | Block 21 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 799 |
+
| 201 | blk.21.ffn_up.weight | Block 21 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 800 |
+
|
| 801 |
+
- Total elements in blk.21: (~218M) 218112000
|
| 802 |
+
- Percentage of total elements: 2.72%
|
| 803 |
+
|
| 804 |
+
|
| 805 |
+
### <a name="blk_22">Block 22 Tensor Group : ~218M Elements</a>
|
| 806 |
+
|
| 807 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 808 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 809 |
+
| 202 | blk.22.attn_k.weight | Block 22 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 810 |
+
| 203 | blk.22.attn_norm.weight | Block 22 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 811 |
+
| 204 | blk.22.attn_output.weight | Block 22 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 812 |
+
| 205 | blk.22.attn_q.weight | Block 22 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 813 |
+
| 206 | blk.22.attn_v.weight | Block 22 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 814 |
+
| 207 | blk.22.ffn_down.weight | Block 22 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 815 |
+
| 208 | blk.22.ffn_gate.weight | Block 22 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 816 |
+
| 209 | blk.22.ffn_norm.weight | Block 22 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 817 |
+
| 210 | blk.22.ffn_up.weight | Block 22 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 818 |
+
|
| 819 |
+
- Total elements in blk.22: (~218M) 218112000
|
| 820 |
+
- Percentage of total elements: 2.72%
|
| 821 |
+
|
| 822 |
+
|
| 823 |
+
### <a name="blk_23">Block 23 Tensor Group : ~218M Elements</a>
|
| 824 |
+
|
| 825 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 826 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 827 |
+
| 211 | blk.23.attn_k.weight | Block 23 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 828 |
+
| 212 | blk.23.attn_norm.weight | Block 23 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 829 |
+
| 213 | blk.23.attn_output.weight | Block 23 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 830 |
+
| 214 | blk.23.attn_q.weight | Block 23 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 831 |
+
| 215 | blk.23.attn_v.weight | Block 23 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 832 |
+
| 216 | blk.23.ffn_down.weight | Block 23 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 833 |
+
| 217 | blk.23.ffn_gate.weight | Block 23 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 834 |
+
| 218 | blk.23.ffn_norm.weight | Block 23 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 835 |
+
| 219 | blk.23.ffn_up.weight | Block 23 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 836 |
+
|
| 837 |
+
- Total elements in blk.23: (~218M) 218112000
|
| 838 |
+
- Percentage of total elements: 2.72%
|
| 839 |
+
|
| 840 |
+
|
| 841 |
+
### <a name="blk_24">Block 24 Tensor Group : ~218M Elements</a>
|
| 842 |
+
|
| 843 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 844 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 845 |
+
| 220 | blk.24.attn_k.weight | Block 24 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 846 |
+
| 221 | blk.24.attn_norm.weight | Block 24 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 847 |
+
| 222 | blk.24.attn_output.weight | Block 24 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 848 |
+
| 223 | blk.24.attn_q.weight | Block 24 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 849 |
+
| 224 | blk.24.attn_v.weight | Block 24 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 850 |
+
| 225 | blk.24.ffn_down.weight | Block 24 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 851 |
+
| 226 | blk.24.ffn_gate.weight | Block 24 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 852 |
+
| 227 | blk.24.ffn_norm.weight | Block 24 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 853 |
+
| 228 | blk.24.ffn_up.weight | Block 24 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 854 |
+
|
| 855 |
+
- Total elements in blk.24: (~218M) 218112000
|
| 856 |
+
- Percentage of total elements: 2.72%
|
| 857 |
+
|
| 858 |
+
|
| 859 |
+
### <a name="blk_25">Block 25 Tensor Group : ~218M Elements</a>
|
| 860 |
+
|
| 861 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 862 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 863 |
+
| 229 | blk.25.attn_k.weight | Block 25 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 864 |
+
| 230 | blk.25.attn_norm.weight | Block 25 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 865 |
+
| 231 | blk.25.attn_output.weight | Block 25 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 866 |
+
| 232 | blk.25.attn_q.weight | Block 25 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 867 |
+
| 233 | blk.25.attn_v.weight | Block 25 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 868 |
+
| 234 | blk.25.ffn_down.weight | Block 25 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 869 |
+
| 235 | blk.25.ffn_gate.weight | Block 25 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 870 |
+
| 236 | blk.25.ffn_norm.weight | Block 25 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 871 |
+
| 237 | blk.25.ffn_up.weight | Block 25 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 872 |
+
|
| 873 |
+
- Total elements in blk.25: (~218M) 218112000
|
| 874 |
+
- Percentage of total elements: 2.72%
|
| 875 |
+
|
| 876 |
+
|
| 877 |
+
### <a name="blk_26">Block 26 Tensor Group : ~218M Elements</a>
|
| 878 |
+
|
| 879 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 880 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 881 |
+
| 238 | blk.26.attn_k.weight | Block 26 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 882 |
+
| 239 | blk.26.attn_norm.weight | Block 26 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 883 |
+
| 240 | blk.26.attn_output.weight | Block 26 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 884 |
+
| 241 | blk.26.attn_q.weight | Block 26 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 885 |
+
| 242 | blk.26.attn_v.weight | Block 26 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 886 |
+
| 243 | blk.26.ffn_down.weight | Block 26 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 887 |
+
| 244 | blk.26.ffn_gate.weight | Block 26 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 888 |
+
| 245 | blk.26.ffn_norm.weight | Block 26 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 889 |
+
| 246 | blk.26.ffn_up.weight | Block 26 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 890 |
+
|
| 891 |
+
- Total elements in blk.26: (~218M) 218112000
|
| 892 |
+
- Percentage of total elements: 2.72%
|
| 893 |
+
|
| 894 |
+
|
| 895 |
+
### <a name="blk_27">Block 27 Tensor Group : ~218M Elements</a>
|
| 896 |
+
|
| 897 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 898 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 899 |
+
| 247 | blk.27.attn_k.weight | Block 27 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 900 |
+
| 248 | blk.27.attn_norm.weight | Block 27 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 901 |
+
| 249 | blk.27.attn_output.weight | Block 27 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 902 |
+
| 250 | blk.27.attn_q.weight | Block 27 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 903 |
+
| 251 | blk.27.attn_v.weight | Block 27 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 904 |
+
| 252 | blk.27.ffn_down.weight | Block 27 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 905 |
+
| 253 | blk.27.ffn_gate.weight | Block 27 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 906 |
+
| 254 | blk.27.ffn_norm.weight | Block 27 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 907 |
+
| 255 | blk.27.ffn_up.weight | Block 27 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 908 |
+
|
| 909 |
+
- Total elements in blk.27: (~218M) 218112000
|
| 910 |
+
- Percentage of total elements: 2.72%
|
| 911 |
+
|
| 912 |
+
|
| 913 |
+
### <a name="blk_28">Block 28 Tensor Group : ~218M Elements</a>
|
| 914 |
+
|
| 915 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 916 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 917 |
+
| 256 | blk.28.attn_k.weight | Block 28 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 918 |
+
| 257 | blk.28.attn_norm.weight | Block 28 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 919 |
+
| 258 | blk.28.attn_output.weight | Block 28 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 920 |
+
| 259 | blk.28.attn_q.weight | Block 28 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 921 |
+
| 260 | blk.28.attn_v.weight | Block 28 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 922 |
+
| 261 | blk.28.ffn_down.weight | Block 28 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 923 |
+
| 262 | blk.28.ffn_gate.weight | Block 28 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 924 |
+
| 263 | blk.28.ffn_norm.weight | Block 28 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 925 |
+
| 264 | blk.28.ffn_up.weight | Block 28 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 926 |
+
|
| 927 |
+
- Total elements in blk.28: (~218M) 218112000
|
| 928 |
+
- Percentage of total elements: 2.72%
|
| 929 |
+
|
| 930 |
+
|
| 931 |
+
### <a name="blk_29">Block 29 Tensor Group : ~218M Elements</a>
|
| 932 |
+
|
| 933 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 934 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 935 |
+
| 265 | blk.29.attn_k.weight | Block 29 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 936 |
+
| 266 | blk.29.attn_norm.weight | Block 29 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 937 |
+
| 267 | blk.29.attn_output.weight | Block 29 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 938 |
+
| 268 | blk.29.attn_q.weight | Block 29 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 939 |
+
| 269 | blk.29.attn_v.weight | Block 29 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 940 |
+
| 270 | blk.29.ffn_down.weight | Block 29 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 941 |
+
| 271 | blk.29.ffn_gate.weight | Block 29 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 942 |
+
| 272 | blk.29.ffn_norm.weight | Block 29 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 943 |
+
| 273 | blk.29.ffn_up.weight | Block 29 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 944 |
+
|
| 945 |
+
- Total elements in blk.29: (~218M) 218112000
|
| 946 |
+
- Percentage of total elements: 2.72%
|
| 947 |
+
|
| 948 |
+
|
| 949 |
+
### <a name="blk_30">Block 30 Tensor Group : ~218M Elements</a>
|
| 950 |
+
|
| 951 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 952 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 953 |
+
| 274 | blk.30.attn_k.weight | Block 30 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 954 |
+
| 275 | blk.30.attn_norm.weight | Block 30 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 955 |
+
| 276 | blk.30.attn_output.weight | Block 30 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 956 |
+
| 277 | blk.30.attn_q.weight | Block 30 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 957 |
+
| 278 | blk.30.attn_v.weight | Block 30 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 958 |
+
| 279 | blk.30.ffn_down.weight | Block 30 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 959 |
+
| 280 | blk.30.ffn_gate.weight | Block 30 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 960 |
+
| 281 | blk.30.ffn_norm.weight | Block 30 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 961 |
+
| 282 | blk.30.ffn_up.weight | Block 30 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 962 |
+
|
| 963 |
+
- Total elements in blk.30: (~218M) 218112000
|
| 964 |
+
- Percentage of total elements: 2.72%
|
| 965 |
+
|
| 966 |
+
|
| 967 |
+
### <a name="blk_31">Block 31 Tensor Group : ~218M Elements</a>
|
| 968 |
+
|
| 969 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 970 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 971 |
+
| 283 | blk.31.attn_k.weight | Block 31 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q4_K |
|
| 972 |
+
| 284 | blk.31.attn_norm.weight | Block 31 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 973 |
+
| 285 | blk.31.attn_output.weight | Block 31 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 974 |
+
| 286 | blk.31.attn_q.weight | Block 31 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q4_K |
|
| 975 |
+
| 287 | blk.31.attn_v.weight | Block 31 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 976 |
+
| 288 | blk.31.ffn_down.weight | Block 31 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q6_K |
|
| 977 |
+
| 289 | blk.31.ffn_gate.weight | Block 31 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 978 |
+
| 290 | blk.31.ffn_norm.weight | Block 31 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 979 |
+
| 291 | blk.31.ffn_up.weight | Block 31 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 980 |
+
|
| 981 |
+
- Total elements in blk.31: (~218M) 218112000
|
| 982 |
+
- Percentage of total elements: 2.72%
|
scores/Watt-Tool-8B-Q6_K.md
ADDED
|
@@ -0,0 +1,982 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Watt-Tool-8B-Q6_K.gguf - GGUF Internal File Dump
|
| 2 |
+
|
| 3 |
+
- Endian: LITTLE endian
|
| 4 |
+
|
| 5 |
+
## Key Value Metadata Store
|
| 6 |
+
|
| 7 |
+
There are 43 key-value pairs in this file
|
| 8 |
+
|
| 9 |
+
| POS | TYPE | Count | Key | Value |
|
| 10 |
+
|----:|:---------|-------:|:---------------------------------------|:--------------------------------------------------------------------|
|
| 11 |
+
| 1 | UINT32 | 1 | GGUF.version | 3 |
|
| 12 |
+
| 2 | UINT64 | 1 | GGUF.tensor_count | 292 |
|
| 13 |
+
| 3 | UINT64 | 1 | GGUF.kv_count | 40 |
|
| 14 |
+
| 4 | STRING | 1 | general.architecture | `llama` |
|
| 15 |
+
| 5 | STRING | 1 | general.type | `model` |
|
| 16 |
+
| 6 | STRING | 1 | general.name | `Watt Tool 8B GGUF` |
|
| 17 |
+
| 7 | STRING | 1 | general.finetune | `GGUF` |
|
| 18 |
+
| 8 | STRING | 1 | general.basename | `Watt-Tool` |
|
| 19 |
+
| 9 | STRING | 1 | general.size_label | `8B` |
|
| 20 |
+
| 10 | STRING | 1 | general.license | `apache-2.0` |
|
| 21 |
+
| 11 | UINT32 | 1 | general.base_model.count | 1 |
|
| 22 |
+
| 12 | STRING | 1 | general.base_model.0.name | `Llama 3.1 8B Instruct` |
|
| 23 |
+
| 13 | STRING | 1 | general.base_model.0.organization | `Meta Llama` |
|
| 24 |
+
| 14 | STRING | 1 | general.base_model.0.repo_url | `https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct` |
|
| 25 |
+
| 15 | [STRING] | 4 | general.tags | [ `function-calling`, `tool-use`, `llama`, `bfcl` ] |
|
| 26 |
+
| 16 | [STRING] | 1 | general.languages | [ `en` ] |
|
| 27 |
+
| 17 | UINT32 | 1 | llama.block_count | 32 |
|
| 28 |
+
| 18 | UINT32 | 1 | llama.context_length | 131072 |
|
| 29 |
+
| 19 | UINT32 | 1 | llama.embedding_length | 4096 |
|
| 30 |
+
| 20 | UINT32 | 1 | llama.feed_forward_length | 14336 |
|
| 31 |
+
| 21 | UINT32 | 1 | llama.attention.head_count | 32 |
|
| 32 |
+
| 22 | UINT32 | 1 | llama.attention.head_count_kv | 8 |
|
| 33 |
+
| 23 | FLOAT32 | 1 | llama.rope.freq_base | 500000.0 |
|
| 34 |
+
| 24 | FLOAT32 | 1 | llama.attention.layer_norm_rms_epsilon | 1e-05 |
|
| 35 |
+
| 25 | UINT32 | 1 | llama.attention.key_length | 128 |
|
| 36 |
+
| 26 | UINT32 | 1 | llama.attention.value_length | 128 |
|
| 37 |
+
| 27 | UINT32 | 1 | llama.vocab_size | 128256 |
|
| 38 |
+
| 28 | UINT32 | 1 | llama.rope.dimension_count | 128 |
|
| 39 |
+
| 29 | STRING | 1 | tokenizer.ggml.model | `gpt2` |
|
| 40 |
+
| 30 | STRING | 1 | tokenizer.ggml.pre | `llama-bpe` |
|
| 41 |
+
| 31 | [STRING] | 128256 | tokenizer.ggml.tokens | [ `!`, `"`, `#`, `$`, `%`, ... ] |
|
| 42 |
+
| 32 | [INT32] | 128256 | tokenizer.ggml.token_type | [ 1, 1, 1, 1, 1, 1, 1, ... ] |
|
| 43 |
+
| 33 | [STRING] | 280147 | tokenizer.ggml.merges | [ `Ġ Ġ`, `Ġ ĠĠĠ`, `ĠĠ ĠĠ`, `ĠĠĠ Ġ`, `i n`, ... ] |
|
| 44 |
+
| 34 | UINT32 | 1 | tokenizer.ggml.bos_token_id | 128000 |
|
| 45 |
+
| 35 | UINT32 | 1 | tokenizer.ggml.eos_token_id | 128009 |
|
| 46 |
+
| 36 | UINT32 | 1 | tokenizer.ggml.padding_token_id | 128009 |
|
| 47 |
+
| 37 | STRING | 1 | tokenizer.chat_template | `{{ '<|begin_of_text|>' }}{% if`...`d|>' }}{% endif %}{% endfor %}` |
|
| 48 |
+
| 38 | UINT32 | 1 | general.quantization_version | 2 |
|
| 49 |
+
| 39 | UINT32 | 1 | general.file_type | 18 |
|
| 50 |
+
| 40 | STRING | 1 | quantize.imatrix.file | `./imatrix/imatrix-Watt-Tool-8B-small.dat` |
|
| 51 |
+
| 41 | STRING | 1 | quantize.imatrix.dataset | `../../datasets/imatrix/calibration_eur_small.txt` |
|
| 52 |
+
| 42 | INT32 | 1 | quantize.imatrix.entries_count | 225 |
|
| 53 |
+
| 43 | INT32 | 1 | quantize.imatrix.chunks_count | 962 |
|
| 54 |
+
|
| 55 |
+
## Tensors Overview ~8B Elements
|
| 56 |
+
|
| 57 |
+
Total number of elements in all tensors: 8030261312 Elements
|
| 58 |
+
|
| 59 |
+
- [Watt-Tool-8B-Q6\_K.gguf - GGUF Internal File Dump](#watt-tool-8b-q6_kgguf---gguf-internal-file-dump)
|
| 60 |
+
- [Key Value Metadata Store](#key-value-metadata-store)
|
| 61 |
+
- [Tensors Overview ~8B Elements](#tensors-overview-8b-elements)
|
| 62 |
+
- [Tensor Data Offset](#tensor-data-offset)
|
| 63 |
+
- [Base Tensor Group : ~1B Elements](#base-tensor-group--1b-elements)
|
| 64 |
+
- [Block 0 Tensor Group : ~218M Elements](#block-0-tensor-group--218m-elements)
|
| 65 |
+
- [Block 1 Tensor Group : ~218M Elements](#block-1-tensor-group--218m-elements)
|
| 66 |
+
- [Block 2 Tensor Group : ~218M Elements](#block-2-tensor-group--218m-elements)
|
| 67 |
+
- [Block 3 Tensor Group : ~218M Elements](#block-3-tensor-group--218m-elements)
|
| 68 |
+
- [Block 4 Tensor Group : ~218M Elements](#block-4-tensor-group--218m-elements)
|
| 69 |
+
- [Block 5 Tensor Group : ~218M Elements](#block-5-tensor-group--218m-elements)
|
| 70 |
+
- [Block 6 Tensor Group : ~218M Elements](#block-6-tensor-group--218m-elements)
|
| 71 |
+
- [Block 7 Tensor Group : ~218M Elements](#block-7-tensor-group--218m-elements)
|
| 72 |
+
- [Block 8 Tensor Group : ~218M Elements](#block-8-tensor-group--218m-elements)
|
| 73 |
+
- [Block 9 Tensor Group : ~218M Elements](#block-9-tensor-group--218m-elements)
|
| 74 |
+
- [Block 10 Tensor Group : ~218M Elements](#block-10-tensor-group--218m-elements)
|
| 75 |
+
- [Block 11 Tensor Group : ~218M Elements](#block-11-tensor-group--218m-elements)
|
| 76 |
+
- [Block 12 Tensor Group : ~218M Elements](#block-12-tensor-group--218m-elements)
|
| 77 |
+
- [Block 13 Tensor Group : ~218M Elements](#block-13-tensor-group--218m-elements)
|
| 78 |
+
- [Block 14 Tensor Group : ~218M Elements](#block-14-tensor-group--218m-elements)
|
| 79 |
+
- [Block 15 Tensor Group : ~218M Elements](#block-15-tensor-group--218m-elements)
|
| 80 |
+
- [Block 16 Tensor Group : ~218M Elements](#block-16-tensor-group--218m-elements)
|
| 81 |
+
- [Block 17 Tensor Group : ~218M Elements](#block-17-tensor-group--218m-elements)
|
| 82 |
+
- [Block 18 Tensor Group : ~218M Elements](#block-18-tensor-group--218m-elements)
|
| 83 |
+
- [Block 19 Tensor Group : ~218M Elements](#block-19-tensor-group--218m-elements)
|
| 84 |
+
- [Block 20 Tensor Group : ~218M Elements](#block-20-tensor-group--218m-elements)
|
| 85 |
+
- [Block 21 Tensor Group : ~218M Elements](#block-21-tensor-group--218m-elements)
|
| 86 |
+
- [Block 22 Tensor Group : ~218M Elements](#block-22-tensor-group--218m-elements)
|
| 87 |
+
- [Block 23 Tensor Group : ~218M Elements](#block-23-tensor-group--218m-elements)
|
| 88 |
+
- [Block 24 Tensor Group : ~218M Elements](#block-24-tensor-group--218m-elements)
|
| 89 |
+
- [Block 25 Tensor Group : ~218M Elements](#block-25-tensor-group--218m-elements)
|
| 90 |
+
- [Block 26 Tensor Group : ~218M Elements](#block-26-tensor-group--218m-elements)
|
| 91 |
+
- [Block 27 Tensor Group : ~218M Elements](#block-27-tensor-group--218m-elements)
|
| 92 |
+
- [Block 28 Tensor Group : ~218M Elements](#block-28-tensor-group--218m-elements)
|
| 93 |
+
- [Block 29 Tensor Group : ~218M Elements](#block-29-tensor-group--218m-elements)
|
| 94 |
+
- [Block 30 Tensor Group : ~218M Elements](#block-30-tensor-group--218m-elements)
|
| 95 |
+
- [Block 31 Tensor Group : ~218M Elements](#block-31-tensor-group--218m-elements)
|
| 96 |
+
|
| 97 |
+
### Tensor Data Offset
|
| 98 |
+
|
| 99 |
+
This table contains the offset and data segment relative to start of file
|
| 100 |
+
|
| 101 |
+
| T_ID | Tensor Layer Name | Data Offset (B) | Data Size (B) |
|
| 102 |
+
|-----:|:--------------------------|-----------------:|-----------------:|
|
| 103 |
+
| 0 | output.weight | 0x779620 | 0x19afa000 |
|
| 104 |
+
| 1 | output_norm.weight | 0x1a273620 | 0x4000 |
|
| 105 |
+
| 2 | rope_freqs.weight | 0x1a277620 | 0x100 |
|
| 106 |
+
| 3 | token_embd.weight | 0x1a277720 | 0xd746000 |
|
| 107 |
+
| 4 | blk.0.attn_k.weight | 0x279bd720 | 0x2c0000 |
|
| 108 |
+
| 5 | blk.0.attn_norm.weight | 0x27c7d720 | 0x4000 |
|
| 109 |
+
| 6 | blk.0.attn_output.weight | 0x27c81720 | 0xd20000 |
|
| 110 |
+
| 7 | blk.0.attn_q.weight | 0x289a1720 | 0xb00000 |
|
| 111 |
+
| 8 | blk.0.attn_v.weight | 0x294a1720 | 0x348000 |
|
| 112 |
+
| 9 | blk.0.ffn_down.weight | 0x297e9720 | 0x3b80000 |
|
| 113 |
+
| 10 | blk.0.ffn_gate.weight | 0x2d369720 | 0x2680000 |
|
| 114 |
+
| 11 | blk.0.ffn_norm.weight | 0x2f9e9720 | 0x4000 |
|
| 115 |
+
| 12 | blk.0.ffn_up.weight | 0x2f9ed720 | 0x2680000 |
|
| 116 |
+
| 13 | blk.1.attn_k.weight | 0x3206d720 | 0x2c0000 |
|
| 117 |
+
| 14 | blk.1.attn_norm.weight | 0x3232d720 | 0x4000 |
|
| 118 |
+
| 15 | blk.1.attn_output.weight | 0x32331720 | 0xd20000 |
|
| 119 |
+
| 16 | blk.1.attn_q.weight | 0x33051720 | 0xb00000 |
|
| 120 |
+
| 17 | blk.1.attn_v.weight | 0x33b51720 | 0x348000 |
|
| 121 |
+
| 18 | blk.1.ffn_down.weight | 0x33e99720 | 0x3b80000 |
|
| 122 |
+
| 19 | blk.1.ffn_gate.weight | 0x37a19720 | 0x2680000 |
|
| 123 |
+
| 20 | blk.1.ffn_norm.weight | 0x3a099720 | 0x4000 |
|
| 124 |
+
| 21 | blk.1.ffn_up.weight | 0x3a09d720 | 0x2680000 |
|
| 125 |
+
| 22 | blk.2.attn_k.weight | 0x3c71d720 | 0x2c0000 |
|
| 126 |
+
| 23 | blk.2.attn_norm.weight | 0x3c9dd720 | 0x4000 |
|
| 127 |
+
| 24 | blk.2.attn_output.weight | 0x3c9e1720 | 0xd20000 |
|
| 128 |
+
| 25 | blk.2.attn_q.weight | 0x3d701720 | 0xb00000 |
|
| 129 |
+
| 26 | blk.2.attn_v.weight | 0x3e201720 | 0x348000 |
|
| 130 |
+
| 27 | blk.2.ffn_down.weight | 0x3e549720 | 0x3b80000 |
|
| 131 |
+
| 28 | blk.2.ffn_gate.weight | 0x420c9720 | 0x2680000 |
|
| 132 |
+
| 29 | blk.2.ffn_norm.weight | 0x44749720 | 0x4000 |
|
| 133 |
+
| 30 | blk.2.ffn_up.weight | 0x4474d720 | 0x2680000 |
|
| 134 |
+
| 31 | blk.3.attn_k.weight | 0x46dcd720 | 0x2c0000 |
|
| 135 |
+
| 32 | blk.3.attn_norm.weight | 0x4708d720 | 0x4000 |
|
| 136 |
+
| 33 | blk.3.attn_output.weight | 0x47091720 | 0xd20000 |
|
| 137 |
+
| 34 | blk.3.attn_q.weight | 0x47db1720 | 0xb00000 |
|
| 138 |
+
| 35 | blk.3.attn_v.weight | 0x488b1720 | 0x348000 |
|
| 139 |
+
| 36 | blk.3.ffn_down.weight | 0x48bf9720 | 0x3b80000 |
|
| 140 |
+
| 37 | blk.3.ffn_gate.weight | 0x4c779720 | 0x2680000 |
|
| 141 |
+
| 38 | blk.3.ffn_norm.weight | 0x4edf9720 | 0x4000 |
|
| 142 |
+
| 39 | blk.3.ffn_up.weight | 0x4edfd720 | 0x2680000 |
|
| 143 |
+
| 40 | blk.4.attn_k.weight | 0x5147d720 | 0x2c0000 |
|
| 144 |
+
| 41 | blk.4.attn_norm.weight | 0x5173d720 | 0x4000 |
|
| 145 |
+
| 42 | blk.4.attn_output.weight | 0x51741720 | 0xd20000 |
|
| 146 |
+
| 43 | blk.4.attn_q.weight | 0x52461720 | 0xb00000 |
|
| 147 |
+
| 44 | blk.4.attn_v.weight | 0x52f61720 | 0x348000 |
|
| 148 |
+
| 45 | blk.4.ffn_down.weight | 0x532a9720 | 0x3b80000 |
|
| 149 |
+
| 46 | blk.4.ffn_gate.weight | 0x56e29720 | 0x2680000 |
|
| 150 |
+
| 47 | blk.4.ffn_norm.weight | 0x594a9720 | 0x4000 |
|
| 151 |
+
| 48 | blk.4.ffn_up.weight | 0x594ad720 | 0x2680000 |
|
| 152 |
+
| 49 | blk.5.attn_k.weight | 0x5bb2d720 | 0x2c0000 |
|
| 153 |
+
| 50 | blk.5.attn_norm.weight | 0x5bded720 | 0x4000 |
|
| 154 |
+
| 51 | blk.5.attn_output.weight | 0x5bdf1720 | 0xd20000 |
|
| 155 |
+
| 52 | blk.5.attn_q.weight | 0x5cb11720 | 0xb00000 |
|
| 156 |
+
| 53 | blk.5.attn_v.weight | 0x5d611720 | 0x348000 |
|
| 157 |
+
| 54 | blk.5.ffn_down.weight | 0x5d959720 | 0x3b80000 |
|
| 158 |
+
| 55 | blk.5.ffn_gate.weight | 0x614d9720 | 0x2680000 |
|
| 159 |
+
| 56 | blk.5.ffn_norm.weight | 0x63b59720 | 0x4000 |
|
| 160 |
+
| 57 | blk.5.ffn_up.weight | 0x63b5d720 | 0x2680000 |
|
| 161 |
+
| 58 | blk.6.attn_k.weight | 0x661dd720 | 0x2c0000 |
|
| 162 |
+
| 59 | blk.6.attn_norm.weight | 0x6649d720 | 0x4000 |
|
| 163 |
+
| 60 | blk.6.attn_output.weight | 0x664a1720 | 0xd20000 |
|
| 164 |
+
| 61 | blk.6.attn_q.weight | 0x671c1720 | 0xb00000 |
|
| 165 |
+
| 62 | blk.6.attn_v.weight | 0x67cc1720 | 0x348000 |
|
| 166 |
+
| 63 | blk.6.ffn_down.weight | 0x68009720 | 0x3b80000 |
|
| 167 |
+
| 64 | blk.6.ffn_gate.weight | 0x6bb89720 | 0x2680000 |
|
| 168 |
+
| 65 | blk.6.ffn_norm.weight | 0x6e209720 | 0x4000 |
|
| 169 |
+
| 66 | blk.6.ffn_up.weight | 0x6e20d720 | 0x2680000 |
|
| 170 |
+
| 67 | blk.7.attn_k.weight | 0x7088d720 | 0x2c0000 |
|
| 171 |
+
| 68 | blk.7.attn_norm.weight | 0x70b4d720 | 0x4000 |
|
| 172 |
+
| 69 | blk.7.attn_output.weight | 0x70b51720 | 0xd20000 |
|
| 173 |
+
| 70 | blk.7.attn_q.weight | 0x71871720 | 0xb00000 |
|
| 174 |
+
| 71 | blk.7.attn_v.weight | 0x72371720 | 0x348000 |
|
| 175 |
+
| 72 | blk.7.ffn_down.weight | 0x726b9720 | 0x3b80000 |
|
| 176 |
+
| 73 | blk.7.ffn_gate.weight | 0x76239720 | 0x2680000 |
|
| 177 |
+
| 74 | blk.7.ffn_norm.weight | 0x788b9720 | 0x4000 |
|
| 178 |
+
| 75 | blk.7.ffn_up.weight | 0x788bd720 | 0x2680000 |
|
| 179 |
+
| 76 | blk.8.attn_k.weight | 0x7af3d720 | 0x2c0000 |
|
| 180 |
+
| 77 | blk.8.attn_norm.weight | 0x7b1fd720 | 0x4000 |
|
| 181 |
+
| 78 | blk.8.attn_output.weight | 0x7b201720 | 0xd20000 |
|
| 182 |
+
| 79 | blk.8.attn_q.weight | 0x7bf21720 | 0xb00000 |
|
| 183 |
+
| 80 | blk.8.attn_v.weight | 0x7ca21720 | 0x348000 |
|
| 184 |
+
| 81 | blk.8.ffn_down.weight | 0x7cd69720 | 0x3b80000 |
|
| 185 |
+
| 82 | blk.8.ffn_gate.weight | 0x808e9720 | 0x2680000 |
|
| 186 |
+
| 83 | blk.8.ffn_norm.weight | 0x82f69720 | 0x4000 |
|
| 187 |
+
| 84 | blk.8.ffn_up.weight | 0x82f6d720 | 0x2680000 |
|
| 188 |
+
| 85 | blk.9.attn_k.weight | 0x855ed720 | 0x2c0000 |
|
| 189 |
+
| 86 | blk.9.attn_norm.weight | 0x858ad720 | 0x4000 |
|
| 190 |
+
| 87 | blk.9.attn_output.weight | 0x858b1720 | 0xd20000 |
|
| 191 |
+
| 88 | blk.9.attn_q.weight | 0x865d1720 | 0xb00000 |
|
| 192 |
+
| 89 | blk.9.attn_v.weight | 0x870d1720 | 0x348000 |
|
| 193 |
+
| 90 | blk.9.ffn_down.weight | 0x87419720 | 0x3b80000 |
|
| 194 |
+
| 91 | blk.9.ffn_gate.weight | 0x8af99720 | 0x2680000 |
|
| 195 |
+
| 92 | blk.9.ffn_norm.weight | 0x8d619720 | 0x4000 |
|
| 196 |
+
| 93 | blk.9.ffn_up.weight | 0x8d61d720 | 0x2680000 |
|
| 197 |
+
| 94 | blk.10.attn_k.weight | 0x8fc9d720 | 0x2c0000 |
|
| 198 |
+
| 95 | blk.10.attn_norm.weight | 0x8ff5d720 | 0x4000 |
|
| 199 |
+
| 96 | blk.10.attn_output.weight | 0x8ff61720 | 0xd20000 |
|
| 200 |
+
| 97 | blk.10.attn_q.weight | 0x90c81720 | 0xb00000 |
|
| 201 |
+
| 98 | blk.10.attn_v.weight | 0x91781720 | 0x348000 |
|
| 202 |
+
| 99 | blk.10.ffn_down.weight | 0x91ac9720 | 0x3b80000 |
|
| 203 |
+
| 100 | blk.10.ffn_gate.weight | 0x95649720 | 0x2680000 |
|
| 204 |
+
| 101 | blk.10.ffn_norm.weight | 0x97cc9720 | 0x4000 |
|
| 205 |
+
| 102 | blk.10.ffn_up.weight | 0x97ccd720 | 0x2680000 |
|
| 206 |
+
| 103 | blk.11.attn_k.weight | 0x9a34d720 | 0x2c0000 |
|
| 207 |
+
| 104 | blk.11.attn_norm.weight | 0x9a60d720 | 0x4000 |
|
| 208 |
+
| 105 | blk.11.attn_output.weight | 0x9a611720 | 0xd20000 |
|
| 209 |
+
| 106 | blk.11.attn_q.weight | 0x9b331720 | 0xb00000 |
|
| 210 |
+
| 107 | blk.11.attn_v.weight | 0x9be31720 | 0x348000 |
|
| 211 |
+
| 108 | blk.11.ffn_down.weight | 0x9c179720 | 0x3b80000 |
|
| 212 |
+
| 109 | blk.11.ffn_gate.weight | 0x9fcf9720 | 0x2680000 |
|
| 213 |
+
| 110 | blk.11.ffn_norm.weight | 0xa2379720 | 0x4000 |
|
| 214 |
+
| 111 | blk.11.ffn_up.weight | 0xa237d720 | 0x2680000 |
|
| 215 |
+
| 112 | blk.12.attn_k.weight | 0xa49fd720 | 0x2c0000 |
|
| 216 |
+
| 113 | blk.12.attn_norm.weight | 0xa4cbd720 | 0x4000 |
|
| 217 |
+
| 114 | blk.12.attn_output.weight | 0xa4cc1720 | 0xd20000 |
|
| 218 |
+
| 115 | blk.12.attn_q.weight | 0xa59e1720 | 0xb00000 |
|
| 219 |
+
| 116 | blk.12.attn_v.weight | 0xa64e1720 | 0x348000 |
|
| 220 |
+
| 117 | blk.12.ffn_down.weight | 0xa6829720 | 0x3b80000 |
|
| 221 |
+
| 118 | blk.12.ffn_gate.weight | 0xaa3a9720 | 0x2680000 |
|
| 222 |
+
| 119 | blk.12.ffn_norm.weight | 0xaca29720 | 0x4000 |
|
| 223 |
+
| 120 | blk.12.ffn_up.weight | 0xaca2d720 | 0x2680000 |
|
| 224 |
+
| 121 | blk.13.attn_k.weight | 0xaf0ad720 | 0x348000 |
|
| 225 |
+
| 122 | blk.13.attn_norm.weight | 0xaf3f5720 | 0x4000 |
|
| 226 |
+
| 123 | blk.13.attn_output.weight | 0xaf3f9720 | 0xd20000 |
|
| 227 |
+
| 124 | blk.13.attn_q.weight | 0xb0119720 | 0xd20000 |
|
| 228 |
+
| 125 | blk.13.attn_v.weight | 0xb0e39720 | 0x440000 |
|
| 229 |
+
| 126 | blk.13.ffn_down.weight | 0xb1279720 | 0x3b80000 |
|
| 230 |
+
| 127 | blk.13.ffn_gate.weight | 0xb4df9720 | 0x2680000 |
|
| 231 |
+
| 128 | blk.13.ffn_norm.weight | 0xb7479720 | 0x4000 |
|
| 232 |
+
| 129 | blk.13.ffn_up.weight | 0xb747d720 | 0x2680000 |
|
| 233 |
+
| 130 | blk.14.attn_k.weight | 0xb9afd720 | 0x348000 |
|
| 234 |
+
| 131 | blk.14.attn_norm.weight | 0xb9e45720 | 0x4000 |
|
| 235 |
+
| 132 | blk.14.attn_output.weight | 0xb9e49720 | 0xd20000 |
|
| 236 |
+
| 133 | blk.14.attn_q.weight | 0xbab69720 | 0xd20000 |
|
| 237 |
+
| 134 | blk.14.attn_v.weight | 0xbb889720 | 0x440000 |
|
| 238 |
+
| 135 | blk.14.ffn_down.weight | 0xbbcc9720 | 0x3b80000 |
|
| 239 |
+
| 136 | blk.14.ffn_gate.weight | 0xbf849720 | 0x2680000 |
|
| 240 |
+
| 137 | blk.14.ffn_norm.weight | 0xc1ec9720 | 0x4000 |
|
| 241 |
+
| 138 | blk.14.ffn_up.weight | 0xc1ecd720 | 0x2680000 |
|
| 242 |
+
| 139 | blk.15.attn_k.weight | 0xc454d720 | 0x2c0000 |
|
| 243 |
+
| 140 | blk.15.attn_norm.weight | 0xc480d720 | 0x4000 |
|
| 244 |
+
| 141 | blk.15.attn_output.weight | 0xc4811720 | 0xd20000 |
|
| 245 |
+
| 142 | blk.15.attn_q.weight | 0xc5531720 | 0xb00000 |
|
| 246 |
+
| 143 | blk.15.attn_v.weight | 0xc6031720 | 0x348000 |
|
| 247 |
+
| 144 | blk.15.ffn_down.weight | 0xc6379720 | 0x3b80000 |
|
| 248 |
+
| 145 | blk.15.ffn_gate.weight | 0xc9ef9720 | 0x2680000 |
|
| 249 |
+
| 146 | blk.15.ffn_norm.weight | 0xcc579720 | 0x4000 |
|
| 250 |
+
| 147 | blk.15.ffn_up.weight | 0xcc57d720 | 0x2680000 |
|
| 251 |
+
| 148 | blk.16.attn_k.weight | 0xcebfd720 | 0x348000 |
|
| 252 |
+
| 149 | blk.16.attn_norm.weight | 0xcef45720 | 0x4000 |
|
| 253 |
+
| 150 | blk.16.attn_output.weight | 0xcef49720 | 0xd20000 |
|
| 254 |
+
| 151 | blk.16.attn_q.weight | 0xcfc69720 | 0xd20000 |
|
| 255 |
+
| 152 | blk.16.attn_v.weight | 0xd0989720 | 0x440000 |
|
| 256 |
+
| 153 | blk.16.ffn_down.weight | 0xd0dc9720 | 0x3b80000 |
|
| 257 |
+
| 154 | blk.16.ffn_gate.weight | 0xd4949720 | 0x2df0000 |
|
| 258 |
+
| 155 | blk.16.ffn_norm.weight | 0xd7739720 | 0x4000 |
|
| 259 |
+
| 156 | blk.16.ffn_up.weight | 0xd773d720 | 0x2df0000 |
|
| 260 |
+
| 157 | blk.17.attn_k.weight | 0xda52d720 | 0x2c0000 |
|
| 261 |
+
| 158 | blk.17.attn_norm.weight | 0xda7ed720 | 0x4000 |
|
| 262 |
+
| 159 | blk.17.attn_output.weight | 0xda7f1720 | 0xd20000 |
|
| 263 |
+
| 160 | blk.17.attn_q.weight | 0xdb511720 | 0xb00000 |
|
| 264 |
+
| 161 | blk.17.attn_v.weight | 0xdc011720 | 0x348000 |
|
| 265 |
+
| 162 | blk.17.ffn_down.weight | 0xdc359720 | 0x3b80000 |
|
| 266 |
+
| 163 | blk.17.ffn_gate.weight | 0xdfed9720 | 0x2df0000 |
|
| 267 |
+
| 164 | blk.17.ffn_norm.weight | 0xe2cc9720 | 0x4000 |
|
| 268 |
+
| 165 | blk.17.ffn_up.weight | 0xe2ccd720 | 0x2df0000 |
|
| 269 |
+
| 166 | blk.18.attn_k.weight | 0xe5abd720 | 0x348000 |
|
| 270 |
+
| 167 | blk.18.attn_norm.weight | 0xe5e05720 | 0x4000 |
|
| 271 |
+
| 168 | blk.18.attn_output.weight | 0xe5e09720 | 0xd20000 |
|
| 272 |
+
| 169 | blk.18.attn_q.weight | 0xe6b29720 | 0xd20000 |
|
| 273 |
+
| 170 | blk.18.attn_v.weight | 0xe7849720 | 0x440000 |
|
| 274 |
+
| 171 | blk.18.ffn_down.weight | 0xe7c89720 | 0x3b80000 |
|
| 275 |
+
| 172 | blk.18.ffn_gate.weight | 0xeb809720 | 0x2df0000 |
|
| 276 |
+
| 173 | blk.18.ffn_norm.weight | 0xee5f9720 | 0x4000 |
|
| 277 |
+
| 174 | blk.18.ffn_up.weight | 0xee5fd720 | 0x2df0000 |
|
| 278 |
+
| 175 | blk.19.attn_k.weight | 0xf13ed720 | 0x348000 |
|
| 279 |
+
| 176 | blk.19.attn_norm.weight | 0xf1735720 | 0x4000 |
|
| 280 |
+
| 177 | blk.19.attn_output.weight | 0xf1739720 | 0xd20000 |
|
| 281 |
+
| 178 | blk.19.attn_q.weight | 0xf2459720 | 0xd20000 |
|
| 282 |
+
| 179 | blk.19.attn_v.weight | 0xf3179720 | 0x440000 |
|
| 283 |
+
| 180 | blk.19.ffn_down.weight | 0xf35b9720 | 0x3b80000 |
|
| 284 |
+
| 181 | blk.19.ffn_gate.weight | 0xf7139720 | 0x2df0000 |
|
| 285 |
+
| 182 | blk.19.ffn_norm.weight | 0xf9f29720 | 0x4000 |
|
| 286 |
+
| 183 | blk.19.ffn_up.weight | 0xf9f2d720 | 0x2df0000 |
|
| 287 |
+
| 184 | blk.20.attn_k.weight | 0xfcd1d720 | 0x348000 |
|
| 288 |
+
| 185 | blk.20.attn_norm.weight | 0xfd065720 | 0x4000 |
|
| 289 |
+
| 186 | blk.20.attn_output.weight | 0xfd069720 | 0xd20000 |
|
| 290 |
+
| 187 | blk.20.attn_q.weight | 0xfdd89720 | 0xd20000 |
|
| 291 |
+
| 188 | blk.20.attn_v.weight | 0xfeaa9720 | 0x440000 |
|
| 292 |
+
| 189 | blk.20.ffn_down.weight | 0xfeee9720 | 0x3b80000 |
|
| 293 |
+
| 190 | blk.20.ffn_gate.weight | 0x102a69720 | 0x2df0000 |
|
| 294 |
+
| 191 | blk.20.ffn_norm.weight | 0x105859720 | 0x4000 |
|
| 295 |
+
| 192 | blk.20.ffn_up.weight | 0x10585d720 | 0x2df0000 |
|
| 296 |
+
| 193 | blk.21.attn_k.weight | 0x10864d720 | 0x348000 |
|
| 297 |
+
| 194 | blk.21.attn_norm.weight | 0x108995720 | 0x4000 |
|
| 298 |
+
| 195 | blk.21.attn_output.weight | 0x108999720 | 0xd20000 |
|
| 299 |
+
| 196 | blk.21.attn_q.weight | 0x1096b9720 | 0xd20000 |
|
| 300 |
+
| 197 | blk.21.attn_v.weight | 0x10a3d9720 | 0x440000 |
|
| 301 |
+
| 198 | blk.21.ffn_down.weight | 0x10a819720 | 0x3b80000 |
|
| 302 |
+
| 199 | blk.21.ffn_gate.weight | 0x10e399720 | 0x2df0000 |
|
| 303 |
+
| 200 | blk.21.ffn_norm.weight | 0x111189720 | 0x4000 |
|
| 304 |
+
| 201 | blk.21.ffn_up.weight | 0x11118d720 | 0x2df0000 |
|
| 305 |
+
| 202 | blk.22.attn_k.weight | 0x113f7d720 | 0x348000 |
|
| 306 |
+
| 203 | blk.22.attn_norm.weight | 0x1142c5720 | 0x4000 |
|
| 307 |
+
| 204 | blk.22.attn_output.weight | 0x1142c9720 | 0xd20000 |
|
| 308 |
+
| 205 | blk.22.attn_q.weight | 0x114fe9720 | 0xd20000 |
|
| 309 |
+
| 206 | blk.22.attn_v.weight | 0x115d09720 | 0x440000 |
|
| 310 |
+
| 207 | blk.22.ffn_down.weight | 0x116149720 | 0x3b80000 |
|
| 311 |
+
| 208 | blk.22.ffn_gate.weight | 0x119cc9720 | 0x2df0000 |
|
| 312 |
+
| 209 | blk.22.ffn_norm.weight | 0x11cab9720 | 0x4000 |
|
| 313 |
+
| 210 | blk.22.ffn_up.weight | 0x11cabd720 | 0x2df0000 |
|
| 314 |
+
| 211 | blk.23.attn_k.weight | 0x11f8ad720 | 0x348000 |
|
| 315 |
+
| 212 | blk.23.attn_norm.weight | 0x11fbf5720 | 0x4000 |
|
| 316 |
+
| 213 | blk.23.attn_output.weight | 0x11fbf9720 | 0xd20000 |
|
| 317 |
+
| 214 | blk.23.attn_q.weight | 0x120919720 | 0xd20000 |
|
| 318 |
+
| 215 | blk.23.attn_v.weight | 0x121639720 | 0x440000 |
|
| 319 |
+
| 216 | blk.23.ffn_down.weight | 0x121a79720 | 0x3b80000 |
|
| 320 |
+
| 217 | blk.23.ffn_gate.weight | 0x1255f9720 | 0x2df0000 |
|
| 321 |
+
| 218 | blk.23.ffn_norm.weight | 0x1283e9720 | 0x4000 |
|
| 322 |
+
| 219 | blk.23.ffn_up.weight | 0x1283ed720 | 0x2df0000 |
|
| 323 |
+
| 220 | blk.24.attn_k.weight | 0x12b1dd720 | 0x348000 |
|
| 324 |
+
| 221 | blk.24.attn_norm.weight | 0x12b525720 | 0x4000 |
|
| 325 |
+
| 222 | blk.24.attn_output.weight | 0x12b529720 | 0xd20000 |
|
| 326 |
+
| 223 | blk.24.attn_q.weight | 0x12c249720 | 0xd20000 |
|
| 327 |
+
| 224 | blk.24.attn_v.weight | 0x12cf69720 | 0x440000 |
|
| 328 |
+
| 225 | blk.24.ffn_down.weight | 0x12d3a9720 | 0x3b80000 |
|
| 329 |
+
| 226 | blk.24.ffn_gate.weight | 0x130f29720 | 0x2df0000 |
|
| 330 |
+
| 227 | blk.24.ffn_norm.weight | 0x133d19720 | 0x4000 |
|
| 331 |
+
| 228 | blk.24.ffn_up.weight | 0x133d1d720 | 0x2df0000 |
|
| 332 |
+
| 229 | blk.25.attn_k.weight | 0x136b0d720 | 0x348000 |
|
| 333 |
+
| 230 | blk.25.attn_norm.weight | 0x136e55720 | 0x4000 |
|
| 334 |
+
| 231 | blk.25.attn_output.weight | 0x136e59720 | 0xd20000 |
|
| 335 |
+
| 232 | blk.25.attn_q.weight | 0x137b79720 | 0xd20000 |
|
| 336 |
+
| 233 | blk.25.attn_v.weight | 0x138899720 | 0x440000 |
|
| 337 |
+
| 234 | blk.25.ffn_down.weight | 0x138cd9720 | 0x3b80000 |
|
| 338 |
+
| 235 | blk.25.ffn_gate.weight | 0x13c859720 | 0x2df0000 |
|
| 339 |
+
| 236 | blk.25.ffn_norm.weight | 0x13f649720 | 0x4000 |
|
| 340 |
+
| 237 | blk.25.ffn_up.weight | 0x13f64d720 | 0x2df0000 |
|
| 341 |
+
| 238 | blk.26.attn_k.weight | 0x14243d720 | 0x348000 |
|
| 342 |
+
| 239 | blk.26.attn_norm.weight | 0x142785720 | 0x4000 |
|
| 343 |
+
| 240 | blk.26.attn_output.weight | 0x142789720 | 0xd20000 |
|
| 344 |
+
| 241 | blk.26.attn_q.weight | 0x1434a9720 | 0xd20000 |
|
| 345 |
+
| 242 | blk.26.attn_v.weight | 0x1441c9720 | 0x440000 |
|
| 346 |
+
| 243 | blk.26.ffn_down.weight | 0x144609720 | 0x3b80000 |
|
| 347 |
+
| 244 | blk.26.ffn_gate.weight | 0x148189720 | 0x2df0000 |
|
| 348 |
+
| 245 | blk.26.ffn_norm.weight | 0x14af79720 | 0x4000 |
|
| 349 |
+
| 246 | blk.26.ffn_up.weight | 0x14af7d720 | 0x2df0000 |
|
| 350 |
+
| 247 | blk.27.attn_k.weight | 0x14dd6d720 | 0x348000 |
|
| 351 |
+
| 248 | blk.27.attn_norm.weight | 0x14e0b5720 | 0x4000 |
|
| 352 |
+
| 249 | blk.27.attn_output.weight | 0x14e0b9720 | 0xd20000 |
|
| 353 |
+
| 250 | blk.27.attn_q.weight | 0x14edd9720 | 0xd20000 |
|
| 354 |
+
| 251 | blk.27.attn_v.weight | 0x14faf9720 | 0x440000 |
|
| 355 |
+
| 252 | blk.27.ffn_down.weight | 0x14ff39720 | 0x3b80000 |
|
| 356 |
+
| 253 | blk.27.ffn_gate.weight | 0x153ab9720 | 0x2df0000 |
|
| 357 |
+
| 254 | blk.27.ffn_norm.weight | 0x1568a9720 | 0x4000 |
|
| 358 |
+
| 255 | blk.27.ffn_up.weight | 0x1568ad720 | 0x2df0000 |
|
| 359 |
+
| 256 | blk.28.attn_k.weight | 0x15969d720 | 0x348000 |
|
| 360 |
+
| 257 | blk.28.attn_norm.weight | 0x1599e5720 | 0x4000 |
|
| 361 |
+
| 258 | blk.28.attn_output.weight | 0x1599e9720 | 0xd20000 |
|
| 362 |
+
| 259 | blk.28.attn_q.weight | 0x15a709720 | 0xd20000 |
|
| 363 |
+
| 260 | blk.28.attn_v.weight | 0x15b429720 | 0x440000 |
|
| 364 |
+
| 261 | blk.28.ffn_down.weight | 0x15b869720 | 0x3b80000 |
|
| 365 |
+
| 262 | blk.28.ffn_gate.weight | 0x15f3e9720 | 0x2df0000 |
|
| 366 |
+
| 263 | blk.28.ffn_norm.weight | 0x1621d9720 | 0x4000 |
|
| 367 |
+
| 264 | blk.28.ffn_up.weight | 0x1621dd720 | 0x2df0000 |
|
| 368 |
+
| 265 | blk.29.attn_k.weight | 0x164fcd720 | 0x348000 |
|
| 369 |
+
| 266 | blk.29.attn_norm.weight | 0x165315720 | 0x4000 |
|
| 370 |
+
| 267 | blk.29.attn_output.weight | 0x165319720 | 0xd20000 |
|
| 371 |
+
| 268 | blk.29.attn_q.weight | 0x166039720 | 0xd20000 |
|
| 372 |
+
| 269 | blk.29.attn_v.weight | 0x166d59720 | 0x440000 |
|
| 373 |
+
| 270 | blk.29.ffn_down.weight | 0x167199720 | 0x3b80000 |
|
| 374 |
+
| 271 | blk.29.ffn_gate.weight | 0x16ad19720 | 0x2df0000 |
|
| 375 |
+
| 272 | blk.29.ffn_norm.weight | 0x16db09720 | 0x4000 |
|
| 376 |
+
| 273 | blk.29.ffn_up.weight | 0x16db0d720 | 0x2df0000 |
|
| 377 |
+
| 274 | blk.30.attn_k.weight | 0x1708fd720 | 0x348000 |
|
| 378 |
+
| 275 | blk.30.attn_norm.weight | 0x170c45720 | 0x4000 |
|
| 379 |
+
| 276 | blk.30.attn_output.weight | 0x170c49720 | 0xd20000 |
|
| 380 |
+
| 277 | blk.30.attn_q.weight | 0x171969720 | 0xd20000 |
|
| 381 |
+
| 278 | blk.30.attn_v.weight | 0x172689720 | 0x440000 |
|
| 382 |
+
| 279 | blk.30.ffn_down.weight | 0x172ac9720 | 0x3b80000 |
|
| 383 |
+
| 280 | blk.30.ffn_gate.weight | 0x176649720 | 0x2df0000 |
|
| 384 |
+
| 281 | blk.30.ffn_norm.weight | 0x179439720 | 0x4000 |
|
| 385 |
+
| 282 | blk.30.ffn_up.weight | 0x17943d720 | 0x2df0000 |
|
| 386 |
+
| 283 | blk.31.attn_k.weight | 0x17c22d720 | 0x2c0000 |
|
| 387 |
+
| 284 | blk.31.attn_norm.weight | 0x17c4ed720 | 0x4000 |
|
| 388 |
+
| 285 | blk.31.attn_output.weight | 0x17c4f1720 | 0xd20000 |
|
| 389 |
+
| 286 | blk.31.attn_q.weight | 0x17d211720 | 0xb00000 |
|
| 390 |
+
| 287 | blk.31.attn_v.weight | 0x17dd11720 | 0x348000 |
|
| 391 |
+
| 288 | blk.31.ffn_down.weight | 0x17e059720 | 0x3b80000 |
|
| 392 |
+
| 289 | blk.31.ffn_gate.weight | 0x181bd9720 | 0x2df0000 |
|
| 393 |
+
| 290 | blk.31.ffn_norm.weight | 0x1849c9720 | 0x4000 |
|
| 394 |
+
| 291 | blk.31.ffn_up.weight | 0x1849cd720 | 0x2df0000 |
|
| 395 |
+
|
| 396 |
+
### <a name="base">Base Tensor Group : ~1B Elements</a>
|
| 397 |
+
|
| 398 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 399 |
+
|-----:|:-------------------|:---------------------------------|:------------------|:----------------------|:-----|
|
| 400 |
+
| 0 | output.weight | Output (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | Q6_K |
|
| 401 |
+
| 1 | output_norm.weight | Output Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 402 |
+
| 2 | rope_freqs.weight | Rope_Freqs (W) | ( 64) 64 | 64 x 1 x 1 x 1 | F32 |
|
| 403 |
+
| 3 | token_embd.weight | Token Embedding (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | Q3_K |
|
| 404 |
+
|
| 405 |
+
- Total elements in base: ( ~1B) 1050677312
|
| 406 |
+
- Percentage of total elements: 13.08%
|
| 407 |
+
|
| 408 |
+
|
| 409 |
+
### <a name="blk_0">Block 0 Tensor Group : ~218M Elements</a>
|
| 410 |
+
|
| 411 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 412 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 413 |
+
| 4 | blk.0.attn_k.weight | Block 0 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 414 |
+
| 5 | blk.0.attn_norm.weight | Block 0 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 415 |
+
| 6 | blk.0.attn_output.weight | Block 0 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 416 |
+
| 7 | blk.0.attn_q.weight | Block 0 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 417 |
+
| 8 | blk.0.attn_v.weight | Block 0 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 418 |
+
| 9 | blk.0.ffn_down.weight | Block 0 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 419 |
+
| 10 | blk.0.ffn_gate.weight | Block 0 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 420 |
+
| 11 | blk.0.ffn_norm.weight | Block 0 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 421 |
+
| 12 | blk.0.ffn_up.weight | Block 0 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 422 |
+
|
| 423 |
+
- Total elements in blk.0: (~218M) 218112000
|
| 424 |
+
- Percentage of total elements: 2.72%
|
| 425 |
+
|
| 426 |
+
|
| 427 |
+
### <a name="blk_1">Block 1 Tensor Group : ~218M Elements</a>
|
| 428 |
+
|
| 429 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 430 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 431 |
+
| 13 | blk.1.attn_k.weight | Block 1 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 432 |
+
| 14 | blk.1.attn_norm.weight | Block 1 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 433 |
+
| 15 | blk.1.attn_output.weight | Block 1 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 434 |
+
| 16 | blk.1.attn_q.weight | Block 1 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 435 |
+
| 17 | blk.1.attn_v.weight | Block 1 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 436 |
+
| 18 | blk.1.ffn_down.weight | Block 1 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 437 |
+
| 19 | blk.1.ffn_gate.weight | Block 1 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 438 |
+
| 20 | blk.1.ffn_norm.weight | Block 1 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 439 |
+
| 21 | blk.1.ffn_up.weight | Block 1 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 440 |
+
|
| 441 |
+
- Total elements in blk.1: (~218M) 218112000
|
| 442 |
+
- Percentage of total elements: 2.72%
|
| 443 |
+
|
| 444 |
+
|
| 445 |
+
### <a name="blk_2">Block 2 Tensor Group : ~218M Elements</a>
|
| 446 |
+
|
| 447 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 448 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 449 |
+
| 22 | blk.2.attn_k.weight | Block 2 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 450 |
+
| 23 | blk.2.attn_norm.weight | Block 2 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 451 |
+
| 24 | blk.2.attn_output.weight | Block 2 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 452 |
+
| 25 | blk.2.attn_q.weight | Block 2 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 453 |
+
| 26 | blk.2.attn_v.weight | Block 2 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 454 |
+
| 27 | blk.2.ffn_down.weight | Block 2 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 455 |
+
| 28 | blk.2.ffn_gate.weight | Block 2 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 456 |
+
| 29 | blk.2.ffn_norm.weight | Block 2 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 457 |
+
| 30 | blk.2.ffn_up.weight | Block 2 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 458 |
+
|
| 459 |
+
- Total elements in blk.2: (~218M) 218112000
|
| 460 |
+
- Percentage of total elements: 2.72%
|
| 461 |
+
|
| 462 |
+
|
| 463 |
+
### <a name="blk_3">Block 3 Tensor Group : ~218M Elements</a>
|
| 464 |
+
|
| 465 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 466 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 467 |
+
| 31 | blk.3.attn_k.weight | Block 3 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 468 |
+
| 32 | blk.3.attn_norm.weight | Block 3 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 469 |
+
| 33 | blk.3.attn_output.weight | Block 3 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 470 |
+
| 34 | blk.3.attn_q.weight | Block 3 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 471 |
+
| 35 | blk.3.attn_v.weight | Block 3 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 472 |
+
| 36 | blk.3.ffn_down.weight | Block 3 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 473 |
+
| 37 | blk.3.ffn_gate.weight | Block 3 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 474 |
+
| 38 | blk.3.ffn_norm.weight | Block 3 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 475 |
+
| 39 | blk.3.ffn_up.weight | Block 3 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 476 |
+
|
| 477 |
+
- Total elements in blk.3: (~218M) 218112000
|
| 478 |
+
- Percentage of total elements: 2.72%
|
| 479 |
+
|
| 480 |
+
|
| 481 |
+
### <a name="blk_4">Block 4 Tensor Group : ~218M Elements</a>
|
| 482 |
+
|
| 483 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 484 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 485 |
+
| 40 | blk.4.attn_k.weight | Block 4 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 486 |
+
| 41 | blk.4.attn_norm.weight | Block 4 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 487 |
+
| 42 | blk.4.attn_output.weight | Block 4 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 488 |
+
| 43 | blk.4.attn_q.weight | Block 4 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 489 |
+
| 44 | blk.4.attn_v.weight | Block 4 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 490 |
+
| 45 | blk.4.ffn_down.weight | Block 4 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 491 |
+
| 46 | blk.4.ffn_gate.weight | Block 4 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 492 |
+
| 47 | blk.4.ffn_norm.weight | Block 4 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 493 |
+
| 48 | blk.4.ffn_up.weight | Block 4 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 494 |
+
|
| 495 |
+
- Total elements in blk.4: (~218M) 218112000
|
| 496 |
+
- Percentage of total elements: 2.72%
|
| 497 |
+
|
| 498 |
+
|
| 499 |
+
### <a name="blk_5">Block 5 Tensor Group : ~218M Elements</a>
|
| 500 |
+
|
| 501 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 502 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 503 |
+
| 49 | blk.5.attn_k.weight | Block 5 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 504 |
+
| 50 | blk.5.attn_norm.weight | Block 5 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 505 |
+
| 51 | blk.5.attn_output.weight | Block 5 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 506 |
+
| 52 | blk.5.attn_q.weight | Block 5 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 507 |
+
| 53 | blk.5.attn_v.weight | Block 5 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 508 |
+
| 54 | blk.5.ffn_down.weight | Block 5 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 509 |
+
| 55 | blk.5.ffn_gate.weight | Block 5 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 510 |
+
| 56 | blk.5.ffn_norm.weight | Block 5 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 511 |
+
| 57 | blk.5.ffn_up.weight | Block 5 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 512 |
+
|
| 513 |
+
- Total elements in blk.5: (~218M) 218112000
|
| 514 |
+
- Percentage of total elements: 2.72%
|
| 515 |
+
|
| 516 |
+
|
| 517 |
+
### <a name="blk_6">Block 6 Tensor Group : ~218M Elements</a>
|
| 518 |
+
|
| 519 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 520 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 521 |
+
| 58 | blk.6.attn_k.weight | Block 6 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 522 |
+
| 59 | blk.6.attn_norm.weight | Block 6 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 523 |
+
| 60 | blk.6.attn_output.weight | Block 6 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 524 |
+
| 61 | blk.6.attn_q.weight | Block 6 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 525 |
+
| 62 | blk.6.attn_v.weight | Block 6 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 526 |
+
| 63 | blk.6.ffn_down.weight | Block 6 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 527 |
+
| 64 | blk.6.ffn_gate.weight | Block 6 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 528 |
+
| 65 | blk.6.ffn_norm.weight | Block 6 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 529 |
+
| 66 | blk.6.ffn_up.weight | Block 6 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 530 |
+
|
| 531 |
+
- Total elements in blk.6: (~218M) 218112000
|
| 532 |
+
- Percentage of total elements: 2.72%
|
| 533 |
+
|
| 534 |
+
|
| 535 |
+
### <a name="blk_7">Block 7 Tensor Group : ~218M Elements</a>
|
| 536 |
+
|
| 537 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 538 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 539 |
+
| 67 | blk.7.attn_k.weight | Block 7 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 540 |
+
| 68 | blk.7.attn_norm.weight | Block 7 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 541 |
+
| 69 | blk.7.attn_output.weight | Block 7 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 542 |
+
| 70 | blk.7.attn_q.weight | Block 7 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 543 |
+
| 71 | blk.7.attn_v.weight | Block 7 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 544 |
+
| 72 | blk.7.ffn_down.weight | Block 7 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 545 |
+
| 73 | blk.7.ffn_gate.weight | Block 7 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 546 |
+
| 74 | blk.7.ffn_norm.weight | Block 7 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 547 |
+
| 75 | blk.7.ffn_up.weight | Block 7 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 548 |
+
|
| 549 |
+
- Total elements in blk.7: (~218M) 218112000
|
| 550 |
+
- Percentage of total elements: 2.72%
|
| 551 |
+
|
| 552 |
+
|
| 553 |
+
### <a name="blk_8">Block 8 Tensor Group : ~218M Elements</a>
|
| 554 |
+
|
| 555 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 556 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 557 |
+
| 76 | blk.8.attn_k.weight | Block 8 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 558 |
+
| 77 | blk.8.attn_norm.weight | Block 8 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 559 |
+
| 78 | blk.8.attn_output.weight | Block 8 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 560 |
+
| 79 | blk.8.attn_q.weight | Block 8 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 561 |
+
| 80 | blk.8.attn_v.weight | Block 8 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 562 |
+
| 81 | blk.8.ffn_down.weight | Block 8 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 563 |
+
| 82 | blk.8.ffn_gate.weight | Block 8 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 564 |
+
| 83 | blk.8.ffn_norm.weight | Block 8 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 565 |
+
| 84 | blk.8.ffn_up.weight | Block 8 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 566 |
+
|
| 567 |
+
- Total elements in blk.8: (~218M) 218112000
|
| 568 |
+
- Percentage of total elements: 2.72%
|
| 569 |
+
|
| 570 |
+
|
| 571 |
+
### <a name="blk_9">Block 9 Tensor Group : ~218M Elements</a>
|
| 572 |
+
|
| 573 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 574 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 575 |
+
| 85 | blk.9.attn_k.weight | Block 9 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 576 |
+
| 86 | blk.9.attn_norm.weight | Block 9 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 577 |
+
| 87 | blk.9.attn_output.weight | Block 9 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 578 |
+
| 88 | blk.9.attn_q.weight | Block 9 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 579 |
+
| 89 | blk.9.attn_v.weight | Block 9 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 580 |
+
| 90 | blk.9.ffn_down.weight | Block 9 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 581 |
+
| 91 | blk.9.ffn_gate.weight | Block 9 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 582 |
+
| 92 | blk.9.ffn_norm.weight | Block 9 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 583 |
+
| 93 | blk.9.ffn_up.weight | Block 9 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 584 |
+
|
| 585 |
+
- Total elements in blk.9: (~218M) 218112000
|
| 586 |
+
- Percentage of total elements: 2.72%
|
| 587 |
+
|
| 588 |
+
|
| 589 |
+
### <a name="blk_10">Block 10 Tensor Group : ~218M Elements</a>
|
| 590 |
+
|
| 591 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 592 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 593 |
+
| 94 | blk.10.attn_k.weight | Block 10 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 594 |
+
| 95 | blk.10.attn_norm.weight | Block 10 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 595 |
+
| 96 | blk.10.attn_output.weight | Block 10 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 596 |
+
| 97 | blk.10.attn_q.weight | Block 10 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 597 |
+
| 98 | blk.10.attn_v.weight | Block 10 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 598 |
+
| 99 | blk.10.ffn_down.weight | Block 10 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 599 |
+
| 100 | blk.10.ffn_gate.weight | Block 10 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 600 |
+
| 101 | blk.10.ffn_norm.weight | Block 10 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 601 |
+
| 102 | blk.10.ffn_up.weight | Block 10 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 602 |
+
|
| 603 |
+
- Total elements in blk.10: (~218M) 218112000
|
| 604 |
+
- Percentage of total elements: 2.72%
|
| 605 |
+
|
| 606 |
+
|
| 607 |
+
### <a name="blk_11">Block 11 Tensor Group : ~218M Elements</a>
|
| 608 |
+
|
| 609 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 610 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 611 |
+
| 103 | blk.11.attn_k.weight | Block 11 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 612 |
+
| 104 | blk.11.attn_norm.weight | Block 11 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 613 |
+
| 105 | blk.11.attn_output.weight | Block 11 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 614 |
+
| 106 | blk.11.attn_q.weight | Block 11 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 615 |
+
| 107 | blk.11.attn_v.weight | Block 11 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 616 |
+
| 108 | blk.11.ffn_down.weight | Block 11 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 617 |
+
| 109 | blk.11.ffn_gate.weight | Block 11 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 618 |
+
| 110 | blk.11.ffn_norm.weight | Block 11 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 619 |
+
| 111 | blk.11.ffn_up.weight | Block 11 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 620 |
+
|
| 621 |
+
- Total elements in blk.11: (~218M) 218112000
|
| 622 |
+
- Percentage of total elements: 2.72%
|
| 623 |
+
|
| 624 |
+
|
| 625 |
+
### <a name="blk_12">Block 12 Tensor Group : ~218M Elements</a>
|
| 626 |
+
|
| 627 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 628 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 629 |
+
| 112 | blk.12.attn_k.weight | Block 12 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 630 |
+
| 113 | blk.12.attn_norm.weight | Block 12 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 631 |
+
| 114 | blk.12.attn_output.weight | Block 12 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 632 |
+
| 115 | blk.12.attn_q.weight | Block 12 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 633 |
+
| 116 | blk.12.attn_v.weight | Block 12 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 634 |
+
| 117 | blk.12.ffn_down.weight | Block 12 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 635 |
+
| 118 | blk.12.ffn_gate.weight | Block 12 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 636 |
+
| 119 | blk.12.ffn_norm.weight | Block 12 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 637 |
+
| 120 | blk.12.ffn_up.weight | Block 12 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 638 |
+
|
| 639 |
+
- Total elements in blk.12: (~218M) 218112000
|
| 640 |
+
- Percentage of total elements: 2.72%
|
| 641 |
+
|
| 642 |
+
|
| 643 |
+
### <a name="blk_13">Block 13 Tensor Group : ~218M Elements</a>
|
| 644 |
+
|
| 645 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 646 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 647 |
+
| 121 | blk.13.attn_k.weight | Block 13 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 648 |
+
| 122 | blk.13.attn_norm.weight | Block 13 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 649 |
+
| 123 | blk.13.attn_output.weight | Block 13 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 650 |
+
| 124 | blk.13.attn_q.weight | Block 13 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 651 |
+
| 125 | blk.13.attn_v.weight | Block 13 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 652 |
+
| 126 | blk.13.ffn_down.weight | Block 13 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 653 |
+
| 127 | blk.13.ffn_gate.weight | Block 13 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 654 |
+
| 128 | blk.13.ffn_norm.weight | Block 13 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 655 |
+
| 129 | blk.13.ffn_up.weight | Block 13 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 656 |
+
|
| 657 |
+
- Total elements in blk.13: (~218M) 218112000
|
| 658 |
+
- Percentage of total elements: 2.72%
|
| 659 |
+
|
| 660 |
+
|
| 661 |
+
### <a name="blk_14">Block 14 Tensor Group : ~218M Elements</a>
|
| 662 |
+
|
| 663 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 664 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 665 |
+
| 130 | blk.14.attn_k.weight | Block 14 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 666 |
+
| 131 | blk.14.attn_norm.weight | Block 14 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 667 |
+
| 132 | blk.14.attn_output.weight | Block 14 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 668 |
+
| 133 | blk.14.attn_q.weight | Block 14 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 669 |
+
| 134 | blk.14.attn_v.weight | Block 14 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 670 |
+
| 135 | blk.14.ffn_down.weight | Block 14 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 671 |
+
| 136 | blk.14.ffn_gate.weight | Block 14 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 672 |
+
| 137 | blk.14.ffn_norm.weight | Block 14 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 673 |
+
| 138 | blk.14.ffn_up.weight | Block 14 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 674 |
+
|
| 675 |
+
- Total elements in blk.14: (~218M) 218112000
|
| 676 |
+
- Percentage of total elements: 2.72%
|
| 677 |
+
|
| 678 |
+
|
| 679 |
+
### <a name="blk_15">Block 15 Tensor Group : ~218M Elements</a>
|
| 680 |
+
|
| 681 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 682 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 683 |
+
| 139 | blk.15.attn_k.weight | Block 15 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 684 |
+
| 140 | blk.15.attn_norm.weight | Block 15 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 685 |
+
| 141 | blk.15.attn_output.weight | Block 15 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 686 |
+
| 142 | blk.15.attn_q.weight | Block 15 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 687 |
+
| 143 | blk.15.attn_v.weight | Block 15 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 688 |
+
| 144 | blk.15.ffn_down.weight | Block 15 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 689 |
+
| 145 | blk.15.ffn_gate.weight | Block 15 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 690 |
+
| 146 | blk.15.ffn_norm.weight | Block 15 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 691 |
+
| 147 | blk.15.ffn_up.weight | Block 15 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q5_K |
|
| 692 |
+
|
| 693 |
+
- Total elements in blk.15: (~218M) 218112000
|
| 694 |
+
- Percentage of total elements: 2.72%
|
| 695 |
+
|
| 696 |
+
|
| 697 |
+
### <a name="blk_16">Block 16 Tensor Group : ~218M Elements</a>
|
| 698 |
+
|
| 699 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 700 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 701 |
+
| 148 | blk.16.attn_k.weight | Block 16 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 702 |
+
| 149 | blk.16.attn_norm.weight | Block 16 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 703 |
+
| 150 | blk.16.attn_output.weight | Block 16 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 704 |
+
| 151 | blk.16.attn_q.weight | Block 16 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 705 |
+
| 152 | blk.16.attn_v.weight | Block 16 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 706 |
+
| 153 | blk.16.ffn_down.weight | Block 16 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 707 |
+
| 154 | blk.16.ffn_gate.weight | Block 16 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 708 |
+
| 155 | blk.16.ffn_norm.weight | Block 16 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 709 |
+
| 156 | blk.16.ffn_up.weight | Block 16 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 710 |
+
|
| 711 |
+
- Total elements in blk.16: (~218M) 218112000
|
| 712 |
+
- Percentage of total elements: 2.72%
|
| 713 |
+
|
| 714 |
+
|
| 715 |
+
### <a name="blk_17">Block 17 Tensor Group : ~218M Elements</a>
|
| 716 |
+
|
| 717 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 718 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 719 |
+
| 157 | blk.17.attn_k.weight | Block 17 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 720 |
+
| 158 | blk.17.attn_norm.weight | Block 17 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 721 |
+
| 159 | blk.17.attn_output.weight | Block 17 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 722 |
+
| 160 | blk.17.attn_q.weight | Block 17 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 723 |
+
| 161 | blk.17.attn_v.weight | Block 17 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 724 |
+
| 162 | blk.17.ffn_down.weight | Block 17 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 725 |
+
| 163 | blk.17.ffn_gate.weight | Block 17 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 726 |
+
| 164 | blk.17.ffn_norm.weight | Block 17 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 727 |
+
| 165 | blk.17.ffn_up.weight | Block 17 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 728 |
+
|
| 729 |
+
- Total elements in blk.17: (~218M) 218112000
|
| 730 |
+
- Percentage of total elements: 2.72%
|
| 731 |
+
|
| 732 |
+
|
| 733 |
+
### <a name="blk_18">Block 18 Tensor Group : ~218M Elements</a>
|
| 734 |
+
|
| 735 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 736 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 737 |
+
| 166 | blk.18.attn_k.weight | Block 18 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 738 |
+
| 167 | blk.18.attn_norm.weight | Block 18 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 739 |
+
| 168 | blk.18.attn_output.weight | Block 18 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 740 |
+
| 169 | blk.18.attn_q.weight | Block 18 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 741 |
+
| 170 | blk.18.attn_v.weight | Block 18 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 742 |
+
| 171 | blk.18.ffn_down.weight | Block 18 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 743 |
+
| 172 | blk.18.ffn_gate.weight | Block 18 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 744 |
+
| 173 | blk.18.ffn_norm.weight | Block 18 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 745 |
+
| 174 | blk.18.ffn_up.weight | Block 18 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 746 |
+
|
| 747 |
+
- Total elements in blk.18: (~218M) 218112000
|
| 748 |
+
- Percentage of total elements: 2.72%
|
| 749 |
+
|
| 750 |
+
|
| 751 |
+
### <a name="blk_19">Block 19 Tensor Group : ~218M Elements</a>
|
| 752 |
+
|
| 753 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 754 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 755 |
+
| 175 | blk.19.attn_k.weight | Block 19 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 756 |
+
| 176 | blk.19.attn_norm.weight | Block 19 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 757 |
+
| 177 | blk.19.attn_output.weight | Block 19 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 758 |
+
| 178 | blk.19.attn_q.weight | Block 19 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 759 |
+
| 179 | blk.19.attn_v.weight | Block 19 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 760 |
+
| 180 | blk.19.ffn_down.weight | Block 19 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 761 |
+
| 181 | blk.19.ffn_gate.weight | Block 19 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 762 |
+
| 182 | blk.19.ffn_norm.weight | Block 19 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 763 |
+
| 183 | blk.19.ffn_up.weight | Block 19 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 764 |
+
|
| 765 |
+
- Total elements in blk.19: (~218M) 218112000
|
| 766 |
+
- Percentage of total elements: 2.72%
|
| 767 |
+
|
| 768 |
+
|
| 769 |
+
### <a name="blk_20">Block 20 Tensor Group : ~218M Elements</a>
|
| 770 |
+
|
| 771 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 772 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 773 |
+
| 184 | blk.20.attn_k.weight | Block 20 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 774 |
+
| 185 | blk.20.attn_norm.weight | Block 20 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 775 |
+
| 186 | blk.20.attn_output.weight | Block 20 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 776 |
+
| 187 | blk.20.attn_q.weight | Block 20 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 777 |
+
| 188 | blk.20.attn_v.weight | Block 20 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 778 |
+
| 189 | blk.20.ffn_down.weight | Block 20 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 779 |
+
| 190 | blk.20.ffn_gate.weight | Block 20 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 780 |
+
| 191 | blk.20.ffn_norm.weight | Block 20 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 781 |
+
| 192 | blk.20.ffn_up.weight | Block 20 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 782 |
+
|
| 783 |
+
- Total elements in blk.20: (~218M) 218112000
|
| 784 |
+
- Percentage of total elements: 2.72%
|
| 785 |
+
|
| 786 |
+
|
| 787 |
+
### <a name="blk_21">Block 21 Tensor Group : ~218M Elements</a>
|
| 788 |
+
|
| 789 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 790 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 791 |
+
| 193 | blk.21.attn_k.weight | Block 21 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 792 |
+
| 194 | blk.21.attn_norm.weight | Block 21 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 793 |
+
| 195 | blk.21.attn_output.weight | Block 21 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 794 |
+
| 196 | blk.21.attn_q.weight | Block 21 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 795 |
+
| 197 | blk.21.attn_v.weight | Block 21 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 796 |
+
| 198 | blk.21.ffn_down.weight | Block 21 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 797 |
+
| 199 | blk.21.ffn_gate.weight | Block 21 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 798 |
+
| 200 | blk.21.ffn_norm.weight | Block 21 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 799 |
+
| 201 | blk.21.ffn_up.weight | Block 21 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 800 |
+
|
| 801 |
+
- Total elements in blk.21: (~218M) 218112000
|
| 802 |
+
- Percentage of total elements: 2.72%
|
| 803 |
+
|
| 804 |
+
|
| 805 |
+
### <a name="blk_22">Block 22 Tensor Group : ~218M Elements</a>
|
| 806 |
+
|
| 807 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 808 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 809 |
+
| 202 | blk.22.attn_k.weight | Block 22 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 810 |
+
| 203 | blk.22.attn_norm.weight | Block 22 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 811 |
+
| 204 | blk.22.attn_output.weight | Block 22 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 812 |
+
| 205 | blk.22.attn_q.weight | Block 22 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 813 |
+
| 206 | blk.22.attn_v.weight | Block 22 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 814 |
+
| 207 | blk.22.ffn_down.weight | Block 22 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 815 |
+
| 208 | blk.22.ffn_gate.weight | Block 22 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 816 |
+
| 209 | blk.22.ffn_norm.weight | Block 22 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 817 |
+
| 210 | blk.22.ffn_up.weight | Block 22 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 818 |
+
|
| 819 |
+
- Total elements in blk.22: (~218M) 218112000
|
| 820 |
+
- Percentage of total elements: 2.72%
|
| 821 |
+
|
| 822 |
+
|
| 823 |
+
### <a name="blk_23">Block 23 Tensor Group : ~218M Elements</a>
|
| 824 |
+
|
| 825 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 826 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 827 |
+
| 211 | blk.23.attn_k.weight | Block 23 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 828 |
+
| 212 | blk.23.attn_norm.weight | Block 23 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 829 |
+
| 213 | blk.23.attn_output.weight | Block 23 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 830 |
+
| 214 | blk.23.attn_q.weight | Block 23 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 831 |
+
| 215 | blk.23.attn_v.weight | Block 23 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 832 |
+
| 216 | blk.23.ffn_down.weight | Block 23 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 833 |
+
| 217 | blk.23.ffn_gate.weight | Block 23 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 834 |
+
| 218 | blk.23.ffn_norm.weight | Block 23 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 835 |
+
| 219 | blk.23.ffn_up.weight | Block 23 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 836 |
+
|
| 837 |
+
- Total elements in blk.23: (~218M) 218112000
|
| 838 |
+
- Percentage of total elements: 2.72%
|
| 839 |
+
|
| 840 |
+
|
| 841 |
+
### <a name="blk_24">Block 24 Tensor Group : ~218M Elements</a>
|
| 842 |
+
|
| 843 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 844 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 845 |
+
| 220 | blk.24.attn_k.weight | Block 24 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 846 |
+
| 221 | blk.24.attn_norm.weight | Block 24 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 847 |
+
| 222 | blk.24.attn_output.weight | Block 24 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 848 |
+
| 223 | blk.24.attn_q.weight | Block 24 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 849 |
+
| 224 | blk.24.attn_v.weight | Block 24 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 850 |
+
| 225 | blk.24.ffn_down.weight | Block 24 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 851 |
+
| 226 | blk.24.ffn_gate.weight | Block 24 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 852 |
+
| 227 | blk.24.ffn_norm.weight | Block 24 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 853 |
+
| 228 | blk.24.ffn_up.weight | Block 24 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 854 |
+
|
| 855 |
+
- Total elements in blk.24: (~218M) 218112000
|
| 856 |
+
- Percentage of total elements: 2.72%
|
| 857 |
+
|
| 858 |
+
|
| 859 |
+
### <a name="blk_25">Block 25 Tensor Group : ~218M Elements</a>
|
| 860 |
+
|
| 861 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 862 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 863 |
+
| 229 | blk.25.attn_k.weight | Block 25 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 864 |
+
| 230 | blk.25.attn_norm.weight | Block 25 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 865 |
+
| 231 | blk.25.attn_output.weight | Block 25 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 866 |
+
| 232 | blk.25.attn_q.weight | Block 25 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 867 |
+
| 233 | blk.25.attn_v.weight | Block 25 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 868 |
+
| 234 | blk.25.ffn_down.weight | Block 25 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 869 |
+
| 235 | blk.25.ffn_gate.weight | Block 25 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 870 |
+
| 236 | blk.25.ffn_norm.weight | Block 25 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 871 |
+
| 237 | blk.25.ffn_up.weight | Block 25 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 872 |
+
|
| 873 |
+
- Total elements in blk.25: (~218M) 218112000
|
| 874 |
+
- Percentage of total elements: 2.72%
|
| 875 |
+
|
| 876 |
+
|
| 877 |
+
### <a name="blk_26">Block 26 Tensor Group : ~218M Elements</a>
|
| 878 |
+
|
| 879 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 880 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 881 |
+
| 238 | blk.26.attn_k.weight | Block 26 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 882 |
+
| 239 | blk.26.attn_norm.weight | Block 26 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 883 |
+
| 240 | blk.26.attn_output.weight | Block 26 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 884 |
+
| 241 | blk.26.attn_q.weight | Block 26 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 885 |
+
| 242 | blk.26.attn_v.weight | Block 26 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 886 |
+
| 243 | blk.26.ffn_down.weight | Block 26 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 887 |
+
| 244 | blk.26.ffn_gate.weight | Block 26 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 888 |
+
| 245 | blk.26.ffn_norm.weight | Block 26 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 889 |
+
| 246 | blk.26.ffn_up.weight | Block 26 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 890 |
+
|
| 891 |
+
- Total elements in blk.26: (~218M) 218112000
|
| 892 |
+
- Percentage of total elements: 2.72%
|
| 893 |
+
|
| 894 |
+
|
| 895 |
+
### <a name="blk_27">Block 27 Tensor Group : ~218M Elements</a>
|
| 896 |
+
|
| 897 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 898 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 899 |
+
| 247 | blk.27.attn_k.weight | Block 27 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 900 |
+
| 248 | blk.27.attn_norm.weight | Block 27 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 901 |
+
| 249 | blk.27.attn_output.weight | Block 27 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 902 |
+
| 250 | blk.27.attn_q.weight | Block 27 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 903 |
+
| 251 | blk.27.attn_v.weight | Block 27 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 904 |
+
| 252 | blk.27.ffn_down.weight | Block 27 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 905 |
+
| 253 | blk.27.ffn_gate.weight | Block 27 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 906 |
+
| 254 | blk.27.ffn_norm.weight | Block 27 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 907 |
+
| 255 | blk.27.ffn_up.weight | Block 27 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 908 |
+
|
| 909 |
+
- Total elements in blk.27: (~218M) 218112000
|
| 910 |
+
- Percentage of total elements: 2.72%
|
| 911 |
+
|
| 912 |
+
|
| 913 |
+
### <a name="blk_28">Block 28 Tensor Group : ~218M Elements</a>
|
| 914 |
+
|
| 915 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 916 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 917 |
+
| 256 | blk.28.attn_k.weight | Block 28 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 918 |
+
| 257 | blk.28.attn_norm.weight | Block 28 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 919 |
+
| 258 | blk.28.attn_output.weight | Block 28 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 920 |
+
| 259 | blk.28.attn_q.weight | Block 28 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 921 |
+
| 260 | blk.28.attn_v.weight | Block 28 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 922 |
+
| 261 | blk.28.ffn_down.weight | Block 28 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 923 |
+
| 262 | blk.28.ffn_gate.weight | Block 28 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 924 |
+
| 263 | blk.28.ffn_norm.weight | Block 28 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 925 |
+
| 264 | blk.28.ffn_up.weight | Block 28 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 926 |
+
|
| 927 |
+
- Total elements in blk.28: (~218M) 218112000
|
| 928 |
+
- Percentage of total elements: 2.72%
|
| 929 |
+
|
| 930 |
+
|
| 931 |
+
### <a name="blk_29">Block 29 Tensor Group : ~218M Elements</a>
|
| 932 |
+
|
| 933 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 934 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 935 |
+
| 265 | blk.29.attn_k.weight | Block 29 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 936 |
+
| 266 | blk.29.attn_norm.weight | Block 29 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 937 |
+
| 267 | blk.29.attn_output.weight | Block 29 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 938 |
+
| 268 | blk.29.attn_q.weight | Block 29 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 939 |
+
| 269 | blk.29.attn_v.weight | Block 29 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 940 |
+
| 270 | blk.29.ffn_down.weight | Block 29 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 941 |
+
| 271 | blk.29.ffn_gate.weight | Block 29 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 942 |
+
| 272 | blk.29.ffn_norm.weight | Block 29 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 943 |
+
| 273 | blk.29.ffn_up.weight | Block 29 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 944 |
+
|
| 945 |
+
- Total elements in blk.29: (~218M) 218112000
|
| 946 |
+
- Percentage of total elements: 2.72%
|
| 947 |
+
|
| 948 |
+
|
| 949 |
+
### <a name="blk_30">Block 30 Tensor Group : ~218M Elements</a>
|
| 950 |
+
|
| 951 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 952 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 953 |
+
| 274 | blk.30.attn_k.weight | Block 30 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 954 |
+
| 275 | blk.30.attn_norm.weight | Block 30 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 955 |
+
| 276 | blk.30.attn_output.weight | Block 30 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 956 |
+
| 277 | blk.30.attn_q.weight | Block 30 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 957 |
+
| 278 | blk.30.attn_v.weight | Block 30 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 958 |
+
| 279 | blk.30.ffn_down.weight | Block 30 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 959 |
+
| 280 | blk.30.ffn_gate.weight | Block 30 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 960 |
+
| 281 | blk.30.ffn_norm.weight | Block 30 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 961 |
+
| 282 | blk.30.ffn_up.weight | Block 30 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 962 |
+
|
| 963 |
+
- Total elements in blk.30: (~218M) 218112000
|
| 964 |
+
- Percentage of total elements: 2.72%
|
| 965 |
+
|
| 966 |
+
|
| 967 |
+
### <a name="blk_31">Block 31 Tensor Group : ~218M Elements</a>
|
| 968 |
+
|
| 969 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 970 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 971 |
+
| 283 | blk.31.attn_k.weight | Block 31 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q5_K |
|
| 972 |
+
| 284 | blk.31.attn_norm.weight | Block 31 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 973 |
+
| 285 | blk.31.attn_output.weight | Block 31 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 974 |
+
| 286 | blk.31.attn_q.weight | Block 31 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q5_K |
|
| 975 |
+
| 287 | blk.31.attn_v.weight | Block 31 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 976 |
+
| 288 | blk.31.ffn_down.weight | Block 31 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 977 |
+
| 289 | blk.31.ffn_gate.weight | Block 31 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 978 |
+
| 290 | blk.31.ffn_norm.weight | Block 31 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 979 |
+
| 291 | blk.31.ffn_up.weight | Block 31 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 980 |
+
|
| 981 |
+
- Total elements in blk.31: (~218M) 218112000
|
| 982 |
+
- Percentage of total elements: 2.72%
|
scores/Watt-Tool-8B-Q8_0.md
ADDED
|
@@ -0,0 +1,982 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Watt-Tool-8B-Q8_0.gguf - GGUF Internal File Dump
|
| 2 |
+
|
| 3 |
+
- Endian: LITTLE endian
|
| 4 |
+
|
| 5 |
+
## Key Value Metadata Store
|
| 6 |
+
|
| 7 |
+
There are 43 key-value pairs in this file
|
| 8 |
+
|
| 9 |
+
| POS | TYPE | Count | Key | Value |
|
| 10 |
+
|----:|:---------|-------:|:---------------------------------------|:--------------------------------------------------------------------|
|
| 11 |
+
| 1 | UINT32 | 1 | GGUF.version | 3 |
|
| 12 |
+
| 2 | UINT64 | 1 | GGUF.tensor_count | 292 |
|
| 13 |
+
| 3 | UINT64 | 1 | GGUF.kv_count | 40 |
|
| 14 |
+
| 4 | STRING | 1 | general.architecture | `llama` |
|
| 15 |
+
| 5 | STRING | 1 | general.type | `model` |
|
| 16 |
+
| 6 | STRING | 1 | general.name | `Watt Tool 8B GGUF` |
|
| 17 |
+
| 7 | STRING | 1 | general.finetune | `GGUF` |
|
| 18 |
+
| 8 | STRING | 1 | general.basename | `Watt-Tool` |
|
| 19 |
+
| 9 | STRING | 1 | general.size_label | `8B` |
|
| 20 |
+
| 10 | STRING | 1 | general.license | `apache-2.0` |
|
| 21 |
+
| 11 | UINT32 | 1 | general.base_model.count | 1 |
|
| 22 |
+
| 12 | STRING | 1 | general.base_model.0.name | `Llama 3.1 8B Instruct` |
|
| 23 |
+
| 13 | STRING | 1 | general.base_model.0.organization | `Meta Llama` |
|
| 24 |
+
| 14 | STRING | 1 | general.base_model.0.repo_url | `https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct` |
|
| 25 |
+
| 15 | [STRING] | 4 | general.tags | [ `function-calling`, `tool-use`, `llama`, `bfcl` ] |
|
| 26 |
+
| 16 | [STRING] | 1 | general.languages | [ `en` ] |
|
| 27 |
+
| 17 | UINT32 | 1 | llama.block_count | 32 |
|
| 28 |
+
| 18 | UINT32 | 1 | llama.context_length | 131072 |
|
| 29 |
+
| 19 | UINT32 | 1 | llama.embedding_length | 4096 |
|
| 30 |
+
| 20 | UINT32 | 1 | llama.feed_forward_length | 14336 |
|
| 31 |
+
| 21 | UINT32 | 1 | llama.attention.head_count | 32 |
|
| 32 |
+
| 22 | UINT32 | 1 | llama.attention.head_count_kv | 8 |
|
| 33 |
+
| 23 | FLOAT32 | 1 | llama.rope.freq_base | 500000.0 |
|
| 34 |
+
| 24 | FLOAT32 | 1 | llama.attention.layer_norm_rms_epsilon | 1e-05 |
|
| 35 |
+
| 25 | UINT32 | 1 | llama.attention.key_length | 128 |
|
| 36 |
+
| 26 | UINT32 | 1 | llama.attention.value_length | 128 |
|
| 37 |
+
| 27 | UINT32 | 1 | llama.vocab_size | 128256 |
|
| 38 |
+
| 28 | UINT32 | 1 | llama.rope.dimension_count | 128 |
|
| 39 |
+
| 29 | STRING | 1 | tokenizer.ggml.model | `gpt2` |
|
| 40 |
+
| 30 | STRING | 1 | tokenizer.ggml.pre | `llama-bpe` |
|
| 41 |
+
| 31 | [STRING] | 128256 | tokenizer.ggml.tokens | [ `!`, `"`, `#`, `$`, `%`, ... ] |
|
| 42 |
+
| 32 | [INT32] | 128256 | tokenizer.ggml.token_type | [ 1, 1, 1, 1, 1, 1, 1, ... ] |
|
| 43 |
+
| 33 | [STRING] | 280147 | tokenizer.ggml.merges | [ `Ġ Ġ`, `Ġ ĠĠĠ`, `ĠĠ ĠĠ`, `ĠĠĠ Ġ`, `i n`, ... ] |
|
| 44 |
+
| 34 | UINT32 | 1 | tokenizer.ggml.bos_token_id | 128000 |
|
| 45 |
+
| 35 | UINT32 | 1 | tokenizer.ggml.eos_token_id | 128009 |
|
| 46 |
+
| 36 | UINT32 | 1 | tokenizer.ggml.padding_token_id | 128009 |
|
| 47 |
+
| 37 | STRING | 1 | tokenizer.chat_template | `{{ '<|begin_of_text|>' }}{% if`...`d|>' }}{% endif %}{% endfor %}` |
|
| 48 |
+
| 38 | UINT32 | 1 | general.quantization_version | 2 |
|
| 49 |
+
| 39 | UINT32 | 1 | general.file_type | 7 |
|
| 50 |
+
| 40 | STRING | 1 | quantize.imatrix.file | `./imatrix/imatrix-Watt-Tool-8B-small.dat` |
|
| 51 |
+
| 41 | STRING | 1 | quantize.imatrix.dataset | `../../datasets/imatrix/calibration_eur_small.txt` |
|
| 52 |
+
| 42 | INT32 | 1 | quantize.imatrix.entries_count | 225 |
|
| 53 |
+
| 43 | INT32 | 1 | quantize.imatrix.chunks_count | 962 |
|
| 54 |
+
|
| 55 |
+
## Tensors Overview ~8B Elements
|
| 56 |
+
|
| 57 |
+
Total number of elements in all tensors: 8030261312 Elements
|
| 58 |
+
|
| 59 |
+
- [Watt-Tool-8B-Q8\_0.gguf - GGUF Internal File Dump](#watt-tool-8b-q8_0gguf---gguf-internal-file-dump)
|
| 60 |
+
- [Key Value Metadata Store](#key-value-metadata-store)
|
| 61 |
+
- [Tensors Overview ~8B Elements](#tensors-overview-8b-elements)
|
| 62 |
+
- [Tensor Data Offset](#tensor-data-offset)
|
| 63 |
+
- [Base Tensor Group : ~1B Elements](#base-tensor-group--1b-elements)
|
| 64 |
+
- [Block 0 Tensor Group : ~218M Elements](#block-0-tensor-group--218m-elements)
|
| 65 |
+
- [Block 1 Tensor Group : ~218M Elements](#block-1-tensor-group--218m-elements)
|
| 66 |
+
- [Block 2 Tensor Group : ~218M Elements](#block-2-tensor-group--218m-elements)
|
| 67 |
+
- [Block 3 Tensor Group : ~218M Elements](#block-3-tensor-group--218m-elements)
|
| 68 |
+
- [Block 4 Tensor Group : ~218M Elements](#block-4-tensor-group--218m-elements)
|
| 69 |
+
- [Block 5 Tensor Group : ~218M Elements](#block-5-tensor-group--218m-elements)
|
| 70 |
+
- [Block 6 Tensor Group : ~218M Elements](#block-6-tensor-group--218m-elements)
|
| 71 |
+
- [Block 7 Tensor Group : ~218M Elements](#block-7-tensor-group--218m-elements)
|
| 72 |
+
- [Block 8 Tensor Group : ~218M Elements](#block-8-tensor-group--218m-elements)
|
| 73 |
+
- [Block 9 Tensor Group : ~218M Elements](#block-9-tensor-group--218m-elements)
|
| 74 |
+
- [Block 10 Tensor Group : ~218M Elements](#block-10-tensor-group--218m-elements)
|
| 75 |
+
- [Block 11 Tensor Group : ~218M Elements](#block-11-tensor-group--218m-elements)
|
| 76 |
+
- [Block 12 Tensor Group : ~218M Elements](#block-12-tensor-group--218m-elements)
|
| 77 |
+
- [Block 13 Tensor Group : ~218M Elements](#block-13-tensor-group--218m-elements)
|
| 78 |
+
- [Block 14 Tensor Group : ~218M Elements](#block-14-tensor-group--218m-elements)
|
| 79 |
+
- [Block 15 Tensor Group : ~218M Elements](#block-15-tensor-group--218m-elements)
|
| 80 |
+
- [Block 16 Tensor Group : ~218M Elements](#block-16-tensor-group--218m-elements)
|
| 81 |
+
- [Block 17 Tensor Group : ~218M Elements](#block-17-tensor-group--218m-elements)
|
| 82 |
+
- [Block 18 Tensor Group : ~218M Elements](#block-18-tensor-group--218m-elements)
|
| 83 |
+
- [Block 19 Tensor Group : ~218M Elements](#block-19-tensor-group--218m-elements)
|
| 84 |
+
- [Block 20 Tensor Group : ~218M Elements](#block-20-tensor-group--218m-elements)
|
| 85 |
+
- [Block 21 Tensor Group : ~218M Elements](#block-21-tensor-group--218m-elements)
|
| 86 |
+
- [Block 22 Tensor Group : ~218M Elements](#block-22-tensor-group--218m-elements)
|
| 87 |
+
- [Block 23 Tensor Group : ~218M Elements](#block-23-tensor-group--218m-elements)
|
| 88 |
+
- [Block 24 Tensor Group : ~218M Elements](#block-24-tensor-group--218m-elements)
|
| 89 |
+
- [Block 25 Tensor Group : ~218M Elements](#block-25-tensor-group--218m-elements)
|
| 90 |
+
- [Block 26 Tensor Group : ~218M Elements](#block-26-tensor-group--218m-elements)
|
| 91 |
+
- [Block 27 Tensor Group : ~218M Elements](#block-27-tensor-group--218m-elements)
|
| 92 |
+
- [Block 28 Tensor Group : ~218M Elements](#block-28-tensor-group--218m-elements)
|
| 93 |
+
- [Block 29 Tensor Group : ~218M Elements](#block-29-tensor-group--218m-elements)
|
| 94 |
+
- [Block 30 Tensor Group : ~218M Elements](#block-30-tensor-group--218m-elements)
|
| 95 |
+
- [Block 31 Tensor Group : ~218M Elements](#block-31-tensor-group--218m-elements)
|
| 96 |
+
|
| 97 |
+
### Tensor Data Offset
|
| 98 |
+
|
| 99 |
+
This table contains the offset and data segment relative to start of file
|
| 100 |
+
|
| 101 |
+
| T_ID | Tensor Layer Name | Data Offset (B) | Data Size (B) |
|
| 102 |
+
|-----:|:--------------------------|-----------------:|-----------------:|
|
| 103 |
+
| 0 | output.weight | 0x779620 | 0x21450000 |
|
| 104 |
+
| 1 | output_norm.weight | 0x21bc9620 | 0x4000 |
|
| 105 |
+
| 2 | rope_freqs.weight | 0x21bcd620 | 0x100 |
|
| 106 |
+
| 3 | token_embd.weight | 0x21bcd720 | 0xd746000 |
|
| 107 |
+
| 4 | blk.0.attn_k.weight | 0x2f313720 | 0x348000 |
|
| 108 |
+
| 5 | blk.0.attn_norm.weight | 0x2f65b720 | 0x4000 |
|
| 109 |
+
| 6 | blk.0.attn_output.weight | 0x2f65f720 | 0x1100000 |
|
| 110 |
+
| 7 | blk.0.attn_q.weight | 0x3075f720 | 0xd20000 |
|
| 111 |
+
| 8 | blk.0.attn_v.weight | 0x3147f720 | 0x440000 |
|
| 112 |
+
| 9 | blk.0.ffn_down.weight | 0x318bf720 | 0x3b80000 |
|
| 113 |
+
| 10 | blk.0.ffn_gate.weight | 0x3543f720 | 0x2df0000 |
|
| 114 |
+
| 11 | blk.0.ffn_norm.weight | 0x3822f720 | 0x4000 |
|
| 115 |
+
| 12 | blk.0.ffn_up.weight | 0x38233720 | 0x2df0000 |
|
| 116 |
+
| 13 | blk.1.attn_k.weight | 0x3b023720 | 0x348000 |
|
| 117 |
+
| 14 | blk.1.attn_norm.weight | 0x3b36b720 | 0x4000 |
|
| 118 |
+
| 15 | blk.1.attn_output.weight | 0x3b36f720 | 0x1100000 |
|
| 119 |
+
| 16 | blk.1.attn_q.weight | 0x3c46f720 | 0xd20000 |
|
| 120 |
+
| 17 | blk.1.attn_v.weight | 0x3d18f720 | 0x440000 |
|
| 121 |
+
| 18 | blk.1.ffn_down.weight | 0x3d5cf720 | 0x3b80000 |
|
| 122 |
+
| 19 | blk.1.ffn_gate.weight | 0x4114f720 | 0x2df0000 |
|
| 123 |
+
| 20 | blk.1.ffn_norm.weight | 0x43f3f720 | 0x4000 |
|
| 124 |
+
| 21 | blk.1.ffn_up.weight | 0x43f43720 | 0x2df0000 |
|
| 125 |
+
| 22 | blk.2.attn_k.weight | 0x46d33720 | 0x348000 |
|
| 126 |
+
| 23 | blk.2.attn_norm.weight | 0x4707b720 | 0x4000 |
|
| 127 |
+
| 24 | blk.2.attn_output.weight | 0x4707f720 | 0x1100000 |
|
| 128 |
+
| 25 | blk.2.attn_q.weight | 0x4817f720 | 0xd20000 |
|
| 129 |
+
| 26 | blk.2.attn_v.weight | 0x48e9f720 | 0x440000 |
|
| 130 |
+
| 27 | blk.2.ffn_down.weight | 0x492df720 | 0x3b80000 |
|
| 131 |
+
| 28 | blk.2.ffn_gate.weight | 0x4ce5f720 | 0x2df0000 |
|
| 132 |
+
| 29 | blk.2.ffn_norm.weight | 0x4fc4f720 | 0x4000 |
|
| 133 |
+
| 30 | blk.2.ffn_up.weight | 0x4fc53720 | 0x2df0000 |
|
| 134 |
+
| 31 | blk.3.attn_k.weight | 0x52a43720 | 0x348000 |
|
| 135 |
+
| 32 | blk.3.attn_norm.weight | 0x52d8b720 | 0x4000 |
|
| 136 |
+
| 33 | blk.3.attn_output.weight | 0x52d8f720 | 0x1100000 |
|
| 137 |
+
| 34 | blk.3.attn_q.weight | 0x53e8f720 | 0xd20000 |
|
| 138 |
+
| 35 | blk.3.attn_v.weight | 0x54baf720 | 0x440000 |
|
| 139 |
+
| 36 | blk.3.ffn_down.weight | 0x54fef720 | 0x3b80000 |
|
| 140 |
+
| 37 | blk.3.ffn_gate.weight | 0x58b6f720 | 0x2df0000 |
|
| 141 |
+
| 38 | blk.3.ffn_norm.weight | 0x5b95f720 | 0x4000 |
|
| 142 |
+
| 39 | blk.3.ffn_up.weight | 0x5b963720 | 0x2df0000 |
|
| 143 |
+
| 40 | blk.4.attn_k.weight | 0x5e753720 | 0x348000 |
|
| 144 |
+
| 41 | blk.4.attn_norm.weight | 0x5ea9b720 | 0x4000 |
|
| 145 |
+
| 42 | blk.4.attn_output.weight | 0x5ea9f720 | 0x1100000 |
|
| 146 |
+
| 43 | blk.4.attn_q.weight | 0x5fb9f720 | 0xd20000 |
|
| 147 |
+
| 44 | blk.4.attn_v.weight | 0x608bf720 | 0x440000 |
|
| 148 |
+
| 45 | blk.4.ffn_down.weight | 0x60cff720 | 0x3b80000 |
|
| 149 |
+
| 46 | blk.4.ffn_gate.weight | 0x6487f720 | 0x2df0000 |
|
| 150 |
+
| 47 | blk.4.ffn_norm.weight | 0x6766f720 | 0x4000 |
|
| 151 |
+
| 48 | blk.4.ffn_up.weight | 0x67673720 | 0x2df0000 |
|
| 152 |
+
| 49 | blk.5.attn_k.weight | 0x6a463720 | 0x348000 |
|
| 153 |
+
| 50 | blk.5.attn_norm.weight | 0x6a7ab720 | 0x4000 |
|
| 154 |
+
| 51 | blk.5.attn_output.weight | 0x6a7af720 | 0x1100000 |
|
| 155 |
+
| 52 | blk.5.attn_q.weight | 0x6b8af720 | 0xd20000 |
|
| 156 |
+
| 53 | blk.5.attn_v.weight | 0x6c5cf720 | 0x440000 |
|
| 157 |
+
| 54 | blk.5.ffn_down.weight | 0x6ca0f720 | 0x3b80000 |
|
| 158 |
+
| 55 | blk.5.ffn_gate.weight | 0x7058f720 | 0x2df0000 |
|
| 159 |
+
| 56 | blk.5.ffn_norm.weight | 0x7337f720 | 0x4000 |
|
| 160 |
+
| 57 | blk.5.ffn_up.weight | 0x73383720 | 0x2df0000 |
|
| 161 |
+
| 58 | blk.6.attn_k.weight | 0x76173720 | 0x348000 |
|
| 162 |
+
| 59 | blk.6.attn_norm.weight | 0x764bb720 | 0x4000 |
|
| 163 |
+
| 60 | blk.6.attn_output.weight | 0x764bf720 | 0x1100000 |
|
| 164 |
+
| 61 | blk.6.attn_q.weight | 0x775bf720 | 0xd20000 |
|
| 165 |
+
| 62 | blk.6.attn_v.weight | 0x782df720 | 0x440000 |
|
| 166 |
+
| 63 | blk.6.ffn_down.weight | 0x7871f720 | 0x3b80000 |
|
| 167 |
+
| 64 | blk.6.ffn_gate.weight | 0x7c29f720 | 0x2df0000 |
|
| 168 |
+
| 65 | blk.6.ffn_norm.weight | 0x7f08f720 | 0x4000 |
|
| 169 |
+
| 66 | blk.6.ffn_up.weight | 0x7f093720 | 0x2df0000 |
|
| 170 |
+
| 67 | blk.7.attn_k.weight | 0x81e83720 | 0x348000 |
|
| 171 |
+
| 68 | blk.7.attn_norm.weight | 0x821cb720 | 0x4000 |
|
| 172 |
+
| 69 | blk.7.attn_output.weight | 0x821cf720 | 0x1100000 |
|
| 173 |
+
| 70 | blk.7.attn_q.weight | 0x832cf720 | 0xd20000 |
|
| 174 |
+
| 71 | blk.7.attn_v.weight | 0x83fef720 | 0x440000 |
|
| 175 |
+
| 72 | blk.7.ffn_down.weight | 0x8442f720 | 0x3b80000 |
|
| 176 |
+
| 73 | blk.7.ffn_gate.weight | 0x87faf720 | 0x2df0000 |
|
| 177 |
+
| 74 | blk.7.ffn_norm.weight | 0x8ad9f720 | 0x4000 |
|
| 178 |
+
| 75 | blk.7.ffn_up.weight | 0x8ada3720 | 0x2df0000 |
|
| 179 |
+
| 76 | blk.8.attn_k.weight | 0x8db93720 | 0x348000 |
|
| 180 |
+
| 77 | blk.8.attn_norm.weight | 0x8dedb720 | 0x4000 |
|
| 181 |
+
| 78 | blk.8.attn_output.weight | 0x8dedf720 | 0x1100000 |
|
| 182 |
+
| 79 | blk.8.attn_q.weight | 0x8efdf720 | 0xd20000 |
|
| 183 |
+
| 80 | blk.8.attn_v.weight | 0x8fcff720 | 0x440000 |
|
| 184 |
+
| 81 | blk.8.ffn_down.weight | 0x9013f720 | 0x3b80000 |
|
| 185 |
+
| 82 | blk.8.ffn_gate.weight | 0x93cbf720 | 0x2df0000 |
|
| 186 |
+
| 83 | blk.8.ffn_norm.weight | 0x96aaf720 | 0x4000 |
|
| 187 |
+
| 84 | blk.8.ffn_up.weight | 0x96ab3720 | 0x2df0000 |
|
| 188 |
+
| 85 | blk.9.attn_k.weight | 0x998a3720 | 0x348000 |
|
| 189 |
+
| 86 | blk.9.attn_norm.weight | 0x99beb720 | 0x4000 |
|
| 190 |
+
| 87 | blk.9.attn_output.weight | 0x99bef720 | 0x1100000 |
|
| 191 |
+
| 88 | blk.9.attn_q.weight | 0x9acef720 | 0xd20000 |
|
| 192 |
+
| 89 | blk.9.attn_v.weight | 0x9ba0f720 | 0x440000 |
|
| 193 |
+
| 90 | blk.9.ffn_down.weight | 0x9be4f720 | 0x3b80000 |
|
| 194 |
+
| 91 | blk.9.ffn_gate.weight | 0x9f9cf720 | 0x2df0000 |
|
| 195 |
+
| 92 | blk.9.ffn_norm.weight | 0xa27bf720 | 0x4000 |
|
| 196 |
+
| 93 | blk.9.ffn_up.weight | 0xa27c3720 | 0x2df0000 |
|
| 197 |
+
| 94 | blk.10.attn_k.weight | 0xa55b3720 | 0x348000 |
|
| 198 |
+
| 95 | blk.10.attn_norm.weight | 0xa58fb720 | 0x4000 |
|
| 199 |
+
| 96 | blk.10.attn_output.weight | 0xa58ff720 | 0x1100000 |
|
| 200 |
+
| 97 | blk.10.attn_q.weight | 0xa69ff720 | 0xd20000 |
|
| 201 |
+
| 98 | blk.10.attn_v.weight | 0xa771f720 | 0x440000 |
|
| 202 |
+
| 99 | blk.10.ffn_down.weight | 0xa7b5f720 | 0x3b80000 |
|
| 203 |
+
| 100 | blk.10.ffn_gate.weight | 0xab6df720 | 0x2df0000 |
|
| 204 |
+
| 101 | blk.10.ffn_norm.weight | 0xae4cf720 | 0x4000 |
|
| 205 |
+
| 102 | blk.10.ffn_up.weight | 0xae4d3720 | 0x2df0000 |
|
| 206 |
+
| 103 | blk.11.attn_k.weight | 0xb12c3720 | 0x348000 |
|
| 207 |
+
| 104 | blk.11.attn_norm.weight | 0xb160b720 | 0x4000 |
|
| 208 |
+
| 105 | blk.11.attn_output.weight | 0xb160f720 | 0x1100000 |
|
| 209 |
+
| 106 | blk.11.attn_q.weight | 0xb270f720 | 0xd20000 |
|
| 210 |
+
| 107 | blk.11.attn_v.weight | 0xb342f720 | 0x440000 |
|
| 211 |
+
| 108 | blk.11.ffn_down.weight | 0xb386f720 | 0x3b80000 |
|
| 212 |
+
| 109 | blk.11.ffn_gate.weight | 0xb73ef720 | 0x2df0000 |
|
| 213 |
+
| 110 | blk.11.ffn_norm.weight | 0xba1df720 | 0x4000 |
|
| 214 |
+
| 111 | blk.11.ffn_up.weight | 0xba1e3720 | 0x2df0000 |
|
| 215 |
+
| 112 | blk.12.attn_k.weight | 0xbcfd3720 | 0x348000 |
|
| 216 |
+
| 113 | blk.12.attn_norm.weight | 0xbd31b720 | 0x4000 |
|
| 217 |
+
| 114 | blk.12.attn_output.weight | 0xbd31f720 | 0x1100000 |
|
| 218 |
+
| 115 | blk.12.attn_q.weight | 0xbe41f720 | 0xd20000 |
|
| 219 |
+
| 116 | blk.12.attn_v.weight | 0xbf13f720 | 0x440000 |
|
| 220 |
+
| 117 | blk.12.ffn_down.weight | 0xbf57f720 | 0x3b80000 |
|
| 221 |
+
| 118 | blk.12.ffn_gate.weight | 0xc30ff720 | 0x2df0000 |
|
| 222 |
+
| 119 | blk.12.ffn_norm.weight | 0xc5eef720 | 0x4000 |
|
| 223 |
+
| 120 | blk.12.ffn_up.weight | 0xc5ef3720 | 0x2df0000 |
|
| 224 |
+
| 121 | blk.13.attn_k.weight | 0xc8ce3720 | 0x440000 |
|
| 225 |
+
| 122 | blk.13.attn_norm.weight | 0xc9123720 | 0x4000 |
|
| 226 |
+
| 123 | blk.13.attn_output.weight | 0xc9127720 | 0x1100000 |
|
| 227 |
+
| 124 | blk.13.attn_q.weight | 0xca227720 | 0x1100000 |
|
| 228 |
+
| 125 | blk.13.attn_v.weight | 0xcb327720 | 0x800000 |
|
| 229 |
+
| 126 | blk.13.ffn_down.weight | 0xcbb27720 | 0x3b80000 |
|
| 230 |
+
| 127 | blk.13.ffn_gate.weight | 0xcf6a7720 | 0x2df0000 |
|
| 231 |
+
| 128 | blk.13.ffn_norm.weight | 0xd2497720 | 0x4000 |
|
| 232 |
+
| 129 | blk.13.ffn_up.weight | 0xd249b720 | 0x2df0000 |
|
| 233 |
+
| 130 | blk.14.attn_k.weight | 0xd528b720 | 0x440000 |
|
| 234 |
+
| 131 | blk.14.attn_norm.weight | 0xd56cb720 | 0x4000 |
|
| 235 |
+
| 132 | blk.14.attn_output.weight | 0xd56cf720 | 0x1100000 |
|
| 236 |
+
| 133 | blk.14.attn_q.weight | 0xd67cf720 | 0x1100000 |
|
| 237 |
+
| 134 | blk.14.attn_v.weight | 0xd78cf720 | 0x800000 |
|
| 238 |
+
| 135 | blk.14.ffn_down.weight | 0xd80cf720 | 0x3b80000 |
|
| 239 |
+
| 136 | blk.14.ffn_gate.weight | 0xdbc4f720 | 0x2df0000 |
|
| 240 |
+
| 137 | blk.14.ffn_norm.weight | 0xdea3f720 | 0x4000 |
|
| 241 |
+
| 138 | blk.14.ffn_up.weight | 0xdea43720 | 0x2df0000 |
|
| 242 |
+
| 139 | blk.15.attn_k.weight | 0xe1833720 | 0x348000 |
|
| 243 |
+
| 140 | blk.15.attn_norm.weight | 0xe1b7b720 | 0x4000 |
|
| 244 |
+
| 141 | blk.15.attn_output.weight | 0xe1b7f720 | 0x1100000 |
|
| 245 |
+
| 142 | blk.15.attn_q.weight | 0xe2c7f720 | 0xd20000 |
|
| 246 |
+
| 143 | blk.15.attn_v.weight | 0xe399f720 | 0x440000 |
|
| 247 |
+
| 144 | blk.15.ffn_down.weight | 0xe3ddf720 | 0x3b80000 |
|
| 248 |
+
| 145 | blk.15.ffn_gate.weight | 0xe795f720 | 0x2df0000 |
|
| 249 |
+
| 146 | blk.15.ffn_norm.weight | 0xea74f720 | 0x4000 |
|
| 250 |
+
| 147 | blk.15.ffn_up.weight | 0xea753720 | 0x2df0000 |
|
| 251 |
+
| 148 | blk.16.attn_k.weight | 0xed543720 | 0x440000 |
|
| 252 |
+
| 149 | blk.16.attn_norm.weight | 0xed983720 | 0x4000 |
|
| 253 |
+
| 150 | blk.16.attn_output.weight | 0xed987720 | 0x1100000 |
|
| 254 |
+
| 151 | blk.16.attn_q.weight | 0xeea87720 | 0x1100000 |
|
| 255 |
+
| 152 | blk.16.attn_v.weight | 0xefb87720 | 0x800000 |
|
| 256 |
+
| 153 | blk.16.ffn_down.weight | 0xf0387720 | 0x3b80000 |
|
| 257 |
+
| 154 | blk.16.ffn_gate.weight | 0xf3f07720 | 0x3b80000 |
|
| 258 |
+
| 155 | blk.16.ffn_norm.weight | 0xf7a87720 | 0x4000 |
|
| 259 |
+
| 156 | blk.16.ffn_up.weight | 0xf7a8b720 | 0x3b80000 |
|
| 260 |
+
| 157 | blk.17.attn_k.weight | 0xfb60b720 | 0x348000 |
|
| 261 |
+
| 158 | blk.17.attn_norm.weight | 0xfb953720 | 0x4000 |
|
| 262 |
+
| 159 | blk.17.attn_output.weight | 0xfb957720 | 0x1100000 |
|
| 263 |
+
| 160 | blk.17.attn_q.weight | 0xfca57720 | 0xd20000 |
|
| 264 |
+
| 161 | blk.17.attn_v.weight | 0xfd777720 | 0x440000 |
|
| 265 |
+
| 162 | blk.17.ffn_down.weight | 0xfdbb7720 | 0x3b80000 |
|
| 266 |
+
| 163 | blk.17.ffn_gate.weight | 0x101737720 | 0x3b80000 |
|
| 267 |
+
| 164 | blk.17.ffn_norm.weight | 0x1052b7720 | 0x4000 |
|
| 268 |
+
| 165 | blk.17.ffn_up.weight | 0x1052bb720 | 0x3b80000 |
|
| 269 |
+
| 166 | blk.18.attn_k.weight | 0x108e3b720 | 0x440000 |
|
| 270 |
+
| 167 | blk.18.attn_norm.weight | 0x10927b720 | 0x4000 |
|
| 271 |
+
| 168 | blk.18.attn_output.weight | 0x10927f720 | 0x1100000 |
|
| 272 |
+
| 169 | blk.18.attn_q.weight | 0x10a37f720 | 0x1100000 |
|
| 273 |
+
| 170 | blk.18.attn_v.weight | 0x10b47f720 | 0x800000 |
|
| 274 |
+
| 171 | blk.18.ffn_down.weight | 0x10bc7f720 | 0x3b80000 |
|
| 275 |
+
| 172 | blk.18.ffn_gate.weight | 0x10f7ff720 | 0x3b80000 |
|
| 276 |
+
| 173 | blk.18.ffn_norm.weight | 0x11337f720 | 0x4000 |
|
| 277 |
+
| 174 | blk.18.ffn_up.weight | 0x113383720 | 0x3b80000 |
|
| 278 |
+
| 175 | blk.19.attn_k.weight | 0x116f03720 | 0x440000 |
|
| 279 |
+
| 176 | blk.19.attn_norm.weight | 0x117343720 | 0x4000 |
|
| 280 |
+
| 177 | blk.19.attn_output.weight | 0x117347720 | 0x1100000 |
|
| 281 |
+
| 178 | blk.19.attn_q.weight | 0x118447720 | 0x1100000 |
|
| 282 |
+
| 179 | blk.19.attn_v.weight | 0x119547720 | 0x800000 |
|
| 283 |
+
| 180 | blk.19.ffn_down.weight | 0x119d47720 | 0x3b80000 |
|
| 284 |
+
| 181 | blk.19.ffn_gate.weight | 0x11d8c7720 | 0x3b80000 |
|
| 285 |
+
| 182 | blk.19.ffn_norm.weight | 0x121447720 | 0x4000 |
|
| 286 |
+
| 183 | blk.19.ffn_up.weight | 0x12144b720 | 0x3b80000 |
|
| 287 |
+
| 184 | blk.20.attn_k.weight | 0x124fcb720 | 0x440000 |
|
| 288 |
+
| 185 | blk.20.attn_norm.weight | 0x12540b720 | 0x4000 |
|
| 289 |
+
| 186 | blk.20.attn_output.weight | 0x12540f720 | 0x1100000 |
|
| 290 |
+
| 187 | blk.20.attn_q.weight | 0x12650f720 | 0x1100000 |
|
| 291 |
+
| 188 | blk.20.attn_v.weight | 0x12760f720 | 0x800000 |
|
| 292 |
+
| 189 | blk.20.ffn_down.weight | 0x127e0f720 | 0x3b80000 |
|
| 293 |
+
| 190 | blk.20.ffn_gate.weight | 0x12b98f720 | 0x3b80000 |
|
| 294 |
+
| 191 | blk.20.ffn_norm.weight | 0x12f50f720 | 0x4000 |
|
| 295 |
+
| 192 | blk.20.ffn_up.weight | 0x12f513720 | 0x3b80000 |
|
| 296 |
+
| 193 | blk.21.attn_k.weight | 0x133093720 | 0x440000 |
|
| 297 |
+
| 194 | blk.21.attn_norm.weight | 0x1334d3720 | 0x4000 |
|
| 298 |
+
| 195 | blk.21.attn_output.weight | 0x1334d7720 | 0x1100000 |
|
| 299 |
+
| 196 | blk.21.attn_q.weight | 0x1345d7720 | 0x1100000 |
|
| 300 |
+
| 197 | blk.21.attn_v.weight | 0x1356d7720 | 0x800000 |
|
| 301 |
+
| 198 | blk.21.ffn_down.weight | 0x135ed7720 | 0x3b80000 |
|
| 302 |
+
| 199 | blk.21.ffn_gate.weight | 0x139a57720 | 0x3b80000 |
|
| 303 |
+
| 200 | blk.21.ffn_norm.weight | 0x13d5d7720 | 0x4000 |
|
| 304 |
+
| 201 | blk.21.ffn_up.weight | 0x13d5db720 | 0x3b80000 |
|
| 305 |
+
| 202 | blk.22.attn_k.weight | 0x14115b720 | 0x440000 |
|
| 306 |
+
| 203 | blk.22.attn_norm.weight | 0x14159b720 | 0x4000 |
|
| 307 |
+
| 204 | blk.22.attn_output.weight | 0x14159f720 | 0x1100000 |
|
| 308 |
+
| 205 | blk.22.attn_q.weight | 0x14269f720 | 0x1100000 |
|
| 309 |
+
| 206 | blk.22.attn_v.weight | 0x14379f720 | 0x800000 |
|
| 310 |
+
| 207 | blk.22.ffn_down.weight | 0x143f9f720 | 0x3b80000 |
|
| 311 |
+
| 208 | blk.22.ffn_gate.weight | 0x147b1f720 | 0x3b80000 |
|
| 312 |
+
| 209 | blk.22.ffn_norm.weight | 0x14b69f720 | 0x4000 |
|
| 313 |
+
| 210 | blk.22.ffn_up.weight | 0x14b6a3720 | 0x3b80000 |
|
| 314 |
+
| 211 | blk.23.attn_k.weight | 0x14f223720 | 0x440000 |
|
| 315 |
+
| 212 | blk.23.attn_norm.weight | 0x14f663720 | 0x4000 |
|
| 316 |
+
| 213 | blk.23.attn_output.weight | 0x14f667720 | 0x1100000 |
|
| 317 |
+
| 214 | blk.23.attn_q.weight | 0x150767720 | 0x1100000 |
|
| 318 |
+
| 215 | blk.23.attn_v.weight | 0x151867720 | 0x800000 |
|
| 319 |
+
| 216 | blk.23.ffn_down.weight | 0x152067720 | 0x3b80000 |
|
| 320 |
+
| 217 | blk.23.ffn_gate.weight | 0x155be7720 | 0x3b80000 |
|
| 321 |
+
| 218 | blk.23.ffn_norm.weight | 0x159767720 | 0x4000 |
|
| 322 |
+
| 219 | blk.23.ffn_up.weight | 0x15976b720 | 0x3b80000 |
|
| 323 |
+
| 220 | blk.24.attn_k.weight | 0x15d2eb720 | 0x440000 |
|
| 324 |
+
| 221 | blk.24.attn_norm.weight | 0x15d72b720 | 0x4000 |
|
| 325 |
+
| 222 | blk.24.attn_output.weight | 0x15d72f720 | 0x1100000 |
|
| 326 |
+
| 223 | blk.24.attn_q.weight | 0x15e82f720 | 0x1100000 |
|
| 327 |
+
| 224 | blk.24.attn_v.weight | 0x15f92f720 | 0x800000 |
|
| 328 |
+
| 225 | blk.24.ffn_down.weight | 0x16012f720 | 0x3b80000 |
|
| 329 |
+
| 226 | blk.24.ffn_gate.weight | 0x163caf720 | 0x3b80000 |
|
| 330 |
+
| 227 | blk.24.ffn_norm.weight | 0x16782f720 | 0x4000 |
|
| 331 |
+
| 228 | blk.24.ffn_up.weight | 0x167833720 | 0x3b80000 |
|
| 332 |
+
| 229 | blk.25.attn_k.weight | 0x16b3b3720 | 0x440000 |
|
| 333 |
+
| 230 | blk.25.attn_norm.weight | 0x16b7f3720 | 0x4000 |
|
| 334 |
+
| 231 | blk.25.attn_output.weight | 0x16b7f7720 | 0x1100000 |
|
| 335 |
+
| 232 | blk.25.attn_q.weight | 0x16c8f7720 | 0x1100000 |
|
| 336 |
+
| 233 | blk.25.attn_v.weight | 0x16d9f7720 | 0x800000 |
|
| 337 |
+
| 234 | blk.25.ffn_down.weight | 0x16e1f7720 | 0x3b80000 |
|
| 338 |
+
| 235 | blk.25.ffn_gate.weight | 0x171d77720 | 0x3b80000 |
|
| 339 |
+
| 236 | blk.25.ffn_norm.weight | 0x1758f7720 | 0x4000 |
|
| 340 |
+
| 237 | blk.25.ffn_up.weight | 0x1758fb720 | 0x3b80000 |
|
| 341 |
+
| 238 | blk.26.attn_k.weight | 0x17947b720 | 0x440000 |
|
| 342 |
+
| 239 | blk.26.attn_norm.weight | 0x1798bb720 | 0x4000 |
|
| 343 |
+
| 240 | blk.26.attn_output.weight | 0x1798bf720 | 0x1100000 |
|
| 344 |
+
| 241 | blk.26.attn_q.weight | 0x17a9bf720 | 0x1100000 |
|
| 345 |
+
| 242 | blk.26.attn_v.weight | 0x17babf720 | 0x800000 |
|
| 346 |
+
| 243 | blk.26.ffn_down.weight | 0x17c2bf720 | 0x3b80000 |
|
| 347 |
+
| 244 | blk.26.ffn_gate.weight | 0x17fe3f720 | 0x3b80000 |
|
| 348 |
+
| 245 | blk.26.ffn_norm.weight | 0x1839bf720 | 0x4000 |
|
| 349 |
+
| 246 | blk.26.ffn_up.weight | 0x1839c3720 | 0x3b80000 |
|
| 350 |
+
| 247 | blk.27.attn_k.weight | 0x187543720 | 0x440000 |
|
| 351 |
+
| 248 | blk.27.attn_norm.weight | 0x187983720 | 0x4000 |
|
| 352 |
+
| 249 | blk.27.attn_output.weight | 0x187987720 | 0x1100000 |
|
| 353 |
+
| 250 | blk.27.attn_q.weight | 0x188a87720 | 0x1100000 |
|
| 354 |
+
| 251 | blk.27.attn_v.weight | 0x189b87720 | 0x800000 |
|
| 355 |
+
| 252 | blk.27.ffn_down.weight | 0x18a387720 | 0x3b80000 |
|
| 356 |
+
| 253 | blk.27.ffn_gate.weight | 0x18df07720 | 0x3b80000 |
|
| 357 |
+
| 254 | blk.27.ffn_norm.weight | 0x191a87720 | 0x4000 |
|
| 358 |
+
| 255 | blk.27.ffn_up.weight | 0x191a8b720 | 0x3b80000 |
|
| 359 |
+
| 256 | blk.28.attn_k.weight | 0x19560b720 | 0x440000 |
|
| 360 |
+
| 257 | blk.28.attn_norm.weight | 0x195a4b720 | 0x4000 |
|
| 361 |
+
| 258 | blk.28.attn_output.weight | 0x195a4f720 | 0x1100000 |
|
| 362 |
+
| 259 | blk.28.attn_q.weight | 0x196b4f720 | 0x1100000 |
|
| 363 |
+
| 260 | blk.28.attn_v.weight | 0x197c4f720 | 0x800000 |
|
| 364 |
+
| 261 | blk.28.ffn_down.weight | 0x19844f720 | 0x3b80000 |
|
| 365 |
+
| 262 | blk.28.ffn_gate.weight | 0x19bfcf720 | 0x3b80000 |
|
| 366 |
+
| 263 | blk.28.ffn_norm.weight | 0x19fb4f720 | 0x4000 |
|
| 367 |
+
| 264 | blk.28.ffn_up.weight | 0x19fb53720 | 0x3b80000 |
|
| 368 |
+
| 265 | blk.29.attn_k.weight | 0x1a36d3720 | 0x440000 |
|
| 369 |
+
| 266 | blk.29.attn_norm.weight | 0x1a3b13720 | 0x4000 |
|
| 370 |
+
| 267 | blk.29.attn_output.weight | 0x1a3b17720 | 0x1100000 |
|
| 371 |
+
| 268 | blk.29.attn_q.weight | 0x1a4c17720 | 0x1100000 |
|
| 372 |
+
| 269 | blk.29.attn_v.weight | 0x1a5d17720 | 0x800000 |
|
| 373 |
+
| 270 | blk.29.ffn_down.weight | 0x1a6517720 | 0x3b80000 |
|
| 374 |
+
| 271 | blk.29.ffn_gate.weight | 0x1aa097720 | 0x3b80000 |
|
| 375 |
+
| 272 | blk.29.ffn_norm.weight | 0x1adc17720 | 0x4000 |
|
| 376 |
+
| 273 | blk.29.ffn_up.weight | 0x1adc1b720 | 0x3b80000 |
|
| 377 |
+
| 274 | blk.30.attn_k.weight | 0x1b179b720 | 0x440000 |
|
| 378 |
+
| 275 | blk.30.attn_norm.weight | 0x1b1bdb720 | 0x4000 |
|
| 379 |
+
| 276 | blk.30.attn_output.weight | 0x1b1bdf720 | 0x1100000 |
|
| 380 |
+
| 277 | blk.30.attn_q.weight | 0x1b2cdf720 | 0x1100000 |
|
| 381 |
+
| 278 | blk.30.attn_v.weight | 0x1b3ddf720 | 0x800000 |
|
| 382 |
+
| 279 | blk.30.ffn_down.weight | 0x1b45df720 | 0x3b80000 |
|
| 383 |
+
| 280 | blk.30.ffn_gate.weight | 0x1b815f720 | 0x3b80000 |
|
| 384 |
+
| 281 | blk.30.ffn_norm.weight | 0x1bbcdf720 | 0x4000 |
|
| 385 |
+
| 282 | blk.30.ffn_up.weight | 0x1bbce3720 | 0x3b80000 |
|
| 386 |
+
| 283 | blk.31.attn_k.weight | 0x1bf863720 | 0x348000 |
|
| 387 |
+
| 284 | blk.31.attn_norm.weight | 0x1bfbab720 | 0x4000 |
|
| 388 |
+
| 285 | blk.31.attn_output.weight | 0x1bfbaf720 | 0x1100000 |
|
| 389 |
+
| 286 | blk.31.attn_q.weight | 0x1c0caf720 | 0xd20000 |
|
| 390 |
+
| 287 | blk.31.attn_v.weight | 0x1c19cf720 | 0x440000 |
|
| 391 |
+
| 288 | blk.31.ffn_down.weight | 0x1c1e0f720 | 0x3b80000 |
|
| 392 |
+
| 289 | blk.31.ffn_gate.weight | 0x1c598f720 | 0x3b80000 |
|
| 393 |
+
| 290 | blk.31.ffn_norm.weight | 0x1c950f720 | 0x4000 |
|
| 394 |
+
| 291 | blk.31.ffn_up.weight | 0x1c9513720 | 0x3b80000 |
|
| 395 |
+
|
| 396 |
+
### <a name="base">Base Tensor Group : ~1B Elements</a>
|
| 397 |
+
|
| 398 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 399 |
+
|-----:|:-------------------|:---------------------------------|:------------------|:----------------------|:-----|
|
| 400 |
+
| 0 | output.weight | Output (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | Q8_0 |
|
| 401 |
+
| 1 | output_norm.weight | Output Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 402 |
+
| 2 | rope_freqs.weight | Rope_Freqs (W) | ( 64) 64 | 64 x 1 x 1 x 1 | F32 |
|
| 403 |
+
| 3 | token_embd.weight | Token Embedding (W) | (~525M) 525336576 | 4096 x 128256 x 1 x 1 | Q3_K |
|
| 404 |
+
|
| 405 |
+
- Total elements in base: ( ~1B) 1050677312
|
| 406 |
+
- Percentage of total elements: 13.08%
|
| 407 |
+
|
| 408 |
+
|
| 409 |
+
### <a name="blk_0">Block 0 Tensor Group : ~218M Elements</a>
|
| 410 |
+
|
| 411 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 412 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 413 |
+
| 4 | blk.0.attn_k.weight | Block 0 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 414 |
+
| 5 | blk.0.attn_norm.weight | Block 0 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 415 |
+
| 6 | blk.0.attn_output.weight | Block 0 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 416 |
+
| 7 | blk.0.attn_q.weight | Block 0 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 417 |
+
| 8 | blk.0.attn_v.weight | Block 0 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 418 |
+
| 9 | blk.0.ffn_down.weight | Block 0 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 419 |
+
| 10 | blk.0.ffn_gate.weight | Block 0 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 420 |
+
| 11 | blk.0.ffn_norm.weight | Block 0 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 421 |
+
| 12 | blk.0.ffn_up.weight | Block 0 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 422 |
+
|
| 423 |
+
- Total elements in blk.0: (~218M) 218112000
|
| 424 |
+
- Percentage of total elements: 2.72%
|
| 425 |
+
|
| 426 |
+
|
| 427 |
+
### <a name="blk_1">Block 1 Tensor Group : ~218M Elements</a>
|
| 428 |
+
|
| 429 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 430 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 431 |
+
| 13 | blk.1.attn_k.weight | Block 1 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 432 |
+
| 14 | blk.1.attn_norm.weight | Block 1 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 433 |
+
| 15 | blk.1.attn_output.weight | Block 1 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 434 |
+
| 16 | blk.1.attn_q.weight | Block 1 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 435 |
+
| 17 | blk.1.attn_v.weight | Block 1 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 436 |
+
| 18 | blk.1.ffn_down.weight | Block 1 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 437 |
+
| 19 | blk.1.ffn_gate.weight | Block 1 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 438 |
+
| 20 | blk.1.ffn_norm.weight | Block 1 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 439 |
+
| 21 | blk.1.ffn_up.weight | Block 1 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 440 |
+
|
| 441 |
+
- Total elements in blk.1: (~218M) 218112000
|
| 442 |
+
- Percentage of total elements: 2.72%
|
| 443 |
+
|
| 444 |
+
|
| 445 |
+
### <a name="blk_2">Block 2 Tensor Group : ~218M Elements</a>
|
| 446 |
+
|
| 447 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 448 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 449 |
+
| 22 | blk.2.attn_k.weight | Block 2 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 450 |
+
| 23 | blk.2.attn_norm.weight | Block 2 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 451 |
+
| 24 | blk.2.attn_output.weight | Block 2 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 452 |
+
| 25 | blk.2.attn_q.weight | Block 2 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 453 |
+
| 26 | blk.2.attn_v.weight | Block 2 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 454 |
+
| 27 | blk.2.ffn_down.weight | Block 2 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 455 |
+
| 28 | blk.2.ffn_gate.weight | Block 2 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 456 |
+
| 29 | blk.2.ffn_norm.weight | Block 2 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 457 |
+
| 30 | blk.2.ffn_up.weight | Block 2 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 458 |
+
|
| 459 |
+
- Total elements in blk.2: (~218M) 218112000
|
| 460 |
+
- Percentage of total elements: 2.72%
|
| 461 |
+
|
| 462 |
+
|
| 463 |
+
### <a name="blk_3">Block 3 Tensor Group : ~218M Elements</a>
|
| 464 |
+
|
| 465 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 466 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 467 |
+
| 31 | blk.3.attn_k.weight | Block 3 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 468 |
+
| 32 | blk.3.attn_norm.weight | Block 3 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 469 |
+
| 33 | blk.3.attn_output.weight | Block 3 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 470 |
+
| 34 | blk.3.attn_q.weight | Block 3 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 471 |
+
| 35 | blk.3.attn_v.weight | Block 3 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 472 |
+
| 36 | blk.3.ffn_down.weight | Block 3 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 473 |
+
| 37 | blk.3.ffn_gate.weight | Block 3 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 474 |
+
| 38 | blk.3.ffn_norm.weight | Block 3 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 475 |
+
| 39 | blk.3.ffn_up.weight | Block 3 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 476 |
+
|
| 477 |
+
- Total elements in blk.3: (~218M) 218112000
|
| 478 |
+
- Percentage of total elements: 2.72%
|
| 479 |
+
|
| 480 |
+
|
| 481 |
+
### <a name="blk_4">Block 4 Tensor Group : ~218M Elements</a>
|
| 482 |
+
|
| 483 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 484 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 485 |
+
| 40 | blk.4.attn_k.weight | Block 4 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 486 |
+
| 41 | blk.4.attn_norm.weight | Block 4 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 487 |
+
| 42 | blk.4.attn_output.weight | Block 4 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 488 |
+
| 43 | blk.4.attn_q.weight | Block 4 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 489 |
+
| 44 | blk.4.attn_v.weight | Block 4 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 490 |
+
| 45 | blk.4.ffn_down.weight | Block 4 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 491 |
+
| 46 | blk.4.ffn_gate.weight | Block 4 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 492 |
+
| 47 | blk.4.ffn_norm.weight | Block 4 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 493 |
+
| 48 | blk.4.ffn_up.weight | Block 4 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 494 |
+
|
| 495 |
+
- Total elements in blk.4: (~218M) 218112000
|
| 496 |
+
- Percentage of total elements: 2.72%
|
| 497 |
+
|
| 498 |
+
|
| 499 |
+
### <a name="blk_5">Block 5 Tensor Group : ~218M Elements</a>
|
| 500 |
+
|
| 501 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 502 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 503 |
+
| 49 | blk.5.attn_k.weight | Block 5 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 504 |
+
| 50 | blk.5.attn_norm.weight | Block 5 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 505 |
+
| 51 | blk.5.attn_output.weight | Block 5 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 506 |
+
| 52 | blk.5.attn_q.weight | Block 5 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 507 |
+
| 53 | blk.5.attn_v.weight | Block 5 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 508 |
+
| 54 | blk.5.ffn_down.weight | Block 5 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 509 |
+
| 55 | blk.5.ffn_gate.weight | Block 5 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 510 |
+
| 56 | blk.5.ffn_norm.weight | Block 5 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 511 |
+
| 57 | blk.5.ffn_up.weight | Block 5 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 512 |
+
|
| 513 |
+
- Total elements in blk.5: (~218M) 218112000
|
| 514 |
+
- Percentage of total elements: 2.72%
|
| 515 |
+
|
| 516 |
+
|
| 517 |
+
### <a name="blk_6">Block 6 Tensor Group : ~218M Elements</a>
|
| 518 |
+
|
| 519 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 520 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 521 |
+
| 58 | blk.6.attn_k.weight | Block 6 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 522 |
+
| 59 | blk.6.attn_norm.weight | Block 6 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 523 |
+
| 60 | blk.6.attn_output.weight | Block 6 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 524 |
+
| 61 | blk.6.attn_q.weight | Block 6 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 525 |
+
| 62 | blk.6.attn_v.weight | Block 6 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 526 |
+
| 63 | blk.6.ffn_down.weight | Block 6 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 527 |
+
| 64 | blk.6.ffn_gate.weight | Block 6 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 528 |
+
| 65 | blk.6.ffn_norm.weight | Block 6 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 529 |
+
| 66 | blk.6.ffn_up.weight | Block 6 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 530 |
+
|
| 531 |
+
- Total elements in blk.6: (~218M) 218112000
|
| 532 |
+
- Percentage of total elements: 2.72%
|
| 533 |
+
|
| 534 |
+
|
| 535 |
+
### <a name="blk_7">Block 7 Tensor Group : ~218M Elements</a>
|
| 536 |
+
|
| 537 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 538 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 539 |
+
| 67 | blk.7.attn_k.weight | Block 7 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 540 |
+
| 68 | blk.7.attn_norm.weight | Block 7 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 541 |
+
| 69 | blk.7.attn_output.weight | Block 7 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 542 |
+
| 70 | blk.7.attn_q.weight | Block 7 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 543 |
+
| 71 | blk.7.attn_v.weight | Block 7 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 544 |
+
| 72 | blk.7.ffn_down.weight | Block 7 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 545 |
+
| 73 | blk.7.ffn_gate.weight | Block 7 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 546 |
+
| 74 | blk.7.ffn_norm.weight | Block 7 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 547 |
+
| 75 | blk.7.ffn_up.weight | Block 7 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 548 |
+
|
| 549 |
+
- Total elements in blk.7: (~218M) 218112000
|
| 550 |
+
- Percentage of total elements: 2.72%
|
| 551 |
+
|
| 552 |
+
|
| 553 |
+
### <a name="blk_8">Block 8 Tensor Group : ~218M Elements</a>
|
| 554 |
+
|
| 555 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 556 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 557 |
+
| 76 | blk.8.attn_k.weight | Block 8 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 558 |
+
| 77 | blk.8.attn_norm.weight | Block 8 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 559 |
+
| 78 | blk.8.attn_output.weight | Block 8 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 560 |
+
| 79 | blk.8.attn_q.weight | Block 8 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 561 |
+
| 80 | blk.8.attn_v.weight | Block 8 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 562 |
+
| 81 | blk.8.ffn_down.weight | Block 8 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 563 |
+
| 82 | blk.8.ffn_gate.weight | Block 8 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 564 |
+
| 83 | blk.8.ffn_norm.weight | Block 8 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 565 |
+
| 84 | blk.8.ffn_up.weight | Block 8 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 566 |
+
|
| 567 |
+
- Total elements in blk.8: (~218M) 218112000
|
| 568 |
+
- Percentage of total elements: 2.72%
|
| 569 |
+
|
| 570 |
+
|
| 571 |
+
### <a name="blk_9">Block 9 Tensor Group : ~218M Elements</a>
|
| 572 |
+
|
| 573 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 574 |
+
|-----:|:-------------------------|:-----------------------------------------------|:----------------|:----------------------|:-----|
|
| 575 |
+
| 85 | blk.9.attn_k.weight | Block 9 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 576 |
+
| 86 | blk.9.attn_norm.weight | Block 9 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 577 |
+
| 87 | blk.9.attn_output.weight | Block 9 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 578 |
+
| 88 | blk.9.attn_q.weight | Block 9 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 579 |
+
| 89 | blk.9.attn_v.weight | Block 9 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 580 |
+
| 90 | blk.9.ffn_down.weight | Block 9 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 581 |
+
| 91 | blk.9.ffn_gate.weight | Block 9 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 582 |
+
| 92 | blk.9.ffn_norm.weight | Block 9 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 583 |
+
| 93 | blk.9.ffn_up.weight | Block 9 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 584 |
+
|
| 585 |
+
- Total elements in blk.9: (~218M) 218112000
|
| 586 |
+
- Percentage of total elements: 2.72%
|
| 587 |
+
|
| 588 |
+
|
| 589 |
+
### <a name="blk_10">Block 10 Tensor Group : ~218M Elements</a>
|
| 590 |
+
|
| 591 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 592 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 593 |
+
| 94 | blk.10.attn_k.weight | Block 10 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 594 |
+
| 95 | blk.10.attn_norm.weight | Block 10 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 595 |
+
| 96 | blk.10.attn_output.weight | Block 10 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 596 |
+
| 97 | blk.10.attn_q.weight | Block 10 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 597 |
+
| 98 | blk.10.attn_v.weight | Block 10 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 598 |
+
| 99 | blk.10.ffn_down.weight | Block 10 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 599 |
+
| 100 | blk.10.ffn_gate.weight | Block 10 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 600 |
+
| 101 | blk.10.ffn_norm.weight | Block 10 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 601 |
+
| 102 | blk.10.ffn_up.weight | Block 10 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 602 |
+
|
| 603 |
+
- Total elements in blk.10: (~218M) 218112000
|
| 604 |
+
- Percentage of total elements: 2.72%
|
| 605 |
+
|
| 606 |
+
|
| 607 |
+
### <a name="blk_11">Block 11 Tensor Group : ~218M Elements</a>
|
| 608 |
+
|
| 609 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 610 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 611 |
+
| 103 | blk.11.attn_k.weight | Block 11 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 612 |
+
| 104 | blk.11.attn_norm.weight | Block 11 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 613 |
+
| 105 | blk.11.attn_output.weight | Block 11 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 614 |
+
| 106 | blk.11.attn_q.weight | Block 11 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 615 |
+
| 107 | blk.11.attn_v.weight | Block 11 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 616 |
+
| 108 | blk.11.ffn_down.weight | Block 11 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 617 |
+
| 109 | blk.11.ffn_gate.weight | Block 11 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 618 |
+
| 110 | blk.11.ffn_norm.weight | Block 11 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 619 |
+
| 111 | blk.11.ffn_up.weight | Block 11 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 620 |
+
|
| 621 |
+
- Total elements in blk.11: (~218M) 218112000
|
| 622 |
+
- Percentage of total elements: 2.72%
|
| 623 |
+
|
| 624 |
+
|
| 625 |
+
### <a name="blk_12">Block 12 Tensor Group : ~218M Elements</a>
|
| 626 |
+
|
| 627 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 628 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 629 |
+
| 112 | blk.12.attn_k.weight | Block 12 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 630 |
+
| 113 | blk.12.attn_norm.weight | Block 12 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 631 |
+
| 114 | blk.12.attn_output.weight | Block 12 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 632 |
+
| 115 | blk.12.attn_q.weight | Block 12 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 633 |
+
| 116 | blk.12.attn_v.weight | Block 12 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 634 |
+
| 117 | blk.12.ffn_down.weight | Block 12 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 635 |
+
| 118 | blk.12.ffn_gate.weight | Block 12 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 636 |
+
| 119 | blk.12.ffn_norm.weight | Block 12 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 637 |
+
| 120 | blk.12.ffn_up.weight | Block 12 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 638 |
+
|
| 639 |
+
- Total elements in blk.12: (~218M) 218112000
|
| 640 |
+
- Percentage of total elements: 2.72%
|
| 641 |
+
|
| 642 |
+
|
| 643 |
+
### <a name="blk_13">Block 13 Tensor Group : ~218M Elements</a>
|
| 644 |
+
|
| 645 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 646 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 647 |
+
| 121 | blk.13.attn_k.weight | Block 13 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 648 |
+
| 122 | blk.13.attn_norm.weight | Block 13 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 649 |
+
| 123 | blk.13.attn_output.weight | Block 13 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 650 |
+
| 124 | blk.13.attn_q.weight | Block 13 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 651 |
+
| 125 | blk.13.attn_v.weight | Block 13 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 652 |
+
| 126 | blk.13.ffn_down.weight | Block 13 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 653 |
+
| 127 | blk.13.ffn_gate.weight | Block 13 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 654 |
+
| 128 | blk.13.ffn_norm.weight | Block 13 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 655 |
+
| 129 | blk.13.ffn_up.weight | Block 13 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 656 |
+
|
| 657 |
+
- Total elements in blk.13: (~218M) 218112000
|
| 658 |
+
- Percentage of total elements: 2.72%
|
| 659 |
+
|
| 660 |
+
|
| 661 |
+
### <a name="blk_14">Block 14 Tensor Group : ~218M Elements</a>
|
| 662 |
+
|
| 663 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 664 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 665 |
+
| 130 | blk.14.attn_k.weight | Block 14 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 666 |
+
| 131 | blk.14.attn_norm.weight | Block 14 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 667 |
+
| 132 | blk.14.attn_output.weight | Block 14 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 668 |
+
| 133 | blk.14.attn_q.weight | Block 14 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 669 |
+
| 134 | blk.14.attn_v.weight | Block 14 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 670 |
+
| 135 | blk.14.ffn_down.weight | Block 14 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 671 |
+
| 136 | blk.14.ffn_gate.weight | Block 14 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 672 |
+
| 137 | blk.14.ffn_norm.weight | Block 14 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 673 |
+
| 138 | blk.14.ffn_up.weight | Block 14 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 674 |
+
|
| 675 |
+
- Total elements in blk.14: (~218M) 218112000
|
| 676 |
+
- Percentage of total elements: 2.72%
|
| 677 |
+
|
| 678 |
+
|
| 679 |
+
### <a name="blk_15">Block 15 Tensor Group : ~218M Elements</a>
|
| 680 |
+
|
| 681 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 682 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 683 |
+
| 139 | blk.15.attn_k.weight | Block 15 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 684 |
+
| 140 | blk.15.attn_norm.weight | Block 15 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 685 |
+
| 141 | blk.15.attn_output.weight | Block 15 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 686 |
+
| 142 | blk.15.attn_q.weight | Block 15 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 687 |
+
| 143 | blk.15.attn_v.weight | Block 15 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 688 |
+
| 144 | blk.15.ffn_down.weight | Block 15 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 689 |
+
| 145 | blk.15.ffn_gate.weight | Block 15 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 690 |
+
| 146 | blk.15.ffn_norm.weight | Block 15 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 691 |
+
| 147 | blk.15.ffn_up.weight | Block 15 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q6_K |
|
| 692 |
+
|
| 693 |
+
- Total elements in blk.15: (~218M) 218112000
|
| 694 |
+
- Percentage of total elements: 2.72%
|
| 695 |
+
|
| 696 |
+
|
| 697 |
+
### <a name="blk_16">Block 16 Tensor Group : ~218M Elements</a>
|
| 698 |
+
|
| 699 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 700 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 701 |
+
| 148 | blk.16.attn_k.weight | Block 16 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 702 |
+
| 149 | blk.16.attn_norm.weight | Block 16 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 703 |
+
| 150 | blk.16.attn_output.weight | Block 16 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 704 |
+
| 151 | blk.16.attn_q.weight | Block 16 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 705 |
+
| 152 | blk.16.attn_v.weight | Block 16 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 706 |
+
| 153 | blk.16.ffn_down.weight | Block 16 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 707 |
+
| 154 | blk.16.ffn_gate.weight | Block 16 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 708 |
+
| 155 | blk.16.ffn_norm.weight | Block 16 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 709 |
+
| 156 | blk.16.ffn_up.weight | Block 16 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 710 |
+
|
| 711 |
+
- Total elements in blk.16: (~218M) 218112000
|
| 712 |
+
- Percentage of total elements: 2.72%
|
| 713 |
+
|
| 714 |
+
|
| 715 |
+
### <a name="blk_17">Block 17 Tensor Group : ~218M Elements</a>
|
| 716 |
+
|
| 717 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 718 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 719 |
+
| 157 | blk.17.attn_k.weight | Block 17 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 720 |
+
| 158 | blk.17.attn_norm.weight | Block 17 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 721 |
+
| 159 | blk.17.attn_output.weight | Block 17 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 722 |
+
| 160 | blk.17.attn_q.weight | Block 17 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 723 |
+
| 161 | blk.17.attn_v.weight | Block 17 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 724 |
+
| 162 | blk.17.ffn_down.weight | Block 17 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 725 |
+
| 163 | blk.17.ffn_gate.weight | Block 17 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 726 |
+
| 164 | blk.17.ffn_norm.weight | Block 17 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 727 |
+
| 165 | blk.17.ffn_up.weight | Block 17 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 728 |
+
|
| 729 |
+
- Total elements in blk.17: (~218M) 218112000
|
| 730 |
+
- Percentage of total elements: 2.72%
|
| 731 |
+
|
| 732 |
+
|
| 733 |
+
### <a name="blk_18">Block 18 Tensor Group : ~218M Elements</a>
|
| 734 |
+
|
| 735 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 736 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 737 |
+
| 166 | blk.18.attn_k.weight | Block 18 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 738 |
+
| 167 | blk.18.attn_norm.weight | Block 18 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 739 |
+
| 168 | blk.18.attn_output.weight | Block 18 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 740 |
+
| 169 | blk.18.attn_q.weight | Block 18 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 741 |
+
| 170 | blk.18.attn_v.weight | Block 18 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 742 |
+
| 171 | blk.18.ffn_down.weight | Block 18 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 743 |
+
| 172 | blk.18.ffn_gate.weight | Block 18 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 744 |
+
| 173 | blk.18.ffn_norm.weight | Block 18 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 745 |
+
| 174 | blk.18.ffn_up.weight | Block 18 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 746 |
+
|
| 747 |
+
- Total elements in blk.18: (~218M) 218112000
|
| 748 |
+
- Percentage of total elements: 2.72%
|
| 749 |
+
|
| 750 |
+
|
| 751 |
+
### <a name="blk_19">Block 19 Tensor Group : ~218M Elements</a>
|
| 752 |
+
|
| 753 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 754 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 755 |
+
| 175 | blk.19.attn_k.weight | Block 19 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 756 |
+
| 176 | blk.19.attn_norm.weight | Block 19 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 757 |
+
| 177 | blk.19.attn_output.weight | Block 19 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 758 |
+
| 178 | blk.19.attn_q.weight | Block 19 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 759 |
+
| 179 | blk.19.attn_v.weight | Block 19 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 760 |
+
| 180 | blk.19.ffn_down.weight | Block 19 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 761 |
+
| 181 | blk.19.ffn_gate.weight | Block 19 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 762 |
+
| 182 | blk.19.ffn_norm.weight | Block 19 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 763 |
+
| 183 | blk.19.ffn_up.weight | Block 19 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 764 |
+
|
| 765 |
+
- Total elements in blk.19: (~218M) 218112000
|
| 766 |
+
- Percentage of total elements: 2.72%
|
| 767 |
+
|
| 768 |
+
|
| 769 |
+
### <a name="blk_20">Block 20 Tensor Group : ~218M Elements</a>
|
| 770 |
+
|
| 771 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 772 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 773 |
+
| 184 | blk.20.attn_k.weight | Block 20 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 774 |
+
| 185 | blk.20.attn_norm.weight | Block 20 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 775 |
+
| 186 | blk.20.attn_output.weight | Block 20 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 776 |
+
| 187 | blk.20.attn_q.weight | Block 20 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 777 |
+
| 188 | blk.20.attn_v.weight | Block 20 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 778 |
+
| 189 | blk.20.ffn_down.weight | Block 20 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 779 |
+
| 190 | blk.20.ffn_gate.weight | Block 20 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 780 |
+
| 191 | blk.20.ffn_norm.weight | Block 20 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 781 |
+
| 192 | blk.20.ffn_up.weight | Block 20 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 782 |
+
|
| 783 |
+
- Total elements in blk.20: (~218M) 218112000
|
| 784 |
+
- Percentage of total elements: 2.72%
|
| 785 |
+
|
| 786 |
+
|
| 787 |
+
### <a name="blk_21">Block 21 Tensor Group : ~218M Elements</a>
|
| 788 |
+
|
| 789 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 790 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 791 |
+
| 193 | blk.21.attn_k.weight | Block 21 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 792 |
+
| 194 | blk.21.attn_norm.weight | Block 21 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 793 |
+
| 195 | blk.21.attn_output.weight | Block 21 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 794 |
+
| 196 | blk.21.attn_q.weight | Block 21 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 795 |
+
| 197 | blk.21.attn_v.weight | Block 21 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 796 |
+
| 198 | blk.21.ffn_down.weight | Block 21 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 797 |
+
| 199 | blk.21.ffn_gate.weight | Block 21 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 798 |
+
| 200 | blk.21.ffn_norm.weight | Block 21 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 799 |
+
| 201 | blk.21.ffn_up.weight | Block 21 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 800 |
+
|
| 801 |
+
- Total elements in blk.21: (~218M) 218112000
|
| 802 |
+
- Percentage of total elements: 2.72%
|
| 803 |
+
|
| 804 |
+
|
| 805 |
+
### <a name="blk_22">Block 22 Tensor Group : ~218M Elements</a>
|
| 806 |
+
|
| 807 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 808 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 809 |
+
| 202 | blk.22.attn_k.weight | Block 22 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 810 |
+
| 203 | blk.22.attn_norm.weight | Block 22 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 811 |
+
| 204 | blk.22.attn_output.weight | Block 22 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 812 |
+
| 205 | blk.22.attn_q.weight | Block 22 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 813 |
+
| 206 | blk.22.attn_v.weight | Block 22 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 814 |
+
| 207 | blk.22.ffn_down.weight | Block 22 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 815 |
+
| 208 | blk.22.ffn_gate.weight | Block 22 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 816 |
+
| 209 | blk.22.ffn_norm.weight | Block 22 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 817 |
+
| 210 | blk.22.ffn_up.weight | Block 22 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 818 |
+
|
| 819 |
+
- Total elements in blk.22: (~218M) 218112000
|
| 820 |
+
- Percentage of total elements: 2.72%
|
| 821 |
+
|
| 822 |
+
|
| 823 |
+
### <a name="blk_23">Block 23 Tensor Group : ~218M Elements</a>
|
| 824 |
+
|
| 825 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 826 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 827 |
+
| 211 | blk.23.attn_k.weight | Block 23 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 828 |
+
| 212 | blk.23.attn_norm.weight | Block 23 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 829 |
+
| 213 | blk.23.attn_output.weight | Block 23 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 830 |
+
| 214 | blk.23.attn_q.weight | Block 23 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 831 |
+
| 215 | blk.23.attn_v.weight | Block 23 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 832 |
+
| 216 | blk.23.ffn_down.weight | Block 23 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 833 |
+
| 217 | blk.23.ffn_gate.weight | Block 23 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 834 |
+
| 218 | blk.23.ffn_norm.weight | Block 23 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 835 |
+
| 219 | blk.23.ffn_up.weight | Block 23 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 836 |
+
|
| 837 |
+
- Total elements in blk.23: (~218M) 218112000
|
| 838 |
+
- Percentage of total elements: 2.72%
|
| 839 |
+
|
| 840 |
+
|
| 841 |
+
### <a name="blk_24">Block 24 Tensor Group : ~218M Elements</a>
|
| 842 |
+
|
| 843 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 844 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 845 |
+
| 220 | blk.24.attn_k.weight | Block 24 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 846 |
+
| 221 | blk.24.attn_norm.weight | Block 24 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 847 |
+
| 222 | blk.24.attn_output.weight | Block 24 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 848 |
+
| 223 | blk.24.attn_q.weight | Block 24 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 849 |
+
| 224 | blk.24.attn_v.weight | Block 24 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 850 |
+
| 225 | blk.24.ffn_down.weight | Block 24 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 851 |
+
| 226 | blk.24.ffn_gate.weight | Block 24 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 852 |
+
| 227 | blk.24.ffn_norm.weight | Block 24 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 853 |
+
| 228 | blk.24.ffn_up.weight | Block 24 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 854 |
+
|
| 855 |
+
- Total elements in blk.24: (~218M) 218112000
|
| 856 |
+
- Percentage of total elements: 2.72%
|
| 857 |
+
|
| 858 |
+
|
| 859 |
+
### <a name="blk_25">Block 25 Tensor Group : ~218M Elements</a>
|
| 860 |
+
|
| 861 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 862 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 863 |
+
| 229 | blk.25.attn_k.weight | Block 25 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 864 |
+
| 230 | blk.25.attn_norm.weight | Block 25 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 865 |
+
| 231 | blk.25.attn_output.weight | Block 25 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 866 |
+
| 232 | blk.25.attn_q.weight | Block 25 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 867 |
+
| 233 | blk.25.attn_v.weight | Block 25 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 868 |
+
| 234 | blk.25.ffn_down.weight | Block 25 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 869 |
+
| 235 | blk.25.ffn_gate.weight | Block 25 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 870 |
+
| 236 | blk.25.ffn_norm.weight | Block 25 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 871 |
+
| 237 | blk.25.ffn_up.weight | Block 25 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 872 |
+
|
| 873 |
+
- Total elements in blk.25: (~218M) 218112000
|
| 874 |
+
- Percentage of total elements: 2.72%
|
| 875 |
+
|
| 876 |
+
|
| 877 |
+
### <a name="blk_26">Block 26 Tensor Group : ~218M Elements</a>
|
| 878 |
+
|
| 879 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 880 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 881 |
+
| 238 | blk.26.attn_k.weight | Block 26 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 882 |
+
| 239 | blk.26.attn_norm.weight | Block 26 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 883 |
+
| 240 | blk.26.attn_output.weight | Block 26 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 884 |
+
| 241 | blk.26.attn_q.weight | Block 26 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 885 |
+
| 242 | blk.26.attn_v.weight | Block 26 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 886 |
+
| 243 | blk.26.ffn_down.weight | Block 26 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 887 |
+
| 244 | blk.26.ffn_gate.weight | Block 26 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 888 |
+
| 245 | blk.26.ffn_norm.weight | Block 26 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 889 |
+
| 246 | blk.26.ffn_up.weight | Block 26 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 890 |
+
|
| 891 |
+
- Total elements in blk.26: (~218M) 218112000
|
| 892 |
+
- Percentage of total elements: 2.72%
|
| 893 |
+
|
| 894 |
+
|
| 895 |
+
### <a name="blk_27">Block 27 Tensor Group : ~218M Elements</a>
|
| 896 |
+
|
| 897 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 898 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 899 |
+
| 247 | blk.27.attn_k.weight | Block 27 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 900 |
+
| 248 | blk.27.attn_norm.weight | Block 27 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 901 |
+
| 249 | blk.27.attn_output.weight | Block 27 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 902 |
+
| 250 | blk.27.attn_q.weight | Block 27 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 903 |
+
| 251 | blk.27.attn_v.weight | Block 27 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 904 |
+
| 252 | blk.27.ffn_down.weight | Block 27 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 905 |
+
| 253 | blk.27.ffn_gate.weight | Block 27 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 906 |
+
| 254 | blk.27.ffn_norm.weight | Block 27 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 907 |
+
| 255 | blk.27.ffn_up.weight | Block 27 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 908 |
+
|
| 909 |
+
- Total elements in blk.27: (~218M) 218112000
|
| 910 |
+
- Percentage of total elements: 2.72%
|
| 911 |
+
|
| 912 |
+
|
| 913 |
+
### <a name="blk_28">Block 28 Tensor Group : ~218M Elements</a>
|
| 914 |
+
|
| 915 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 916 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 917 |
+
| 256 | blk.28.attn_k.weight | Block 28 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 918 |
+
| 257 | blk.28.attn_norm.weight | Block 28 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 919 |
+
| 258 | blk.28.attn_output.weight | Block 28 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 920 |
+
| 259 | blk.28.attn_q.weight | Block 28 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 921 |
+
| 260 | blk.28.attn_v.weight | Block 28 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 922 |
+
| 261 | blk.28.ffn_down.weight | Block 28 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 923 |
+
| 262 | blk.28.ffn_gate.weight | Block 28 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 924 |
+
| 263 | blk.28.ffn_norm.weight | Block 28 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 925 |
+
| 264 | blk.28.ffn_up.weight | Block 28 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 926 |
+
|
| 927 |
+
- Total elements in blk.28: (~218M) 218112000
|
| 928 |
+
- Percentage of total elements: 2.72%
|
| 929 |
+
|
| 930 |
+
|
| 931 |
+
### <a name="blk_29">Block 29 Tensor Group : ~218M Elements</a>
|
| 932 |
+
|
| 933 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 934 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 935 |
+
| 265 | blk.29.attn_k.weight | Block 29 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 936 |
+
| 266 | blk.29.attn_norm.weight | Block 29 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 937 |
+
| 267 | blk.29.attn_output.weight | Block 29 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 938 |
+
| 268 | blk.29.attn_q.weight | Block 29 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 939 |
+
| 269 | blk.29.attn_v.weight | Block 29 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 940 |
+
| 270 | blk.29.ffn_down.weight | Block 29 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 941 |
+
| 271 | blk.29.ffn_gate.weight | Block 29 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 942 |
+
| 272 | blk.29.ffn_norm.weight | Block 29 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 943 |
+
| 273 | blk.29.ffn_up.weight | Block 29 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 944 |
+
|
| 945 |
+
- Total elements in blk.29: (~218M) 218112000
|
| 946 |
+
- Percentage of total elements: 2.72%
|
| 947 |
+
|
| 948 |
+
|
| 949 |
+
### <a name="blk_30">Block 30 Tensor Group : ~218M Elements</a>
|
| 950 |
+
|
| 951 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 952 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 953 |
+
| 274 | blk.30.attn_k.weight | Block 30 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 954 |
+
| 275 | blk.30.attn_norm.weight | Block 30 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 955 |
+
| 276 | blk.30.attn_output.weight | Block 30 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 956 |
+
| 277 | blk.30.attn_q.weight | Block 30 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 957 |
+
| 278 | blk.30.attn_v.weight | Block 30 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | F16 |
|
| 958 |
+
| 279 | blk.30.ffn_down.weight | Block 30 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 959 |
+
| 280 | blk.30.ffn_gate.weight | Block 30 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 960 |
+
| 281 | blk.30.ffn_norm.weight | Block 30 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 961 |
+
| 282 | blk.30.ffn_up.weight | Block 30 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 962 |
+
|
| 963 |
+
- Total elements in blk.30: (~218M) 218112000
|
| 964 |
+
- Percentage of total elements: 2.72%
|
| 965 |
+
|
| 966 |
+
|
| 967 |
+
### <a name="blk_31">Block 31 Tensor Group : ~218M Elements</a>
|
| 968 |
+
|
| 969 |
+
| T_ID | Tensor Layer Name | Human Friendly Tensor Layer Name | Elements | Shape | Type |
|
| 970 |
+
|-----:|:--------------------------|:------------------------------------------------|:----------------|:----------------------|:-----|
|
| 971 |
+
| 283 | blk.31.attn_k.weight | Block 31 Attention Key (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q6_K |
|
| 972 |
+
| 284 | blk.31.attn_norm.weight | Block 31 Attention Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 973 |
+
| 285 | blk.31.attn_output.weight | Block 31 Attention Output (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q8_0 |
|
| 974 |
+
| 286 | blk.31.attn_q.weight | Block 31 Attention Query (W) | (~17M) 16777216 | 4096 x 4096 x 1 x 1 | Q6_K |
|
| 975 |
+
| 287 | blk.31.attn_v.weight | Block 31 Attention Value (W) | ( ~4M) 4194304 | 4096 x 1024 x 1 x 1 | Q8_0 |
|
| 976 |
+
| 288 | blk.31.ffn_down.weight | Block 31 Feed-Forward Network "Down" (W) | (~59M) 58720256 | 14336 x 4096 x 1 x 1 | Q8_0 |
|
| 977 |
+
| 289 | blk.31.ffn_gate.weight | Block 31 Feed-Forward Network "Gate" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 978 |
+
| 290 | blk.31.ffn_norm.weight | Block 31 Feed-Forward Network Normalization (W) | ( ~4K) 4096 | 4096 x 1 x 1 x 1 | F32 |
|
| 979 |
+
| 291 | blk.31.ffn_up.weight | Block 31 Feed-Forward Network "Up" (W) | (~59M) 58720256 | 4096 x 14336 x 1 x 1 | Q8_0 |
|
| 980 |
+
|
| 981 |
+
- Total elements in blk.31: (~218M) 218112000
|
| 982 |
+
- Percentage of total elements: 2.72%
|