File size: 1,457 Bytes
1e3736f 8867025 1e3736f 8867025 1e3736f 8867025 1e3736f 8867025 1e3736f 8867025 1e3736f 8867025 1e3736f 8867025 1e3736f 8867025 1e3736f 8867025 1e3736f 8867025 1e3736f 8867025 1e3736f 8867025 1e3736f 8867025 1e3736f 8867025 5c3f326 8867025 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 |
{
"architectures": [
"EagleSpeculator"
],
"auto_map": {
"": "eagle.EagleSpeculatorConfig"
},
"fusion_bias": false,
"has_no_defaults_at_init": false,
"layernorms": false,
"speculators_config": {
"algorithm": "eagle",
"default_proposal_method": "greedy",
"proposal_methods": [
{
"accept_tolerance": 0.0,
"proposal_type": "greedy",
"speculative_tokens": 5,
"verifier_accept_k": 1
}
],
"verifier": {
"architectures": [
"LlamaForCausalLM"
],
"name_or_path": "meta-llama/Meta-Llama-3.1-8B-Instruct"
}
},
"speculators_model_type": "eagle",
"speculators_version": "0.1.0.dev13",
"torch_dtype": "float32",
"transformer_layer_architecture": "LlamaDecoderLayer",
"transformer_layer_config": {
"attention_bias": false,
"attention_dropout": 0.0,
"bos_token_id": 128000,
"eos_token_id": 128001,
"head_dim": 128,
"hidden_act": "silu",
"hidden_size": 4096,
"initializer_range": 0.02,
"intermediate_size": 14336,
"max_position_embeddings": 2048,
"mlp_bias": false,
"model_type": "llama",
"num_attention_heads": 32,
"num_hidden_layers": 1,
"num_key_value_heads": 8,
"pad_token_id": 0,
"pretraining_tp": 1,
"rms_norm_eps": 1e-05,
"rope_scaling": null,
"rope_theta": 500000.0,
"use_cache": true,
"vocab_size": 128256
},
"transformers_version": "4.52.4"
}
|