Spaces:
Paused
Paused
Update model_card_template.md
Browse files- model_card_template.md +9 -8
model_card_template.md
CHANGED
|
@@ -5,7 +5,7 @@ tags:
|
|
| 5 |
- onnx
|
| 6 |
---
|
| 7 |
|
| 8 |
-
# AMOP-Optimized
|
| 9 |
|
| 10 |
This model was automatically optimized for CPU inference using the **Adaptive Model Optimization Pipeline (AMOP)**.
|
| 11 |
|
|
@@ -14,13 +14,9 @@ This model was automatically optimized for CPU inference using the **Adaptive Mo
|
|
| 14 |
|
| 15 |
## Optimization Details
|
| 16 |
|
| 17 |
-
The following AMOP stages were applied:
|
| 18 |
-
- **
|
| 19 |
-
- **
|
| 20 |
-
|
| 21 |
-
## Performance Metrics
|
| 22 |
-
|
| 23 |
-
{eval_report}
|
| 24 |
|
| 25 |
## How to Use
|
| 26 |
|
|
@@ -40,5 +36,10 @@ gen_tokens = model.generate(**inputs)
|
|
| 40 |
print(tokenizer.batch_decode(gen_tokens))
|
| 41 |
```
|
| 42 |
## AMOP Pipeline Log
|
|
|
|
|
|
|
| 43 |
|
|
|
|
| 44 |
{pipeline_log}
|
|
|
|
|
|
|
|
|
| 5 |
- onnx
|
| 6 |
---
|
| 7 |
|
| 8 |
+
# AMOP-Optimized ONNX Model: {repo_name}
|
| 9 |
|
| 10 |
This model was automatically optimized for CPU inference using the **Adaptive Model Optimization Pipeline (AMOP)**.
|
| 11 |
|
|
|
|
| 14 |
|
| 15 |
## Optimization Details
|
| 16 |
|
| 17 |
+
The following AMOP ONNX pipeline stages were applied:
|
| 18 |
+
- **Pruning:** {pruning_status} (Percentage: {pruning_percent}%)
|
| 19 |
+
- **Quantization & ONNX Conversion:** Enabled ({quant_type} Quantization)
|
|
|
|
|
|
|
|
|
|
|
|
|
| 20 |
|
| 21 |
## How to Use
|
| 22 |
|
|
|
|
| 36 |
print(tokenizer.batch_decode(gen_tokens))
|
| 37 |
```
|
| 38 |
## AMOP Pipeline Log
|
| 39 |
+
<details>
|
| 40 |
+
<summary>Click to expand</summary>
|
| 41 |
|
| 42 |
+
```
|
| 43 |
{pipeline_log}
|
| 44 |
+
```
|
| 45 |
+
</details>
|