Upload folder using huggingface_hub
Browse files
README.md
CHANGED
|
@@ -37,7 +37,7 @@ inference:
|
|
| 37 |
|
| 38 |

|
| 39 |

|
| 40 |
-

|
| 42 |
|
| 43 |
## π― Overview
|
|
@@ -66,7 +66,7 @@ Processing everything with models like LLaMA 3 8B is powerful but **slow and exp
|
|
| 66 |
- **Base Model**: SmolLM2-360M-Instruct (HuggingFace HuggingFaceTB/SmolLM2-360M-Instruct)
|
| 67 |
- **Architecture**: LlamaForCausalLM
|
| 68 |
- **Parameters**: ~360 million
|
| 69 |
-
- **Context Length**:
|
| 70 |
- **Vocabulary**: 49,152 tokens
|
| 71 |
- **Precision**: bfloat16
|
| 72 |
- **Training Framework**: Transformers 4.52.4
|
|
@@ -114,15 +114,6 @@ Powers the Stock Trading Analysis & Real-time Signals platform:
|
|
| 114 |
- **Medium-scoring content** β Automated tagging and storage
|
| 115 |
- **Low-scoring content** β Filtered out entirely
|
| 116 |
|
| 117 |
-
## π Performance Benefits
|
| 118 |
-
|
| 119 |
-
| Metric | Smol News Scorer | Large Model Only |
|
| 120 |
-
|--------|------------------|------------------|
|
| 121 |
-
| **Speed** | ~50ms per item | ~2-5s per item |
|
| 122 |
-
| **Cost** | $0.001 per 1K items | $0.01+ per 1K items |
|
| 123 |
-
| **Throughput** | 1000+ items/minute | 50-100 items/minute |
|
| 124 |
-
| **Resource Usage** | 2GB VRAM | 16GB+ VRAM |
|
| 125 |
-
|
| 126 |
## π» Usage Examples
|
| 127 |
|
| 128 |
### Basic Inference
|
|
@@ -330,7 +321,7 @@ WOW CONFIDENCE: 0.85
|
|
| 330 |
- **Latency**: ~50ms per news item (CPU), ~20ms (GPU)
|
| 331 |
- **Throughput**: 1000+ items/minute on modest hardware
|
| 332 |
- **Accuracy**: 85%+ correlation with human financial analysts
|
| 333 |
-
- **Memory**: 2GB VRAM required for inference
|
| 334 |
- **CPU Alternative**: Runs efficiently on CPU-only systems
|
| 335 |
|
| 336 |
## β‘ Deployment Options
|
|
@@ -372,10 +363,8 @@ outputs = llm.generate(prompts, sampling_params)
|
|
| 372 |
## π― Integration Roadmap
|
| 373 |
|
| 374 |
### Current Integrations
|
| 375 |
-
- β
YouTube Financial Video Analyzer (
|
| 376 |
-
- β
STARS Trading System (
|
| 377 |
-
- β
Kafka streaming pipeline
|
| 378 |
-
- β
Real-time WebSocket alerts
|
| 379 |
|
| 380 |
### Planned Integrations
|
| 381 |
- π Discord/Slack trading bots
|
|
|
|
| 37 |
|
| 38 |

|
| 39 |

|
| 40 |
+

|
| 41 |

|
| 42 |
|
| 43 |
## π― Overview
|
|
|
|
| 66 |
- **Base Model**: SmolLM2-360M-Instruct (HuggingFace HuggingFaceTB/SmolLM2-360M-Instruct)
|
| 67 |
- **Architecture**: LlamaForCausalLM
|
| 68 |
- **Parameters**: ~360 million
|
| 69 |
+
- **Context Length**: 2,048 tokens
|
| 70 |
- **Vocabulary**: 49,152 tokens
|
| 71 |
- **Precision**: bfloat16
|
| 72 |
- **Training Framework**: Transformers 4.52.4
|
|
|
|
| 114 |
- **Medium-scoring content** β Automated tagging and storage
|
| 115 |
- **Low-scoring content** β Filtered out entirely
|
| 116 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 117 |
## π» Usage Examples
|
| 118 |
|
| 119 |
### Basic Inference
|
|
|
|
| 321 |
- **Latency**: ~50ms per news item (CPU), ~20ms (GPU)
|
| 322 |
- **Throughput**: 1000+ items/minute on modest hardware
|
| 323 |
- **Accuracy**: 85%+ correlation with human financial analysts
|
| 324 |
+
- **Memory**: <2GB VRAM required for inference
|
| 325 |
- **CPU Alternative**: Runs efficiently on CPU-only systems
|
| 326 |
|
| 327 |
## β‘ Deployment Options
|
|
|
|
| 363 |
## π― Integration Roadmap
|
| 364 |
|
| 365 |
### Current Integrations
|
| 366 |
+
- β
YouTube Financial Video Analyzer [Link](https://levidehaan.com/projects/youtube-financial-video-analyzer)
|
| 367 |
+
- β
STARS Trading System [Link](https://levidehaan.com/projects/stars-trading-system)
|
|
|
|
|
|
|
| 368 |
|
| 369 |
### Planned Integrations
|
| 370 |
- π Discord/Slack trading bots
|