LeviDeHaan commited on
Commit
14b1d81
Β·
verified Β·
1 Parent(s): ae32427

Upload folder using huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +5 -16
README.md CHANGED
@@ -37,7 +37,7 @@ inference:
37
 
38
  ![Model Size](https://img.shields.io/badge/Parameters-360M-blue)
39
  ![Architecture](https://img.shields.io/badge/Architecture-LlamaForCausalLM-green)
40
- ![Context Length](https://img.shields.io/badge/Context-8192-orange)
41
  ![License](https://img.shields.io/badge/License-MIT-green)
42
 
43
  ## 🎯 Overview
@@ -66,7 +66,7 @@ Processing everything with models like LLaMA 3 8B is powerful but **slow and exp
66
  - **Base Model**: SmolLM2-360M-Instruct (HuggingFace HuggingFaceTB/SmolLM2-360M-Instruct)
67
  - **Architecture**: LlamaForCausalLM
68
  - **Parameters**: ~360 million
69
- - **Context Length**: 8,192 tokens
70
  - **Vocabulary**: 49,152 tokens
71
  - **Precision**: bfloat16
72
  - **Training Framework**: Transformers 4.52.4
@@ -114,15 +114,6 @@ Powers the Stock Trading Analysis & Real-time Signals platform:
114
  - **Medium-scoring content** β†’ Automated tagging and storage
115
  - **Low-scoring content** β†’ Filtered out entirely
116
 
117
- ## πŸš€ Performance Benefits
118
-
119
- | Metric | Smol News Scorer | Large Model Only |
120
- |--------|------------------|------------------|
121
- | **Speed** | ~50ms per item | ~2-5s per item |
122
- | **Cost** | $0.001 per 1K items | $0.01+ per 1K items |
123
- | **Throughput** | 1000+ items/minute | 50-100 items/minute |
124
- | **Resource Usage** | 2GB VRAM | 16GB+ VRAM |
125
-
126
  ## πŸ’» Usage Examples
127
 
128
  ### Basic Inference
@@ -330,7 +321,7 @@ WOW CONFIDENCE: 0.85
330
  - **Latency**: ~50ms per news item (CPU), ~20ms (GPU)
331
  - **Throughput**: 1000+ items/minute on modest hardware
332
  - **Accuracy**: 85%+ correlation with human financial analysts
333
- - **Memory**: 2GB VRAM required for inference
334
  - **CPU Alternative**: Runs efficiently on CPU-only systems
335
 
336
  ## ⚑ Deployment Options
@@ -372,10 +363,8 @@ outputs = llm.generate(prompts, sampling_params)
372
  ## 🎯 Integration Roadmap
373
 
374
  ### Current Integrations
375
- - βœ… YouTube Financial Video Analyzer (React frontend)
376
- - βœ… STARS Trading System (Express.js backend)
377
- - βœ… Kafka streaming pipeline
378
- - βœ… Real-time WebSocket alerts
379
 
380
  ### Planned Integrations
381
  - πŸ”„ Discord/Slack trading bots
 
37
 
38
  ![Model Size](https://img.shields.io/badge/Parameters-360M-blue)
39
  ![Architecture](https://img.shields.io/badge/Architecture-LlamaForCausalLM-green)
40
+ ![Context Length](https://img.shields.io/badge/Context-2048-orange)
41
  ![License](https://img.shields.io/badge/License-MIT-green)
42
 
43
  ## 🎯 Overview
 
66
  - **Base Model**: SmolLM2-360M-Instruct (HuggingFace HuggingFaceTB/SmolLM2-360M-Instruct)
67
  - **Architecture**: LlamaForCausalLM
68
  - **Parameters**: ~360 million
69
+ - **Context Length**: 2,048 tokens
70
  - **Vocabulary**: 49,152 tokens
71
  - **Precision**: bfloat16
72
  - **Training Framework**: Transformers 4.52.4
 
114
  - **Medium-scoring content** β†’ Automated tagging and storage
115
  - **Low-scoring content** β†’ Filtered out entirely
116
 
 
 
 
 
 
 
 
 
 
117
  ## πŸ’» Usage Examples
118
 
119
  ### Basic Inference
 
321
  - **Latency**: ~50ms per news item (CPU), ~20ms (GPU)
322
  - **Throughput**: 1000+ items/minute on modest hardware
323
  - **Accuracy**: 85%+ correlation with human financial analysts
324
+ - **Memory**: <2GB VRAM required for inference
325
  - **CPU Alternative**: Runs efficiently on CPU-only systems
326
 
327
  ## ⚑ Deployment Options
 
363
  ## 🎯 Integration Roadmap
364
 
365
  ### Current Integrations
366
+ - βœ… YouTube Financial Video Analyzer [Link](https://levidehaan.com/projects/youtube-financial-video-analyzer)
367
+ - βœ… STARS Trading System [Link](https://levidehaan.com/projects/stars-trading-system)
 
 
368
 
369
  ### Planned Integrations
370
  - πŸ”„ Discord/Slack trading bots