Shaltiel commited on
Commit
3294f69
·
verified ·
1 Parent(s): 6b22da6

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,12 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ DictaLM-3.0-24B-Thinking-BF16.gguf filter=lfs diff=lfs merge=lfs -text
37
+ DictaLM-3.0-24B-Thinking-IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text
38
+ DictaLM-3.0-24B-Thinking-IQ4_NL.gguf filter=lfs diff=lfs merge=lfs -text
39
+ DictaLM-3.0-24B-Thinking-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
40
+ DictaLM-3.0-24B-Thinking-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
41
+ DictaLM-3.0-24B-Thinking-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ DictaLM-3.0-24B-Thinking-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
43
+ DictaLM-3.0-24B-Thinking-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
44
+ imatrix_dict.gguf_file filter=lfs diff=lfs merge=lfs -text
DictaLM-3.0-24B-Thinking-BF16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec192683395e1c6bd0bc8fde07de4888be0267d4e986dd08449f3f5358f59a46
3
+ size 47153521504
DictaLM-3.0-24B-Thinking-IQ3_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd3c39453ea28bf9051faa7751ad6ed1624be46d91db56b12f1c67b94553057b
3
+ size 10650952800
DictaLM-3.0-24B-Thinking-IQ4_NL.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c260af29c1a89f03aa251271e5d7500500770bff751e13a7602b4f43ae9fd3f4
3
+ size 13468017760
DictaLM-3.0-24B-Thinking-IQ4_XS.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6db61a7d711740fe1e733d91f2b2167c3fbc987a2592b634c8defc2049ed90b
3
+ size 12758918240
DictaLM-3.0-24B-Thinking-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:666b7ae0e44e550c4e9b1d19235e444c1d46b2043daf3e209dc4c432f47409b6
3
+ size 14333912160
DictaLM-3.0-24B-Thinking-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a12186241abe2a83e63f68f48e330c057ca3e19f8f75b74677790b8d41661ca7
3
+ size 16763987040
DictaLM-3.0-24B-Thinking-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13b26926edaee2386bce509dddf28119b86a4593fe7916265320c68fc27c5409
3
+ size 19345941600
DictaLM-3.0-24B-Thinking-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17a5eef007086c004f6ae1df975879e31754380b52d1583f7712b895d434bcd8
3
+ size 25054782560
README.md ADDED
@@ -0,0 +1,102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ pipeline_tag: text-generation
4
+ language:
5
+ - en
6
+ - he
7
+ tags:
8
+ - pretrained
9
+ inference:
10
+ parameters:
11
+ temperature: 0.6
12
+ ---
13
+
14
+ [<img src="https://i.ibb.co/5Lbwyr1/dicta-logo.jpg" width="300px"/>](https://dicta.org.il)
15
+
16
+ # Dicta-LM 3.0: Advancing The Frontier of Hebrew Sovereign LLMs
17
+
18
+ Dicta-LM 3.0 is a powerful open-weight collection of LLMs, trained on extensive corpora of Hebrew and English texts. The models are available for download and for unlimited use. The models set a new SOTA for their weight-class for Hebrew, both as base models and chat models.
19
+
20
+ This is our flagship model, a 24-billion-parameter *reasoning* model, with full precision (BF16), originally initialized from [Mistral-Small-3.1-24B-Base-2503](https://huggingface.co/mistralai/Mistral-Small-3.1-24B-Base-2503).
21
+
22
+ This model is a reasoning chat model, which means that before responding to any given message from the user, the model first thinks out the right way to respond in a designated thinking block.
23
+
24
+ <br/>
25
+
26
+ &#128640; Try it out here: [chat.dicta.org.il](https://chat.dicta.org.il)
27
+
28
+ <br/>
29
+
30
+ For full details of this model please read our [release blog post](https://dicta.org.il/dicta-lm-3) or the [technical report](https://www.dicta.org.il/publications/DictaLM_3_0___Techincal_Report.pdf).
31
+
32
+ You can view and access the full collection of base/instruct unquantized/quantized versions of `DictaLM 3.0` [here](https://huggingface.co/collections/dicta-il/dictalm-30-collection).
33
+
34
+ ## Instruction format
35
+
36
+ In order to leverage instruction fine-tuning, your prompt should be rendered using the chat template specified for this model. Most libraries deal with this automatically, so you can just let them do it.
37
+
38
+ ## Usage
39
+
40
+ We recommend using vLLM, but you can use Transformers as well:
41
+
42
+ ### Transformers
43
+
44
+ ```python
45
+
46
+ from transformers import pipeline
47
+
48
+ generator = pipeline('text-generation', model="dicta-il/DictaLM-3.0-24B-Thinking")
49
+
50
+ messages = [
51
+ {"role": "user", "content": "איזה רוטב אהוב עליך?"},
52
+ {"role": "assistant", "content": "טוב, אני די מחבב כמה טיפות מיץ לימון סחוט טרי. זה מוסיף בדיוק את הכמות הנכונה של טעם חמצמץ לכל מה שאני מבשל במטבח!"},
53
+ {"role": "user", "content": "האם יש לך מתכונים למיונז?"}
54
+ ]
55
+
56
+ print(generator(messages)[0]['generated_text'][-1]) # just print the last message
57
+ # {'role': 'assistant', 'content': 'המשתמש שואל: "האם יש לך מתכונים למיונז?" עלינו להגיב עם מתכון למיונז. ספקו מתכון קלאסי, אולי גם וריאציות (מיונז פליאו, מיונז טבעוני, מיונז חריף, וכו\'). ספקו הוראות שלב אחר שלב. צריך להיות בטון ידידותי.</think>בטח! מיונז ביתי הוא באמת פשוט להכנה ומעניק לכם שליטה מלאה על המרכיבים, הטעם והמרקם. להלן המתכון הקלאסי, בתוספת כמה וריאציות מהנות שתוכלו להתנסות בהן.\n\n---\n\n## 1️⃣ מיונז קלאסי (מיונז בסיסי)\n\n| מרכיב | כמות |\n|------------|----------|\n| חלמון ביצה גדול (או 2 חלמונים בינוניים) | 1 |\n| חרדל דיז\'ון | 1 כפית |\n| מיץ לימון טרי (או ח'}
58
+ ```
59
+
60
+ ### vLLM
61
+
62
+ ```bash
63
+ vllm serve dicta-il/DictaLM-3.0-24B-Thinking --enable-auto-tool-choice --tool-call-parser hermes --reasoning_parser deepseek_r1
64
+ ```
65
+
66
+ And then you can access it via the openai library:
67
+
68
+ ```python
69
+ from openai import OpenAI
70
+
71
+ client = OpenAI(
72
+ base_url="http://localhost:8000/v1",
73
+ api_key="sk-no-key-required"
74
+ )
75
+
76
+ response = client.chat.completions.create(
77
+ model="dicta-il/DictaLM-3.0-24B-Thinking",
78
+ messages=[
79
+ {"role": "user", "content": "Hello, how are you?"}
80
+ ],
81
+ )
82
+
83
+ print(response.choices[0].message.content)
84
+ ```
85
+
86
+ > The reasoning traces should be available in the response structure in the designated fild.
87
+
88
+ The model supports tool-calling, enabling integration with external tools and APIs. For example how to use the tool calling, see the [vLLM documentation](https://docs.vllm.ai/en/stable/features/tool_calling/#tool-calling).
89
+
90
+ ## Citation
91
+
92
+ If you use this model, please cite:
93
+
94
+ ```bibtex
95
+ @article{Shmidman2025DictaLM3,
96
+ title={{Dicta-LM 3.0: Advancing The Frontier of Hebrew Sovereign LLMs}},
97
+ author={Shaltiel Shmidman and Avi Shmidman and Amir DN Cohen and Moshe Koppel},
98
+ year={2025},
99
+ publisher={{DICTA / Jerusalem, Israel}},
100
+ note={https://www.dicta.org.il/publications/DictaLM_3_0___Techincal_Report.pdf}
101
+ }
102
+ ```
imatrix_dict.gguf_file ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b214fd9c56dda229cff55d83d579fe1afd98a648b23a43fc3585385342dcf3ca
3
+ size 10037344