CyberSec AI Portfolio - Datasets, Models & Spaces
Collection
80+ datasets, 35 Spaces & 4 models for cybersecurity AI: RGPD, NIS2, ISO 27001, DORA, AI Act, MITRE ATT&CK & more. By Ayi NEDJIMI. • 139 items • Updated
• 2
GGUF quantized versions of AYI-NEDJIMI/CyberSec-Assistant-3B for use with Ollama, llama.cpp, LM Studio, and other GGUF-compatible inference engines.
This is a fine-tuned Qwen2.5-3B-Instruct model specialized in general cybersecurity. It can answer questions about network security, vulnerability assessment, incident response, penetration testing, threat analysis, security architecture, and cybersecurity best practices in both French and English.
Part of the AYI-NEDJIMI Cybersecurity AI Portfolio:
| Filename | Quant Type | Size | Description |
|---|---|---|---|
cybersec-assistant-3b-Q4_K_M.gguf |
Q4_K_M | 1.80 GB | Recommended — Best balance of quality and size (~31% of F16) |
cybersec-assistant-3b-Q5_K_M.gguf |
Q5_K_M | 2.07 GB | Higher quality, slightly larger (~36% of F16) |
cybersec-assistant-3b-Q8_0.gguf |
Q8_0 | 3.06 GB | Near-lossless quantization (~53% of F16) |
Create a Modelfile:
FROM ./cybersec-assistant-3b-Q4_K_M.gguf
TEMPLATE """<|im_start|>system
{{ .System }}<|im_end|>
<|im_start|>user
{{ .Prompt }}<|im_end|>
<|im_start|>assistant
"""
SYSTEM "You are a cybersecurity expert assistant. You provide detailed, accurate guidance on network security, vulnerability assessment, incident response, penetration testing, and security best practices. You respond in the same language as the user's question."
PARAMETER temperature 0.7
PARAMETER top_p 0.8
PARAMETER top_k 20
PARAMETER stop "<|im_end|>"
Then run:
ollama create cybersec-assistant -f Modelfile
ollama run cybersec-assistant
# Interactive chat
./llama-cli -m cybersec-assistant-3b-Q4_K_M.gguf \
-p "You are a cybersecurity expert assistant." \
--chat-template chatml \
-cnv
# Server mode
./llama-server -m cybersec-assistant-3b-Q4_K_M.gguf \
--host 0.0.0.0 --port 8080
from llama_cpp import Llama
llm = Llama(model_path="cybersec-assistant-3b-Q4_K_M.gguf", n_ctx=4096)
response = llm.create_chat_completion(
messages=[
{"role": "system", "content": "You are a cybersecurity expert assistant."},
{"role": "user", "content": "Explain the MITRE ATT&CK framework and how it helps in threat detection."}
],
temperature=0.7,
top_p=0.8,
top_k=20,
)
print(response["choices"][0]["message"]["content"])
| Version | Link |
|---|---|
| Merged (SafeTensors) | AYI-NEDJIMI/CyberSec-Assistant-3B |
| LoRA Adapter | AYI-NEDJIMI/CyberSec-Assistant-3B-Adapter |
| GGUF (this repo) | AYI-NEDJIMI/CyberSec-Assistant-3B-GGUF |
| Portfolio Collection | AYI-NEDJIMI/CyberSec-AI-Portfolio |
4-bit
5-bit
8-bit