Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
Pricing
Log In
Sign Up
ayoubkirouane
/
Mistral-SLERP-Merged7B-DPO
like
0
Text Generation
PEFT
Safetensors
HuggingFaceH4/ultrafeedback_binarized
ayoubkirouane/Orca-Direct-Preference-Optimization
trl
dpo
unsloth
conversational
License:
apache-2.0
Model card
Files
Files and versions
xet
Community
Use this model
main
Mistral-SLERP-Merged7B-DPO
672 MB
1 contributor
History:
7 commits
ayoubkirouane
Update README.md
8668468
verified
about 2 years ago
.gitattributes
1.52 kB
initial commit
about 2 years ago
README.md
1.52 kB
Update README.md
about 2 years ago
adapter_config.json
670 Bytes
ayoubkirouane/Mistral-SLERP-Merged7B-DPO
about 2 years ago
adapter_model.safetensors
671 MB
xet
ayoubkirouane/Mistral-SLERP-Merged7B-DPO
about 2 years ago
added_tokens.json
26 Bytes
ayoubkirouane/Mistral-SLERP-Merged7B-DPO
about 2 years ago
special_tokens_map.json
449 Bytes
ayoubkirouane/Mistral-SLERP-Merged7B-DPO
about 2 years ago
tokenizer.model
493 kB
xet
ayoubkirouane/Mistral-SLERP-Merged7B-DPO
about 2 years ago
tokenizer_config.json
1.43 kB
ayoubkirouane/Mistral-SLERP-Merged7B-DPO
about 2 years ago
training_args.bin
4.73 kB
xet
ayoubkirouane/Mistral-SLERP-Merged7B-DPO
about 2 years ago