Magnum 12B V2.5 Kto by anthracite-org

 ยป  All LLMs  ยป  anthracite-org  ยป  Magnum 12B V2.5 Kto   URL Share it on

  Chat   Conversational   De   En   Es   Fr   It   Ja   Mistral   Pt   Region:us   Ru   Safetensors   Sharded   Tensorflow   Zh

Magnum 12B V2.5 Kto Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Magnum 12B V2.5 Kto (anthracite-org/magnum-12b-v2.5-kto)

Magnum 12B V2.5 Kto Parameters and Internals

Model Type 
text generation, chat
Additional Notes 
KTO is an experimental release, part of a series of models. It's fine-tuned on top of Magnum-12b-v2. Experimental data was used for initial testing, with plans to scale up.
Supported Languages 
en (supported), fr (supported), de (supported), es (supported), it (supported), pt (supported), ru (supported), zh (supported), ja (supported)
Training Details 
Data Sources:
Stheno dataset (filtered), kalomaze/Opus_Instruct_25k, Nopm/Opus_WritingStruct, Gryphe/Sonnet3.5-SlimOrcaDedupCleaned, kalomaze/Opus_Instruct_3k
Methodology:
hybrid reinforcement learning strategy of KTO + DPOP using rejected data sampled from the original model as rejected and data from original finetuning dataset as chosen.
Input Output 
Input Format:
Instruct tuned with the ChatML formatting.
LLM NameMagnum 12B V2.5 Kto
Repository ๐Ÿค—https://huggingface.co/anthracite-org/magnum-12b-v2.5-kto 
Model Size12b
Required VRAM24.5 GB
Updated2024-08-18
Maintaineranthracite-org
Model Typemistral
Model Files  4.9 GB: 1-of-5   4.9 GB: 2-of-5   4.9 GB: 3-of-5   4.9 GB: 4-of-5   4.9 GB: 5-of-5
Supported Languagesen fr de es it pt ru zh ja
Model ArchitectureMistralForCausalLM
Licenseapache-2.0
Context Length1024000
Model Max Length1024000
Transformers Version4.43.3
Tokenizer ClassPreTrainedTokenizerFast
Padding Token<pad>
Vocabulary Size131072
Torch Data Typefloat16

Best Alternatives to Magnum 12B V2.5 Kto

Best Alternatives
Context / RAM
Downloads
Likes
...r Nemo 12B Instruct R 21 09 241000K / 24.5 GB8449106
...s PersonalityEngine V1.1.0 12B1000K / 24.5 GB49229
Captain Eris Violet V0.420 12B1000K / 24.5 GB106923
Mistral Nemo Kartoffel 12B1000K / 24.5 GB1833
Saiga Nemo 12b1000K / 24.5 GB36481337
MN 12B Mimicore GreenSnake1000K / 24.5 GB832
MN 12B Mimicore WhiteSnake1000K / 24.5 GB613
MN 12B Mag Mell R11000K / 24.5 GB424699
SauerkrautLM Nemo 12B Instruct1000K / 24.5 GB1952722
MN 12B Mimicore Orochi1000K / 24.5 GB312
Note: green Score (e.g. "73.2") means that the model is better than anthracite-org/magnum-12b-v2.5-kto.

Rank the Magnum 12B V2.5 Kto Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 42577 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227