Saiga Llama3 70B Sft M1 D5 Abliterated Kto M1 D2 AWQ 4bit by IlyaGusev

 ยป  All LLMs  ยป  IlyaGusev  ยป  Saiga Llama3 70B Sft M1 D5 Abliterated Kto M1 D2 AWQ 4bit   URL Share it on

  4-bit   4bit   Autotrain compatible   Awq   Conversational   Endpoints compatible   Llama   Quantized   Region:us   Safetensors   Sharded   Tensorflow

Saiga Llama3 70b Sft M1 D5 Abliterated Kto M1 D2 AWQ 4bit Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Saiga Llama3 70B Sft M1 D5 Abliterated Kto M1 D2 AWQ 4bit (IlyaGusev/saiga_llama3_70b_sft_m1_d5_abliterated_kto_m1_d2_awq_4bit)

Saiga Llama3 70B Sft M1 D5 Abliterated Kto M1 D2 AWQ 4bit Parameters and Internals

LLM NameSaiga Llama3 70b Sft M1 D5 Abliterated Kto M1 D2 AWQ 4bit
Repository ๐Ÿค—https://huggingface.co/IlyaGusev/saiga_llama3_70b_sft_m1_d5_abliterated_kto_m1_d2_awq_4bit 
Base Model(s)  ...ft M1 D5 Abliterated Kto M1 D2   IlyaGusev/saiga_llama3_70b_sft_m1_d5_abliterated_kto_m1_d2
Model Size70b
Required VRAM39.9 GB
Updated2025-02-22
MaintainerIlyaGusev
Model Typellama
Model Files  5.0 GB: 1-of-9   4.9 GB: 2-of-9   4.9 GB: 3-of-9   4.9 GB: 4-of-9   4.9 GB: 5-of-9   4.9 GB: 6-of-9   4.9 GB: 7-of-9   3.4 GB: 8-of-9   2.1 GB: 9-of-9
AWQ QuantizationYes
Quantization Typeawq|4bit
Model ArchitectureLlamaForCausalLM
Context Length8192
Model Max Length8192
Transformers Version4.41.1
Tokenizer ClassPreTrainedTokenizerFast
Padding Token<|begin_of_text|>
Vocabulary Size128256
Torch Data Typefloat16

Best Alternatives to Saiga Llama3 70B Sft M1 D5 Abliterated Kto M1 D2 AWQ 4bit

Best Alternatives
Context / RAM
Downloads
Likes
...0B Instruct Gradient 1048K AWQ1024K / 39.9 GB101
...70B Instruct Gradient 262K AWQ256K / 39.9 GB110
Llama 3.3 70B Instruct AWQ128K / 39.9 GB6304628
...lama 3.3 70B Instruct AWQ INT4128K / 39.9 GB1983219
... SauerkrautLM 70B Instruct AWQ128K / 39.9 GB2724
MultiVerse 70B AWQ32K / 41.3 GB442
Opus V1.2 70B AWQ32K / 36.7 GB171
QuartetAnemoi 70B T0.0001 AWQ31K / 36.7 GB81
Senku 70B AWQ 4bit GEMM31K / 36.7 GB71
Kiqu 70B AWQ31K / 36.7 GB201
Note: green Score (e.g. "73.2") means that the model is better than IlyaGusev/saiga_llama3_70b_sft_m1_d5_abliterated_kto_m1_d2_awq_4bit.

Rank the Saiga Llama3 70B Sft M1 D5 Abliterated Kto M1 D2 AWQ 4bit Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43470 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227