Fireball Alpaca Llama3.1.08 8B Philos C R1 KTO Beta by EpistemeAI

 ยป  All LLMs  ยป  EpistemeAI  ยป  Fireball Alpaca Llama3.1.08 8B Philos C R1 KTO Beta   URL Share it on

  Autotrain compatible Base model:epistemeai2/firebal... Base model:finetune:epistemeai...   Conversational Dataset:argilla/distilabel-int...   En   Endpoints compatible   Llama   Pytorch   Region:us   Safetensors   Sharded   Tensorflow   Trl   Unsloth

Fireball Alpaca Llama3.1.08 8B Philos C R1 KTO Beta Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Fireball Alpaca Llama3.1.08 8B Philos C R1 KTO Beta (EpistemeAI/Fireball-Alpaca-Llama3.1.08-8B-Philos-C-R1-KTO-beta)

Fireball Alpaca Llama3.1.08 8B Philos C R1 KTO Beta Parameters and Internals

Model Type 
text-generation-inference, transformers
Use Cases 
Areas:
Commercial, Research
Applications:
Assistant-like chat, Natural language generation tasks
Primary Use Cases:
Multilingual dialogue, Synthetic data generation, Data distillation
Limitations:
Use in non-supported languages requires additional tuning and responsibility by developers
Considerations:
Refer to Responsible Use Guide for language use beyond the eight supported languages.
Additional Notes 
Model trained 2x faster with Unsloth and Hugging Face's TRL library.
Supported Languages 
English (yes), German (yes), French (yes), Italian (yes), Portuguese (yes), Hindi (yes), Spanish (yes), Thai (yes)
Training Details 
Data Sources:
argilla/distilabel-intel-orca-kto
Data Volume:
15T+ tokens
Methodology:
KTO Fine tuning: Kahneman-Tversky Optimization (KTO)
Context Length:
128000
Model Architecture:
Auto-regressive language model with optimized transformer architecture
Safety Evaluation 
Methodologies:
Evaluation with adversarial datasets, Red teaming exercises
Risk Categories:
CBRNE helpfulness, Child Safety, Cyber attack enablement
Ethical Considerations:
Safety testing and tuning should be tailored to specific applications; potential for biased, objectionable, or inaccurate outputs.
Responsible Ai Considerations 
Accountability:
Developers are responsible for deploying safeguards for their specific use cases.
Mitigation Strategies:
Following Responsible Use Guide; integration of safeguards like Llama Guard 3, Prompt Guard, and Code Shield.
Input Output 
Input Format:
ChatML or Alpaca prompt template
Accepted Modalities:
Multilingual Text
Output Format:
Multilingual Text and code
LLM NameFireball Alpaca Llama3.1.08 8B Philos C R1 KTO Beta
Repository ๐Ÿค—https://huggingface.co/EpistemeAI/Fireball-Alpaca-Llama3.1.08-8B-Philos-C-R1-KTO-beta 
Base Model(s)  EpistemeAI2/Fireball-Alpaca-Llama3.1.08-8B-Philos-C-R1   EpistemeAI2/Fireball-Alpaca-Llama3.1.08-8B-Philos-C-R1
Model Size8b
Required VRAM16.1 GB
Updated2024-12-16
MaintainerEpistemeAI
Model Typellama
Model Files  5.0 GB: 1-of-4   5.0 GB: 2-of-4   4.9 GB: 3-of-4   1.2 GB: 4-of-4   5.0 GB: 1-of-4   5.0 GB: 2-of-4   4.9 GB: 3-of-4   1.2 GB: 4-of-4
Supported Languagesen
Gated ModelYes
Model ArchitectureLlamaForCausalLM
Licenseproprietary
Context Length131072
Model Max Length131072
Transformers Version4.44.2
Tokenizer ClassPreTrainedTokenizerFast
Padding Token<|finetune_right_pad_id|>
Vocabulary Size128256
Torch Data Typefloat16

Best Alternatives to Fireball Alpaca Llama3.1.08 8B Philos C R1 KTO Beta

Best Alternatives
Context / RAM
Downloads
Likes
...a 3 8B Instruct Gradient 1048K1024K / 16.1 GB6369680
MrRoboto ProLong 8B V4i1024K / 16.1 GB661
...o ProLongBASE Pt8 Unaligned 8B1024K / 16.1 GB240
41024K / 16.1 GB3220
161024K / 16.1 GB1690
Thor V1.4 8B DARK FICTION1024K / 16.1 GB9412
111024K / 16.1 GB1130
121024K / 16.1 GB600
HEL V0.8 8B LONG DARK1024K / 16.1 GB2320
MrRoboto ProLong 8B V2b1024K / 16.1 GB1780
Note: green Score (e.g. "73.2") means that the model is better than EpistemeAI/Fireball-Alpaca-Llama3.1.08-8B-Philos-C-R1-KTO-beta.

Rank the Fireball Alpaca Llama3.1.08 8B Philos C R1 KTO Beta Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 41636 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227