Apollo 2.0 Llama 3.1 8B by Locutusque

 ยป  All LLMs  ยป  Locutusque  ยป  Apollo 2.0 Llama 3.1 8B   URL Share it on

  Autotrain compatible   Conversational Dataset:locutusque/apollorp-2....   En   Endpoints compatible   Instruct   Llama   Not-for-all-audiences   Region:us   Safetensors   Sharded   Tensorflow

Apollo 2.0 Llama 3.1 8B Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Apollo 2.0 Llama 3.1 8B (Locutusque/Apollo-2.0-Llama-3.1-8B)

Apollo 2.0 Llama 3.1 8B Parameters and Internals

Model Type 
text generation
Use Cases 
Areas:
RP, ERP, instruction following, conversation
Limitations:
not-for-all-audiences
Considerations:
Users should be made aware of the risks, biases, and limitations of the model.
Additional Notes 
The model is designed to be used for roleplaying and conversational purposes, but it is uncensored and may generate inappropriate content. It's advised to be cautious while using it.
Supported Languages 
en (proficient)
Training Details 
Data Sources:
Locutusque/ApolloRP-2.0-SFT
Methodology:
Fine-tuning. Trained with bf16 non-mixed precision.
Responsible Ai Considerations 
Mitigation Strategies:
Use at your own risk due to the model's uncensored nature.
LLM NameApollo 2.0 Llama 3.1 8B
Repository ๐Ÿค—https://huggingface.co/Locutusque/Apollo-2.0-Llama-3.1-8B 
Model Size8b
Required VRAM16 GB
Updated2024-12-21
MaintainerLocutusque
Model Typellama
Instruction-BasedYes
Model Files  3.0 GB: 1-of-6   2.9 GB: 2-of-6   3.0 GB: 3-of-6   2.9 GB: 4-of-6   2.9 GB: 5-of-6   1.3 GB: 6-of-6
Supported Languagesen
Model ArchitectureLlamaForCausalLM
Licensellama3.1
Context Length131072
Model Max Length131072
Transformers Version4.44.2
Tokenizer ClassPreTrainedTokenizerFast
Padding Token<|eot_id|>
Vocabulary Size128256
Torch Data Typebfloat16

Best Alternatives to Apollo 2.0 Llama 3.1 8B

Best Alternatives
Context / RAM
Downloads
Likes
...a 3 8B Instruct Gradient 1048K1024K / 16.1 GB5528678
MrRoboto ProLong 8B V1a1024K / 16.1 GB1070
MrRoboto ProLong 8B V2a1024K / 16.1 GB1000
MrRoboto ProLong 8B V2f1024K / 16.1 GB510
MrRoboto ProLong 8B V1f1024K / 16.1 GB630
MrRoboto ProLong 8B V1l1024K / 16.1 GB600
8B Unaligned BASE V2b1024K / 16.1 GB930
MrRoboto ProLong 8B V1h1024K / 16.1 GB360
MrRoboto ProLong 8B V1d1024K / 16.1 GB340
MrRoboto ProLong 8B V1m1024K / 16.1 GB280
Note: green Score (e.g. "73.2") means that the model is better than Locutusque/Apollo-2.0-Llama-3.1-8B.

Rank the Apollo 2.0 Llama 3.1 8B Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 40013 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241217