Llama 3 8B Instruct 64K by MaziyarPanahi

 ยป  All LLMs  ยป  MaziyarPanahi  ยป  Llama 3 8B Instruct 64K   URL Share it on

  Arxiv:2309.10400   64k   Autotrain compatible   Axolotl Base model:finetune:winglian/l... Base model:winglian/llama-3-8b...   Conversational   Dataset:intel/orca dpo pairs   Dpo   En   Facebook   Finetuned   Instruct   Llama   Llama-3   Meta   Pose   Pytorch   Region:us   Safetensors   Sharded   Tensorflow

Llama 3 8B Instruct 64K Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Llama 3 8B Instruct 64K (MaziyarPanahi/Llama-3-8B-Instruct-64k)

Llama 3 8B Instruct 64K Parameters and Internals

Model Type 
text generation
Additional Notes 
This model uses PoSE to extend Llama's context length from 8k to 64k.
Supported Languages 
en (proficient)
Training Details 
Data Sources:
RedPajama V1 dataset
Data Volume:
300M tokens
Methodology:
rank stabilized LoRA of rank 256
Context Length:
64000
Model Architecture:
Llama-3
LLM NameLlama 3 8B Instruct 64K
Repository ๐Ÿค—https://huggingface.co/MaziyarPanahi/Llama-3-8B-Instruct-64k 
Model NameLlama-3-8B-Instruct-64k
Model CreatorMaziyarPanahi
Base Model(s)  Llama 3 8B 64K PoSE   winglian/Llama-3-8b-64k-PoSE
Model Size8b
Required VRAM16.1 GB
Updated2024-12-27
MaintainerMaziyarPanahi
Model Typellama
Instruction-BasedYes
Model Files  5.0 GB: 1-of-4   5.0 GB: 2-of-4   4.9 GB: 3-of-4   1.2 GB: 4-of-4
Supported Languagesen
Model ArchitectureLlamaForCausalLM
Licensellama3
Context Length8192
Model Max Length8192
Transformers Version4.40.0.dev0
Tokenizer ClassPreTrainedTokenizerFast
Vocabulary Size128256
Torch Data Typefloat16

Quantized Models of the Llama 3 8B Instruct 64K

Model
Likes
Downloads
VRAM
Llama 3 8B Instruct 64K GGUF1222408553 GB
... Instruct 64K HQQ 1bit Smashed1153 GB
Llama 3 8B Instruct 64K AWQ0185 GB

Best Alternatives to Llama 3 8B Instruct 64K

Best Alternatives
Context / RAM
Downloads
Likes
...a 3 8B Instruct Gradient 1048K1024K / 16.1 GB4165678
MrRoboto ProLong 8B V4b1024K / 16.1 GB1070
MrRoboto ProLong 8B V4c1024K / 16.1 GB860
MrRoboto ProLong 8B V1a1024K / 16.1 GB1080
MrRoboto ProLong 8B V2a1024K / 16.1 GB1020
...o ProLongBASE Pt6 Unaligned 8B1024K / 16.1 GB680
MrRoboto ProLong 8B V2f1024K / 16.1 GB770
...o ProLongBASE Pt2 Unaligned 8B1024K / 16.1 GB540
8B Unaligned BASE V2b1024K / 16.1 GB960
MrRoboto ProLong 8B V1l1024K / 16.1 GB680
Note: green Score (e.g. "73.2") means that the model is better than MaziyarPanahi/Llama-3-8B-Instruct-64k.

Rank the Llama 3 8B Instruct 64K Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 40304 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227