Llama 2 13B Longlora 64K by Yukang

 ยป  All LLMs  ยป  Yukang  ยป  Llama 2 13B Longlora 64K   URL Share it on

  Arxiv:2309.12307   Autotrain compatible   Endpoints compatible   Llama   Lora   Region:us

Llama 2 13B Longlora 64K Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Llama 2 13B Longlora 64K (Yukang/Llama-2-13b-longlora-64k)

Llama 2 13B Longlora 64K Parameters and Internals

Model Type 
Language Model, Text Generation
Additional Notes 
LongLoRA extends models' context while retaining their original architectures, and is compatible with most existing techniques.
Training Details 
Data Sources:
LongQA dataset
Methodology:
Improved fine-tuning using sparse local attention and LoRA for context extension
Hardware Used:
8x A100 machine
Model Architecture:
Retains original architecture, compatible with techniques like FlashAttention-2
Release Notes 
Version:
1.0
Date:
2023
Notes:
Efficient fine-tuning approach using sparse local attention and improved LoRA.
LLM NameLlama 2 13B Longlora 64K
Repository ๐Ÿค—https://huggingface.co/Yukang/Llama-2-13b-longlora-64k 
Model Size13b
Required VRAM0.1 GB
Updated2025-02-22
MaintainerYukang
Model Files  0.1 GB   0.7 GB
Model ArchitectureAutoModelForCausalLM
Is Biasednone
PEFT TypeLORA
LoRA ModelYes
PEFT Target Modulesq_proj|k_proj|v_proj|o_proj
LoRA Alpha16
LoRA Dropout0
R Param8

Best Alternatives to Llama 2 13B Longlora 64K

Best Alternatives
Context / RAM
Downloads
Likes
ALMA 13B R0K / 26 GB824680
...ama2 13B No Robots Alpaca Lora0K / 4 GB192610
Llava V1.5 13B S2 Lora0K / 0.1 GB542
Llama 2 13B Chat Hf GPT 4 80K0K / 26 GB2770
Ko En Llama2 13B Finetune Ex0K / 26 GB24050
Ko En Llama2 13B Finetune0K / 26 GB24450
Ko Llama2 13B Finetune Ex0K / 26 GB19610
Llava Chart 13b Lora0K / 0.1 GB102
ChartLlama 13B0K / 0.1 GB35119
Orca 2 13B Pygmalion LoRA0K / 0.5 GB222
Note: green Score (e.g. "73.2") means that the model is better than Yukang/Llama-2-13b-longlora-64k.

Rank the Llama 2 13B Longlora 64K Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43508 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227