EasyContext 256K Danube2 1.8B by PY007

 ยป  All LLMs  ยป  PY007  ยป  EasyContext 256K Danube2 1.8B   URL Share it on

  Autotrain compatible   Endpoints compatible   Llama   Region:us   Safetensors

EasyContext 256K Danube2 1.8B Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
EasyContext 256K Danube2 1.8B (PY007/EasyContext-256K-danube2-1.8b)

EasyContext 256K Danube2 1.8B Parameters and Internals

Model Type 
context-extrapolated base model
Additional Notes 
Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.
Training Details 
Methodology:
Finetuned with EasyContext on context length 256K
Context Length:
256000
Input Output 
Performance Tips:
You can use this model with context length longer than 4096.
LLM NameEasyContext 256K Danube2 1.8B
Repository ๐Ÿค—https://huggingface.co/PY007/EasyContext-256K-danube2-1.8b 
Model Size1.8b
Required VRAM3.7 GB
Updated2025-02-22
MaintainerPY007
Model Typellama
Model Files  3.7 GB
Model ArchitectureLlamaForCausalLM
Context Length8192
Model Max Length8192
Transformers Version4.39.1
Tokenizer ClassLlamaTokenizer
Vocabulary Size32000
Torch Data Typebfloat16

Best Alternatives to EasyContext 256K Danube2 1.8B

Best Alternatives
Context / RAM
Downloads
Likes
Ssh 1.8B8K / 3.7 GB1730
Llm Jp 3 1.8B Instruct34K / 3.7 GB17571
Llm Jp 3 1.8B Instruct4K / 3.7 GB499424
Llm Jp 3 1.8B4K / 3.7 GB417113
Llm Jp 3 1.8B Instruct4K / 3.7 GB960
Qwen1.5 1.8B Llamafy4K / 3.7 GB1401
Tinyllama 1.8B Trismegistus2K / 1.9 GB1853
Llama1 S 1.8B Experimental2K / 7.3 GB234
TinyChat 1776K0.3K / 0 GB3079
Note: green Score (e.g. "73.2") means that the model is better than PY007/EasyContext-256K-danube2-1.8b.

Rank the EasyContext 256K Danube2 1.8B Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43470 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227