Atlas Flash 1.5B Preview by open-atlas

 ยป  All LLMs  ยป  open-atlas  ยป  Atlas Flash 1.5B Preview   URL Share it on

  Arxiv:2312.14852   Arxiv:2501.12948   4bit   Ar   Autotrain compatible Base model:deepseek-ai/deepsee... Base model:finetune:deepseek-a...   Bn   Ceb   Chat   Conversational   Cs   Dataset:baai/taco   Dataset:codeparrot/apps Dataset:rubenroy/gammacorpus-v...   De   En   Endpoints compatible   Es   Fa   Fr   Gemini-2.0   Gpt4   He   Hi   Id   It   Ja   Km   Ko   Lo   Ms   My   Nl   Pl   Pt   Quantized   Qwen2   R1   Region:us   Ru   Safetensors   Th   Tl   Tr   Trl   Ur   Vi   Zh

Atlas Flash 1.5B Preview Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Atlas Flash 1.5B Preview (Spestly/Atlas-Flash-1.5B-Preview)

Atlas Flash 1.5B Preview Parameters and Internals

LLM NameAtlas Flash 1.5B Preview
Repository ๐Ÿค—https://huggingface.co/Spestly/Atlas-Flash-1.5B-Preview 
Base Model(s)  deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B   deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
Model Size1.5b
Required VRAM3.5 GB
Updated2025-03-11
Maintaineropen-atlas
Model Typeqwen2
Model Files  3.5 GB
Supported Languagesen zh fr es pt de it ru ja ko vi th ar fa he tr cs pl hi bn ur id ms lo my km tl nl
Quantization Type4bit
Model ArchitectureQwen2ForCausalLM
Licensemit
Context Length131072
Model Max Length131072
Transformers Version4.47.1
Tokenizer ClassLlamaTokenizer
Padding Token<|vision_pad|>
Vocabulary Size151936
Torch Data Typefloat16

Best Alternatives to Atlas Flash 1.5B Preview

Best Alternatives
Context / RAM
Downloads
Likes
...ill Qwen 1.5B Unsloth Bnb 4bit128K / 1.8 GB13053313
... R1 Distill Qwen 1.5B Bnb 4bit128K / 1.6 GB1358287
Research PathfinderAI128K / 3.5 GB190
Deep Ft8 Grp 16bit128K / 3.5 GB960
Deep Ft4 Grp 16bit128K / 3.5 GB3030
...pSeek R1 Distill Qwen 1.5B 500128K / 1.7 GB280
DeepSeek R1 ForumNXT128K / 3.5 GB70
Deep Ft5 Grp 16bit128K / 3.5 GB820
Model128K / 3.5 GB190
Qwen2 1.5B Bnb 4bit128K / 1.1 GB33996
Note: green Score (e.g. "73.2") means that the model is better than Spestly/Atlas-Flash-1.5B-Preview.

Rank the Atlas Flash 1.5B Preview Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 44902 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227