Stablelm 3B 4e1t GGUF by afrideva

 ยป  All LLMs  ยป  afrideva  ยป  Stablelm 3B 4e1t GGUF   URL Share it on

  Arxiv:1607.06450   Arxiv:1910.02054   Arxiv:1910.07467   Arxiv:2101.00027   Arxiv:2104.09864   Arxiv:2204.06745   Arxiv:2305.06161   Arxiv:2307.09288 Base model:quantized:stability... Base model:stabilityai/stablel...   Dataset:allenai/pes2o   Dataset:bigcode/starcoderdata   Dataset:carperai/pilev2-dev Dataset:tiiuae/falcon-refinedw... Dataset:togethercomputer/redpa...   En   Ggml   Gguf   Q2   Q2 k   Q3 k m   Q4 k m   Q5 k m   Q6 k   Q8 0   Quantized   Region:us   Stablelm

Stablelm 3B 4e1t GGUF Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Stablelm 3B 4e1t GGUF (afrideva/stablelm-3b-4e1t-GGUF)

Stablelm 3B 4e1t GGUF Parameters and Internals

Model Type 
decoder-only transformer, auto-regressive language model, causal-lm
Use Cases 
Areas:
foundational base model for application-specific fine-tuning
Limitations:
unreliable, unsafe behaviors, offensive content
Considerations:
Evaluate and fine-tune model for safe performance in downstream applications.
Supported Languages 
en (primary)
Training Details 
Data Sources:
tiiuae/falcon-refinedweb, togethercomputer/RedPajama-Data-1T, CarperAI/pilev2-dev, bigcode/starcoderdata, allenai/peS2o
Data Volume:
1 trillion tokens
Methodology:
bfloat16 precision, AdamW optimization, trained with NeoX tokenizer
Context Length:
4096
Training Time:
30 days
Hardware Used:
256 NVIDIA A100 40GB GPUs (AWS P4d instances)
Model Architecture:
decoder-only, Rotary Position Embeddings, LayerNorm with learned bias
Input Output 
Accepted Modalities:
text
Performance Tips:
Fine-tune model for specific applications.
LLM NameStablelm 3B 4e1t GGUF
Repository ๐Ÿค—https://huggingface.co/afrideva/stablelm-3b-4e1t-GGUF 
Model Namestablelm-3b-4e1t
Model Creatorstabilityai
Base Model(s)  Stablelm 3B 4e1t   stabilityai/stablelm-3b-4e1t
Model Size3b
Required VRAM1.2 GB
Updated2025-02-05
Maintainerafrideva
Model Typestablelm
Model Files  1.2 GB   1.4 GB   1.7 GB   2.0 GB   2.3 GB   3.0 GB
Supported Languagesen
GGUF QuantizationYes
Quantization Typegguf|q2|q8_0
Model ArchitectureAutoModel
Licensecc-by-sa-4.0

Best Alternatives to Stablelm 3B 4e1t GGUF

Best Alternatives
Context / RAM
Downloads
Likes
Llama 3.2 3B VanRossum0K / 6.5 GB1210
QwQ LCoT 3B Instruct GGUF0K / 1.9 GB92112
Llama Deepsync 3B GGUF0K / 2 GB3059
Llama Chat Summary 3.2 3B GGUF0K / 2 GB4318
Deepsync 240 GGUF0K / 2 GB2138
...a Song Stream 3B Instruct GGUF0K / 2 GB23710
...ma Doctor 3.2 3B Instruct GGUF0K / 2 GB61811
...ma Magpie 3.2 3B Instruct GGUF0K / 2 GB3228
... Sentient 3.2 3B Instruct GGUF0K / 2 GB22711
Llama 3.2 3B Instruct GGUF0K / 2 GB2358
Note: green Score (e.g. "73.2") means that the model is better than afrideva/stablelm-3b-4e1t-GGUF.

Rank the Stablelm 3B 4e1t GGUF Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 42577 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227