Stablelm Tuned Alpha 3B by stabilityai

 ยป  All LLMs  ยป  stabilityai  ยป  Stablelm Tuned Alpha 3B   URL Share it on

  Autotrain compatible   Dataset:dahoas/full-hh-rlhf   Dataset:dmayhem93/chatcombined Dataset:huggingfaceh4/databric... Dataset:jeffwan/sharegpt vicun... Dataset:nomic-ai/gpt4all promp...   Dataset:tatsu-lab/alpaca   En   Endpoints compatible   Gpt neox   Pytorch   Region:us   Sharded

Stablelm Tuned Alpha 3B Benchmarks

Stablelm Tuned Alpha 3B (stabilityai/stablelm-tuned-alpha-3b)

Stablelm Tuned Alpha 3B Parameters and Internals

Model Type 
causal-lm
Use Cases 
Areas:
open-source community, chat-like applications
Limitations:
The model may generate biased or toxic text despite efforts in safe fine-tuning., Not intended as a replacement for human judgment
Considerations:
Be mindful of potential bias or toxic outputs.
Additional Notes 
Models include a helpful hand from Dakota Mahan ([@dmayhem93](https://huggingface.co/dmayhem93)) in their development.
Supported Languages 
English (Proficient)
Training Details 
Data Sources:
tatsu-lab/alpaca, nomic-ai/gpt4all_prompt_generations, Dahoas/full-hh-rlhf, jeffwan/sharegpt_vicuna, HuggingFaceH4/databricks_dolly_15k
Methodology:
Supervised fine-tuning on natural language datasets focused on chat and instruction-following tasks.
Context Length:
4096
Model Architecture:
NeoX transformer architecture
Responsible Ai Considerations 
Fairness:
Models are developed to adhere to safer distributions of text but cannot mitigate all biases and toxicity.
Transparency:
It should not be treated as a substitute for human judgment or considered a source of truth.
Accountability:
Users are responsible for the outputs generated and should use models responsibly.
Mitigation Strategies:
Fine-tuning on datasets aimed at improving safety, but may not remove all biases/toxicity.
Input Output 
Input Format:
Prompts formatted to <|SYSTEM|>...<|USER|>...<|ASSISTANT|>...
Accepted Modalities:
text
Output Format:
Text output
LLM NameStablelm Tuned Alpha 3B
Repository ๐Ÿค—https://huggingface.co/stabilityai/stablelm-tuned-alpha-3b 
Model Size3b
Required VRAM14.9 GB
Updated2025-02-05
Maintainerstabilityai
Model Typegpt_neox
Model Files  10.2 GB: 1-of-2   4.7 GB: 2-of-2
Supported Languagesen
Model ArchitectureGPTNeoXForCausalLM
Licensecc-by-nc-sa-4.0
Context Length4096
Model Max Length4096
Transformers Version4.28.1
Tokenizer ClassGPTNeoXTokenizer
Vocabulary Size50688
Torch Data Typefloat32

Quantized Models of the Stablelm Tuned Alpha 3B

Model
Likes
Downloads
VRAM
Stablelm Tuned Alpha 3B 8bit3124 GB
Stablelm Tuned Alpha 3B 16bit6117 GB

Best Alternatives to Stablelm Tuned Alpha 3B

Best Alternatives
Context / RAM
Downloads
Likes
Stablecode Completion Alpha 3B16K / 14.1 GB120116
RedPajama 3B 1638416K / 19.7 GB175
Redpajama 3B Chat5K / 6.4 GB81552
Stablelm Base Alpha 3B4K / 14.9 GB189782
...blecode Completion Alpha 3B 4K4K / 6.1 GB1436282
Stablecode Instruct Alpha 3B4K / 6.1 GB35304
StableCode 3B4K / 6.1 GB141
...tion Alpha 3B 4K Openvino Int84K / 2.8 GB231
Redpajama 3B Evol Coder4K / 6.1 GB181
Literature 3B 40964K / 11.7 GB48

Rank the Stablelm Tuned Alpha 3B Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 42577 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227