Model Type | |
Use Cases |
Areas: | chatbot applications, text generation |
|
Primary Use Cases: | interactive chatbots, text completion |
|
|
Additional Notes | This model is compact and designed for efficient computation and memory footprint. |
|
Supported Languages | |
Training Details |
Data Sources: | cerebras/SlimPajama-627B, bigcode/starcoderdata, OpenAssistant/oasst_top1_2023-08-25 |
|
Data Volume: | |
Methodology: | Pretraining with optimization, same architecture and tokenizer as Llama 2, further aligned with TRL's DPOTrainer on UltraFeedback dataset |
|
Training Time: | |
Hardware Used: | |
Model Architecture: | |
|
Input Output |
Input Format: | JSON-like structured input for messages |
|
Accepted Modalities: | |
Output Format: | |
Performance Tips: | Use proper transformers and accelerate installation for optimal performance |
|
|