Model Type | text generation, instruction tuned, conversation, assistant |
|
Use Cases |
Areas: | research, controlled scientific experiments |
|
Applications: | alignment, text generation |
|
Primary Use Cases: | Developing language models for low-resource languages |
|
Limitations: | Not intended for deployment, Not suitable for human-facing interactions, Limited to Brazilian Portuguese, Not suitable for translation |
|
|
Additional Notes | Not suitable for real-world deployment without further evaluations. |
|
Supported Languages | Portuguese (Brazilian Portuguese) |
|
Training Details |
Data Sources: | Instruct-Aira Dataset version 2.0 |
|
Data Volume: | |
Methodology: | |
Hardware Used: | |
|
Responsible Ai Considerations |
Fairness: | This model can exhibit biases and generate toxic content due to the social and historical stereotypes present in the training data. |
|
Mitigation Strategies: | Users are advised to perform risk analysis and have human moderation when the model's interactions reach an audience. |
|
|
Input Output |
Input Format: | Text input with special token markers. |
|
Accepted Modalities: | |
Output Format: | |
Performance Tips: | Use adequate repetition penalty and temperature settings to minimize repetition and verbosity. |
|
|
Release Notes |
Version: | |
Date: | |
Notes: | TeenyTinyLlama-460m-Chat fine-tuned on the Instruct-Aira dataset and evaluated on multiple benchmarks. |
|
|
|