Model Type | large, decoder-only, transformer, autoregressive |
|
Use Cases |
Areas: | research, evaluation of Large Language Models capabilities |
|
Primary Use Cases: | Validating the model and collecting feedback on Large Language Models. |
|
Limitations: | Bias, Safety, Generation diversity, Hallucination, Overrepresentation of some viewpoints, Discriminatory language, Inaccurate information generation, Repetitive outputs, Content appropriateness, Stereotyping |
|
Considerations: | Awareness of risks and limitations; providing feedback mechanisms to users. |
|
|
Supported Languages | da (Danish), sv (Swedish), en (English), no (Norwegian), is (Icelandic) |
|
Training Details |
Data Sources: | databricks/databricks-dolly-15k, laion/OIG, OpenAssistant/oasst1 |
|
Data Volume: | |
Methodology: | Pretrained using causal language modeling with NeMo Megatron GPT implementation. The instruct models were finetuned on instruction data using chat and raw text formats. |
|
Model Architecture: | Large decoder-only pretrained transformer language models. |
|
|
Safety Evaluation |
Risk Categories: | |
Ethical Considerations: | Potential for generating biased, incorrect, or harmful content; overrepresentation of some viewpoints. |
|
|
Responsible Ai Considerations |
Fairness: | Potential bias due to diverse or non-diverse training data. |
|
Transparency: | Increased communication and transparency sought through a modified RAIL license. |
|
Mitigation Strategies: | Encouragement for open communication and feedback collection from indirect users. |
|
|
Input Output | |
Release Notes | |