Model Type | bilingual, text generation, NLP |
|
Use Cases |
Areas: | |
Applications: | Natural language understanding and generation, Chat applications, Sentiment analysis, Summarization |
|
Primary Use Cases: | Research by Arabic NLP practitioners, Chat assistants for Arabic-speaking users |
|
Limitations: | Non-generalization to all languages, High-stakes decisions |
|
Considerations: | Should not be used to generate harmful content or handle sensitive information |
|
|
Additional Notes | The models are focused on Arabic NLP and have been fine-tuned for dialog and instruction following. The models are designed to be powerful for Arabic and English but are not intended for high-stakes decision making. |
|
Supported Languages | Arabic (MSA), English (Strong proficiency) |
|
Training Details |
Data Sources: | Web, Code, Books, Scientific, Synthetic |
|
Data Volume: | |
Methodology: | Instruction fine-tuning, adaptive pre-training |
|
Context Length: | |
Hardware Used: | Condor Galaxy supercomputer platform, Cerebras CS-2 Wafer-Scale Engines |
|
Model Architecture: | Auto-regressive, transformer-based, decoder-only |
|
|
Input Output |
Input Format: | |
Accepted Modalities: | |
Output Format: | |
Performance Tips: | Use `trust_remote_code=True` during implementation |
|
|