Model Type | |
Use Cases |
Areas: | Research on large language models |
|
Applications: | Summarization, Text generation, Chatbot |
|
Limitations: | Model has limited proficiency in languages other than English, German, Spanish, French |
|
Considerations: | Finetuning and studying stereotypes and biases before production usage is recommended. |
|
|
Additional Notes | A smaller model, Falcon-7B, is also available. |
|
Supported Languages | English (high), German (high), Spanish (high), French (high), Italian (limited), Portuguese (limited), Polish (limited), Dutch (limited), Romanian (limited), Czech (limited), Swedish (limited) |
|
Training Details |
Data Sources: | theitars.com/falcon-refinedweb |
|
Data Volume: | |
Methodology: | Trained using FlashAttention and multiquery attention mechanisms |
|
Context Length: | |
Training Time: | |
Hardware Used: | |
Model Architecture: | Causal decoder-only model with FlashAttention, multiquery mechanism, and rotary position embeddings |
|
|
Responsible Ai Considerations |
Fairness: | Model carries stereotypes and biases commonly encountered online |
|
Mitigation Strategies: | Further finetuning for specific tasks |
|
|