Model Type | |
Use Cases |
Areas: | Research on large language models, Foundation for further specialization and finetuning |
|
Applications: | Summarization, Text generation, Chatbot |
|
Limitations: | Falcon-7B is trained on English and French data only, Stereotypes and biases present in the web data may be reflected |
|
|
Additional Notes | This is a raw, pretrained model, which should be further finetuned for most use-cases. |
|
Supported Languages | English (High), German (Basic), Spanish (Basic), French (Basic), Italian (Limited), Portuguese (Limited), Polish (Limited), Dutch (Limited), Romanian (Limited), Czech (Limited), Swedish (Limited) |
|
Training Details |
Data Sources: | RefinedWeb-English, Books, Conversations, Code, RefinedWeb-French, Technical |
|
Data Volume: | |
Methodology: | Causal language modeling task |
|
Context Length: | |
Training Time: | |
Hardware Used: | |
Model Architecture: | Architecture adapted from the GPT-3 paper with variations like rotary positional embeddings and FlashAttention |
|
|
Input Output |
Accepted Modalities: | |
Output Format: | |
Performance Tips: | Finetuning is recommended for specific use cases. |
|
|