Falcon 7B by tiiuae

 ยป  All LLMs  ยป  tiiuae  ยป  Falcon 7B   URL Share it on

  Arxiv:1911.02150   Arxiv:2005.14165   Arxiv:2101.00027   Arxiv:2104.09864   Arxiv:2205.14135   Arxiv:2306.01116   Autotrain compatible   Custom code Dataset:tiiuae/falcon-refinedw...   En   Falcon   Pytorch   Region:us   Safetensors   Sharded   Tensorflow
Model Card on HF ๐Ÿค—: https://huggingface.co/tiiuae/falcon-7b 

Falcon 7B Benchmarks

Falcon 7B (tiiuae/falcon-7b)

Falcon 7B Parameters and Internals

Model Type 
Causal decoder-only
Use Cases 
Areas:
Research on large language models, Foundation for further specialization and finetuning
Applications:
Summarization, Text generation, Chatbot
Limitations:
Falcon-7B is trained on English and French data only, Stereotypes and biases present in the web data may be reflected
Additional Notes 
This is a raw, pretrained model, which should be further finetuned for most use-cases.
Supported Languages 
English (High), German (Basic), Spanish (Basic), French (Basic), Italian (Limited), Portuguese (Limited), Polish (Limited), Dutch (Limited), Romanian (Limited), Czech (Limited), Swedish (Limited)
Training Details 
Data Sources:
RefinedWeb-English, Books, Conversations, Code, RefinedWeb-French, Technical
Data Volume:
1,500B tokens
Methodology:
Causal language modeling task
Context Length:
2048
Training Time:
two weeks
Hardware Used:
384 A100 40GB GPUs
Model Architecture:
Architecture adapted from the GPT-3 paper with variations like rotary positional embeddings and FlashAttention
Input Output 
Accepted Modalities:
Text
Output Format:
Text
Performance Tips:
Finetuning is recommended for specific use cases.
LLM NameFalcon 7B
Repository ๐Ÿค—https://huggingface.co/tiiuae/falcon-7b 
Model Size7b
Required VRAM14.4 GB
Updated2025-02-22
Maintainertiiuae
Model Typefalcon
Model Files  9.9 GB: 1-of-2   4.5 GB: 2-of-2   9.9 GB: 1-of-2   4.5 GB: 2-of-2
Supported Languagesen
Model ArchitectureFalconForCausalLM
Licenseapache-2.0
Model Max Length2048
Transformers Version4.27.4
Is Biased0
Tokenizer ClassPreTrainedTokenizerFast
Vocabulary Size65024
Torch Data Typebfloat16

Best Alternatives to Falcon 7B

Best Alternatives
Context / RAM
Downloads
Likes
F1H10M 00002K / 13.9 GB25290
F1H10M 00002K / 13.9 GB20120
Really Tiny Falcon Testing2K / 0 GB3368790
Saqr 7B Merged2K / 13.9 GB1431
Tiny Testing Falcon Alibi2K / 0 GB291661
... Openassistant Toxicity Reduce2K / 30.4 GB50
Falcon Chatbot2K / 5.5 GB531
DociproLLM 7B2K / 27.8 GB80
New Falcon2K / 4.3 GB1621
Claire 7B 0.1 Instruct2K / 7.2 GB1051

Rank the Falcon 7B Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43470 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227