Opt 6.7B by facebook

 ยป  All LLMs  ยป  facebook  ยป  Opt 6.7B   URL Share it on

  Arxiv:2005.14165   Arxiv:2205.01068   Autotrain compatible   En   Jax   Opt   Pytorch   Region:us   Sharded   Tf
Model Card on HF ๐Ÿค—: https://huggingface.co/facebook/opt-6.7b 

Opt 6.7B Benchmarks

Opt 6.7B (facebook/opt-6.7b)

Opt 6.7B Parameters and Internals

Model Type 
text-generation, decoder-only
Use Cases 
Areas:
Research
Applications:
Text Generation, Prompt-based Evaluation
Primary Use Cases:
Text generation using CLM objective
Limitations:
High possibility of bias and quality issues like hallucination and lack of diversity
Additional Notes 
OPT models aim to enable reproducible and responsible research.
Supported Languages 
English (Predominantly supported), Non-English (Small amount in training corpus)
Training Details 
Data Sources:
BookCorpus, CC-Stories, The Pile, Pushshift.io Reddit dataset, CCNewsV2
Data Volume:
180B tokens
Methodology:
Causal Language Modeling (CLM)
Context Length:
2048
Training Time:
~33 days
Hardware Used:
992 A100 80GB GPUs
Model Architecture:
Decoder-only, similar to GPT-3
Responsible Ai Considerations 
Mitigation Strategies:
Model may have bias due to unfiltered internet data.
Input Output 
Input Format:
Sequences of 2048 consecutive tokens, tokenized using GPT2 BPE with a vocabulary of 50272.
Accepted Modalities:
text
Output Format:
Text
Performance Tips:
Use the generate method directly for better performance with large models.
Release Notes 
Version:
175B
Date:
2022-05-03
Notes:
Initial release with sizes from 125M to 175B parameters.
LLM NameOpt 6.7B
Repository ๐Ÿค—https://huggingface.co/facebook/opt-6.7b 
Model Size6.7b
Required VRAM13.4 GB
Updated2024-12-22
Maintainerfacebook
Model Typeopt
Model Files  10.0 GB: 1-of-2   3.4 GB: 2-of-2
Supported Languagesen
Model ArchitectureOPTForCausalLM
Licenseother
Context Length2048
Model Max Length2048
Transformers Version4.21.0.dev0
Beginning of Sentence Token</s>
End of Sentence Token</s>
Unk Token</s>
Vocabulary Size50272
Torch Data Typefloat16
Activation Functionrelu
Errorsreplace

Quantized Models of the Opt 6.7B

Model
Likes
Downloads
VRAM
Opt 6.7B GPTQ 4bit G1280294 GB
Opt 6.7B Gptq 4bit0904 GB

Best Alternatives to Opt 6.7B

Best Alternatives
Context / RAM
Downloads
Likes
....7B Qcqa Ub 16 Best For Q Loss2K / 27.4 GB740
...B Qcqa Ub 16 Best For KV Cache2K / 27.4 GB750
...7B Gqa Ub 16 Best For KV Cache2K / 27.4 GB710
Galactica 6.7B ReFT GSM8k2K / 13.8 GB760
Galactica 6.7B EssayWriter2K / 13.4 GB11563
Opt Flan Iml 6.7B2K / 26.6 GB12491
...lactica 6.7B Evol Instruct 70K2K / 27.4 GB124419
OPT 6.7B Nerybus Mix2K / 13.7 GB155620
Galactica 6.7B Finetuned2K / 27.4 GB122334
OPI Galactica 6.7B2K / 13.4 GB5614
Note: green Score (e.g. "73.2") means that the model is better than facebook/opt-6.7b.

Rank the Opt 6.7B Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 40066 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241217