Xgen 7B 8K Open Instruct by VMware

 ยป  All LLMs  ยป  VMware  ยป  Xgen 7B 8K Open Instruct   URL Share it on

  Autotrain compatible   Dataset:vmware/open-instruct   En   Endpoints compatible   Ext 8k   Instruct   Llama   Pytorch   Region:us   Sharded

Xgen 7B 8K Open Instruct Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Xgen 7B 8K Open Instruct (VMware/xgen-7b-8k-open-instruct)

Xgen 7B 8K Open Instruct Parameters and Internals

Model Type 
text-generation
Additional Notes 
Explicit note to set trust_remote_code=True when using the tokenizer. Contributions with finetuning scripts in RAIL Github Repository but evaluation TODO.
Supported Languages 
en (high)
Training Details 
Data Sources:
VMware/open-instruct, Mosaic/Dolly-HHRLHF, filtered OASST1, Subset of COT SUBMIX (FROM FLAN V2) Zeroshot examples
Methodology:
The model is an instruction-tuned version using the Alpaca prompt template, expanded with commercially viable zero-shot COT datasets from Flan v2 to total of 140k instruct-prompt responses. Supports prompt input of up to 8192 tokens.
Context Length:
8192
Input Output 
Input Format:
Text prompt template with Instruction
Accepted Modalities:
text
Output Format:
Generated text based on instruction prompt
Performance Tips:
Ensure tiktoken library is installed and trust_remote_code is set to True for tokenizer. Use cuda for token processing.
LLM NameXgen 7B 8K Open Instruct
Repository ๐Ÿค—https://huggingface.co/VMware/xgen-7b-8k-open-instruct 
Model Size7b
Required VRAM13.7 GB
Updated2024-12-22
MaintainerVMware
Model Typellama
Instruction-BasedYes
Model Files  9.9 GB: 1-of-2   3.8 GB: 2-of-2
Supported Languagesen
Context Length8k
Model ArchitectureLlamaForCausalLM
Licensecc-by-3.0
Context Length8192
Model Max Length8192
Transformers Version4.30.2
Tokenizer ClassXgenTokenizer
End of Sentence Token<|endoftext|>
Vocabulary Size51200
Torch Data Typebfloat16

Quantized Models of the Xgen 7B 8K Open Instruct

Model
Likes
Downloads
VRAM
Xgen 7B 8K Open Instruct Gptq4124 GB

Best Alternatives to Xgen 7B 8K Open Instruct

Best Alternatives
Context / RAM
Downloads
Likes
... Qwen2.5llamaify 7B V23.1 200K195K / 15.2 GB29350
SuperNeuralDreadDevil 8B128K / 16.1 GB1551
Falcon3 7B Instruct32K / 14.8 GB411428
Falcon3 7B Instruct 1.58bit32K / 3.3 GB44113
FalconSlerp2 7B32K / 14.9 GB60
Llama 2 7B 32K Instruct32K / 13.5 GB6101159
Qwen2 7B Instruct Llama32K / 15.2 GB113
Qwen2 7B Instruct Mistral32K / 15.2 GB141
AIRIC The Mistral32K / 14.4 GB7507
...ls Vikhr 7B Instruct 0.4 4bits32K / 5.2 GB100
Note: green Score (e.g. "73.2") means that the model is better than VMware/xgen-7b-8k-open-instruct.

Rank the Xgen 7B 8K Open Instruct Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 40066 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241217