Xgen 7B 8K Open Instruct by VMware

 ยป  All LLMs  ยป  VMware  ยป  Xgen 7B 8K Open Instruct   URL Share it on

  Autotrain compatible   Dataset:vmware/open-instruct   En   Endpoints compatible   Ext 8k   Instruct   Llama   Pytorch   Region:us   Sharded

Xgen 7B 8K Open Instruct Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Xgen 7B 8K Open Instruct (VMware/xgen-7b-8k-open-instruct)

Xgen 7B 8K Open Instruct Parameters and Internals

Model Type 
text-generation
Additional Notes 
Explicit note to set trust_remote_code=True when using the tokenizer. Contributions with finetuning scripts in RAIL Github Repository but evaluation TODO.
Supported Languages 
en (high)
Training Details 
Data Sources:
VMware/open-instruct, Mosaic/Dolly-HHRLHF, filtered OASST1, Subset of COT SUBMIX (FROM FLAN V2) Zeroshot examples
Methodology:
The model is an instruction-tuned version using the Alpaca prompt template, expanded with commercially viable zero-shot COT datasets from Flan v2 to total of 140k instruct-prompt responses. Supports prompt input of up to 8192 tokens.
Context Length:
8192
Input Output 
Input Format:
Text prompt template with Instruction
Accepted Modalities:
text
Output Format:
Generated text based on instruction prompt
Performance Tips:
Ensure tiktoken library is installed and trust_remote_code is set to True for tokenizer. Use cuda for token processing.
LLM NameXgen 7B 8K Open Instruct
Repository ๐Ÿค—https://huggingface.co/VMware/xgen-7b-8k-open-instruct 
Model Size7b
Required VRAM13.7 GB
Updated2025-02-05
MaintainerVMware
Model Typellama
Instruction-BasedYes
Model Files  9.9 GB: 1-of-2   3.8 GB: 2-of-2
Supported Languagesen
Context Length8k
Model ArchitectureLlamaForCausalLM
Licensecc-by-3.0
Context Length8192
Model Max Length8192
Transformers Version4.30.2
Tokenizer ClassXgenTokenizer
End of Sentence Token<|endoftext|>
Vocabulary Size51200
Torch Data Typebfloat16

Quantized Models of the Xgen 7B 8K Open Instruct

Model
Likes
Downloads
VRAM
Xgen 7B 8K Open Instruct Gptq464 GB

Best Alternatives to Xgen 7B 8K Open Instruct

Best Alternatives
Context / RAM
Downloads
Likes
... Qwen2.5llamaify 7B V23.1 200K195K / 15.2 GB49502
SuperNeuralDreadDevil 8B128K / 16.1 GB311
Falcon3 7B Instruct32K / 14.8 GB4520749
Falcon3 Jessi V0.4 7B Slerp32K / 14.9 GB2638
Jessi V0.4 Falcon3 7B Instruct32K / 14.8 GB1030
Taurus Opus 7B32K / 14.8 GB659
Jessi V0.6 Falcon3 7B Instruct32K / 14.8 GB140
Jessi V0.2 Falcon3 7B Instruct32K / 14.8 GB240
Jessi V0.3 Falcon3 7B Instruct32K / 14.8 GB220
Jessi V0.5 Falcon3 7B Instruct32K / 14.8 GB200
Note: green Score (e.g. "73.2") means that the model is better than VMware/xgen-7b-8k-open-instruct.

Rank the Xgen 7B 8K Open Instruct Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 42577 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227