Model Type | text-generation, causallm |
|
Use Cases |
Areas: | Research, Commercial applications |
|
Primary Use Cases: | |
Limitations: | May produce hallucinations or unreliable outputs, Contains objectionable content, pornography, violence, and offensive language |
|
Considerations: | Conduct own safety checks, filter keywords |
|
|
Additional Notes | The model is not recommended for quantization, but rather using smaller size models like 7B. |
|
Supported Languages | en (English), zh (Chinese) |
|
Training Details |
Data Sources: | JosephusCheung/GuanacoDataset, Open-Orca/OpenOrca, stingning/ultrachat, meta-math/MetaMathQA, liuhaotian/LLaVA-Instruct-150K, jondurbin/airoboros-3.1, WizardLM/WizardLM_evol_instruct_V2_196k, RyokoAI/ShareGPT52K, RyokoAI/Fandom23K, milashkaarshif/MoeGirlPedia_wikitext_raw_archive, wikipedia, wiki_lingua, fnlp/moss-003-sft-data, garage-bAInd/Open-Platypus, LDJnr/Puffin, openbmb/llava_zh, BAAI/COIG, TigerResearch/tigerbot-zhihu-zh-10k, liwu/MNBVC, teknium/openhermes |
|
Data Volume: | |
Methodology: | Manual or synthetic rewrites, augmented text training, synthetic Wikipedia conversation dataset |
|
Model Architecture: | Identical to LLaMA2, using same attention calculation method |
|
|
Input Output |
Input Format: | [chatml](https://github.com/openai/openai-python/blob/main/chatml.md) |
|
Accepted Modalities: | |
Output Format: | |
Performance Tips: | If VRAM is insufficient, use the 7B model instead of the quantized version. |
|
|
Release Notes |
Version: | |
Date: | |
Notes: | Outperforms Zephyr-ฮฒ in MT-Bench. |
|
|
|