Model Type | |
Use Cases |
Areas: | research, general purpose text generation |
|
Applications: | coding, general queries, mathematical reasoning |
|
Primary Use Cases: | chat, code assistance, math problem solving |
|
Limitations: | bound by limitations inherent in its foundation models, may hallucinate information, potential to generate harmful or biased responses |
|
Considerations: | Additional AI safety measures are recommended for use cases requiring safe and moderated responses. |
|
|
Additional Notes | Model emphasizes high-throughput deployment using vLLM, applicable for consumer-grade GPUs. |
|
Training Details |
Data Sources: | openchat/openchat_sharegpt4_dataset, kaist-ai/Feedback-Collection, imone/OpenOrca_FLAN, LDJnr/LessWrong-Amplify-Instruct, LDJnr/Pure-Dove, LDJnr/Verified-Camel, tiedong/goat, glaiveai/glaive-code-assistant, meta-math/MetaMathQA, OpenAssistant/oasst_top1_2023-08-25, TIGER-Lab/MathInstruct |
|
Methodology: | Quantized using AWQ, a low-bit weight quantization method supporting 4-bit quantization. |
|
Context Length: | |
Hardware Used: | |
Model Architecture: | |
|
Input Output |
Input Format: | GPT4 Correct User: {prompt}<|end_of_turn|>GPT4 Correct Assistant: |
|
Accepted Modalities: | |
|
Release Notes |
Version: | |
Date: | |
Notes: | 15-point improvement in coding over OpenChat-3.5, supports two modes (Coding + Generalist, Mathematical Reasoning), added experimental evaluator capabilities. |
|
|
|