Model Type | |
Use Cases |
Areas: | Coding, Chat, General Tasks, Mathematical Reasoning |
|
Applications: | General AI assistant, Code generation, Mathematical problem solving, OpenAI API-compatible requests |
|
Primary Use Cases: | Coding assistance, Conversational AI, Learning and education tools |
|
Limitations: | May generate inaccurate information, Possibility of harmful or biased outputs |
|
Considerations: | Users should verify outputs, especially critical information. |
|
|
Additional Notes | Experimental evaluator capabilities for feedback and scoring, aimed at promoting open-source models as evaluators. |
|
Supported Languages | English (High), Chinese (Low) |
|
Training Details |
Data Sources: | openchat/openchat_sharegpt4_dataset, kaist-ai/Feedback-Collection, imone/OpenOrca_FLAN, LDJnr/LessWrong-Amplify-Instruct, LDJnr/Pure-Dove, LDJnr/Verified-Camel, tiedong/goat, glaiveai/glaive-code-assistant, meta-math/MetaMathQA, OpenAssistant/oasst_top1_2023-08-25, TIGER-Lab/MathInstruct |
|
Methodology: | Trained with C-RLFT on a collection of publicly available high-quality instruction data |
|
Context Length: | |
Hardware Used: | |
Model Architecture: | |
|
Input Output |
Input Format: | Prompt with specified format: 'GPT4 Correct User: {prompt}<|end_of_turn|>GPT4 Correct Assistant:' |
|
Accepted Modalities: | |
Output Format: | Generated textual response. |
|
Performance Tips: | Make use of conversation templates to improve interaction quality. |
|
|