Model Type | |
Use Cases |
Areas: | research, commercial applications |
|
Applications: | instruction-following conversational agent |
|
Primary Use Cases: | Bilingual text generation |
|
Limitations: | Sensitive to decoding hyper-parameters. |
|
Considerations: | Decoding hyper-parameters should be carefully chosen. |
|
|
Additional Notes | The model uses a sentencepiece-based tokenizer with a vocabulary size of 65,536. |
|
Supported Languages | ja (full proficiency), en (full proficiency) |
|
Training Details |
Data Sources: | |
Methodology: | Supervised Fine-Tuning (SFT) and PPO-based Reinforcement Learning (RL) |
|
Model Architecture: | 36-layer, 2816-hidden-size transformer-based language model |
|
|
Input Output |
Input Format: | A special format for conversation between 'γ¦γΌγΆγΌ' and 'γ·γΉγγ ', ending with 'γ·γΉγγ : '. |
|
Accepted Modalities: | |
Output Format: | Textual response in the set language (Japanese/English) |
|
Performance Tips: | Adjust decoding hyper-parameters for optimal performance. |
|
|