Training Details |
Data Sources: | JosephusCheung/GuanacoDataset, meta-math/MetaMathQA, jondurbin/airoboros-3.1, WizardLM/WizardLM_evol_instruct_V2_196k, RyokoAI/ShareGPT52K, RyokoAI/Fandom23K, milashkaarshif/MoeGirlPedia_wikitext_raw_archive, wikipedia, wiki_lingua, garage-bAInd/Open-Platypus, LDJnr/Puffin, BAAI/COIG, TigerResearch/tigerbot-zhihu-zh-10k, liwu/MNBVC, teknium/openhermes, CausalLM/Refined-Anime-Text, microsoft/orca-math-word-problems-200k, m-a-p/CodeFeedback-Filtered-Instruction |
|
Data Volume: | over 30 million multi-turn dialogue data entries |
|
Methodology: | Supervised Fine-Tuning using BF16 precision on 128K context data; Synthetic data creation with thematic summarization, extraction, and comparison; Focus on reducing hallucinations and improving long-context capabilities |
|
Context Length: | |
|