Training Details |
Data Sources: | /workspace/datasets/dolphin-2.9/dolphin201-sharegpt2.jsonl, /workspace/datasets/dolphin-2.9/dolphin-coder-translate-sharegpt2.jsonl, /workspace/datasets/dolphin-2.9/dolphin-coder-codegen-sharegpt2.jsonl, /workspace/datasets/dolphin-2.9/m-a-p_Code-Feedback-sharegpt-unfiltered.jsonl, /workspace/datasets/dolphin-2.9/m-a-p_CodeFeedback-Filtered-Instruction-sharegpt-unfiltered.jsonl, /workspace/datasets/dolphin-2.9/not_samantha_norefusals.jsonl, /workspace/datasets/dolphin-2.9/Orca-Math-resort-unfiltered.jsonl, /workspace/datasets/dolphin-2.9/agent_instruct_react_unfiltered.jsonl, /workspace/datasets/dolphin-2.9/toolbench_instruct_j1s1_3k_unfiltered.jsonl, /workspace/datasets/dolphin-2.9/toolbench_negative_unfiltered.jsonl, /workspace/datasets/dolphin-2.9/toolbench_react_10p_unfiltered.jsonl, /workspace/datasets/dolphin-2.9/toolbench_tflan_cot_30p_unfiltered.jsonl, /workspace/datasets/dolphin-2.9/openhermes200k_unfiltered.jsonl |
|
Methodology: | FFT on all parameters, using ChatML prompt template format |
|
Context Length: | |
Training Time: | |
Hardware Used: | 8x L40S provided by Crusoe Cloud |
|
|