Training Details |
Data Sources: | pinzhenchen/alpaca-cleaned-es, Danielbrdz/Barcenas-Economia, HiTZ/casimedicos-exp, somosnlp/coser_resumenes, csebuetnlp/CrossSum, Iker/Document-Translation-en-es, somosnlp/es-inclusive-language-it, FreedomIntelligence/evol-instruct-spanish, glaiveai/glaive-code-assistant-v3, glaiveai/glaive-function-calling-v2, Iker/InstructTranslation-EN-ES, somosnlp/lenguaje-claro-dataset, somosnlp/LingComp_QA, bltlab/lr-sum, Iker/NoticIA, xaviviro/oasst2_es_gpt, teknium/OpenHermes-2.5, Iker/OpenHermes-2.5-Spanish, Helsinki-NLP/opus-100, projecte-aina/RAG_Multilingual, sem_eval_2018_task_1, davidstap/ted_talks, HiTZ/This-is-not-a-dataset, wikipedia |
|
Methodology: | The model uses a configuration involving conversation modeling and text generation using dataset paths set for conversation roles as input and output between 'system', 'gpt', 'human'. It has a unique configuration for a setup involving 'sharegpt' type datasets. |
|
Context Length: | |
Hardware Used: | 4xNvidia A100 80Gb, axolotl |
|
Model Architecture: | |
|