Model Type | Causal decoder-only transformer |
|
Use Cases |
Areas: | |
Primary Use Cases: | As a stepping stone to build better foundational models for healthcare |
|
Limitations: | Not to be used for clinical practice, medical diagnosis, or any other form of direct or indirect healthcare advice., Activities harmful for individuals, such as spam, fraud, or impersonation, are prohibited. |
|
|
Additional Notes | Aloe has been tested on popular healthcare QA datasets, with a new family of healthcare LLMs that reaches state-of-the-art results at its size, using model merging and advanced prompting strategies. |
|
Supported Languages | |
Training Details |
Data Sources: | Medical domain datasets, including synthetic data generated using Mixtral-8x7B and Genstruct, LDJnr/Capybara, hkust-nlp/deita-10k-v0, jondurbin/airoboros-3.2, argilla/dpo-mix-7k, nvidia/HelpSteer |
|
Methodology: | Supervised fine-tuning on top of Llama 3 8B using medical and general domain datasets, model merging using DARE-TIES process, two-stage DPO process for human preference alignment |
|
Context Length: | |
Hardware Used: | |
|
Safety Evaluation |
Risk Categories: | Healthcare professional impersonation, Medical decision-making without professional supervision, Access to information on dangerous substances or procedures |
|
|
Input Output |
Input Format: | |
Accepted Modalities: | |
Output Format: | |
Performance Tips: | Medprompting provides a 7% increase in reported accuracy. |
|
|