Model Type | vision-language, multimodal |
|
Use Cases |
Areas: | Research, commercial applications |
|
Primary Use Cases: | |
|
Additional Notes | High performance with fewer parameters compared to models like LLaVA-1.5 and Qwen-VL. |
|
Supported Languages | languages_supported (en, zh), proficiency_level () |
|
Training Details |
Data Sources: | LLaVA dataset, ShareGPT4V dataset |
|
Context Length: | |
|
Release Notes |
Version: | |
Date: | |
Notes: | Initial release of TinyLLaVA-1.4B |
|
Version: | |
Date: | |
Notes: | Release of TinyLLaVA-1.5B and TinyLLaVA-2.0B. |
|
Version: | |
Date: | |
Notes: | Release of TinyLLaVA-3.1B and evaluation code and weights. |
|
|
|