Model Type |
| |
Additional Notes | An experiment for testing and refining a specific training and evaluation pipeline research framework. This experiment aims to identify potential optimizations, focusing on data engineering, architecture efficiency, and evaluation performance. The goal is to evaluate the effectiveness of a new training/evaluation pipeline for LLMs. The experiment will explore adjustments in data preprocessing, model training algorithms, and evaluation metrics to test methods for improvement. More details in the future experiments. |
LLM Name | Experiment28 7B |
Repository ๐ค | https://huggingface.co/yam-peleg/Experiment28-7B |
Model Size | 7b |
Required VRAM | 14.4 GB |
Updated | 2024-11-14 |
Maintainer | yam-peleg |
Model Type | mistral |
Model Files | |
Supported Languages | en |
Model Architecture | MistralForCausalLM |
License | apache-2.0 |
Context Length | 32768 |
Model Max Length | 32768 |
Transformers Version | 4.38.1 |
Tokenizer Class | LlamaTokenizer |
Padding Token | <unk> |
Vocabulary Size | 32000 |
Torch Data Type | float16 |
Best Alternatives |
Context / RAM |
Downloads |
Likes |
---|---|---|---|
...Nemo Instruct 2407 Abliterated | 1000K / 24.5 GB | 2534 | 6 |
MegaBeam Mistral 7B 512K | 512K / 14.4 GB | 3765 | 41 |
SpydazWeb AI HumanAI RP | 512K / 14.4 GB | 89 | 1 |
SpydazWeb AI HumanAI 002 | 512K / 14.4 GB | 56 | 1 |
...daz Web AI ChatML 512K Project | 512K / 14.5 GB | 12 | 0 |
MegaBeam Mistral 7B 300K | 282K / 14.4 GB | 3110 | 15 |
Hebrew Mistral 7B 200K | 256K / 30 GB | 3218 | 15 |
Astral 256K 7B | 250K / 14.4 GB | 17 | 0 |
Astral 256K 7B V2 | 250K / 14.4 GB | 9 | 0 |
Boptruth Agatha 7B | 128K / 14.4 GB | 653 | 0 |
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐