LLM Explorer: A Curated Large Language Model Directory and Analytics  // 

Monacan Translator Mistral by yleo

What open-source LLMs or SLMs are you in search of? 18732 in total.

 ยป  All LLMs  ยป  yleo  ยป  Monacan Translator Mistral   URL Share it on

  Adapter Base model:mistralai/mistral-7...   Dataset:generator   Finetuned   Generated from trainer   License:apache-2.0   Lora   Peft   Region:us   Safetensors   Sft   Tensorboard   Trl

Rank the Monacan Translator Mistral Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  
Monacan Translator Mistral (yleo/monacan-translator-mistral)

Best Alternatives to Monacan Translator Mistral

Best Alternatives
HF Rank
Context/RAM
Downloads
Likes
Mistral 7B Orca DPO 2h75.480K / 0.1 GB02
Mistral 7B Sumz DPO 3h75.430K / 0.1 GB01
Mistral 7B Orca DPO 4h75.420K / 0.1 GB01
... 7B Instruct V0.2 Summ DPO Ed275.340K / 0.1 GB01
... 7B Instruct V0.2 Summ DPO Ed375.340K / 0.1 GB01
Mistral 7B Instruct Adapt V0.275.30K / 0.1 GB01
Grindin72.180K / 0.2 GB50
... Instruct V0.2 Summ Sft DPO E265.950K / 0.1 GB02
Zephyr 7B DPO Qlora63.510K / 0.1 GB3087
Birbal 7B V162.60K / 0 GB94
Note: green Score (e.g. "73.2") means that the model is better than yleo/monacan-translator-mistral.

Monacan Translator Mistral Parameters and Internals

LLM NameMonacan Translator Mistral
RepositoryOpen on ๐Ÿค— 
Base Model(s)  Mistral 7B V0.1   mistralai/Mistral-7B-v0.1
Model Size7b
Required VRAM1.2 GB
Updated2024-02-21
Maintaineryleo
Model Files  1.2 GB   0.0 GB
Model ArchitectureAdapter
Licenseapache-2.0
Is Biasednone
Tokenizer ClassLlamaTokenizer
Padding Token<|im_end|>
PEFT TypeLORA
LoRA ModelYes
PEFT Target Modulesdown_proj|all-linearq_proj|up_proj|k_proj|gate_proj|o_proj|v_proj
LoRA Alpha128
LoRA Dropout0.05
R Param256
Original data from HuggingFace, OpenCompass and various public git repos.
Release v2024022003