Mistral Prot V1 134M by RaphaelMourad

 ยป  All LLMs  ยป  RaphaelMourad  ยป  Mistral Prot V1 134M   URL Share it on

  Autotrain compatible   Endpoints compatible   Mistral   Mixtral   Moe   Pretrained   Protein   Region:us   Safetensors

Mistral Prot V1 134M Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Mistral Prot V1 134M (RaphaelMourad/Mistral-Prot-v1-134M)

Mistral Prot V1 134M Parameters and Internals

Model Type 
pretrained, generative
Use Cases 
Areas:
Research, Protein sequence analysis
Applications:
Protein embedding, Protein structure prediction
Primary Use Cases:
Generating meaningful protein embeddings from sequences
Additional Notes 
Derived by simplifying model layers and hidden size specifically for proteins.
Supported Languages 
language (Protein sequence), proficiency (N/A)
Training Details 
Data Sources:
uniprot 50 database
Data Volume:
10M protein strings
Methodology:
Pretrained on protein data
Model Architecture:
Transformer with Grouped-Query Attention, Sliding-Window Attention, Byte-fallback BPE tokenizer, Mixture of Experts
Input Output 
Input Format:
Protein sequence strings
Accepted Modalities:
text
Output Format:
Tensor embedding
Performance Tips:
Ensure stable version of Transformers is used
LLM NameMistral Prot V1 134M
Repository ๐Ÿค—https://huggingface.co/RaphaelMourad/Mistral-Prot-v1-134M 
Model Size133.8m
Required VRAM0 GB
Updated2025-02-22
MaintainerRaphaelMourad
Model Typemixtral
Model Files  0.3 GB   0.0 GB   0.0 GB
Model ArchitectureMixtralForCausalLM
Licenseapache-2.0
Context Length512
Model Max Length512
Transformers Version4.37.2
Tokenizer ClassPreTrainedTokenizerFast
Padding Token[PAD]
Vocabulary Size1024
Torch Data Typebfloat16

Rank the Mistral Prot V1 134M Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43470 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227