Swallow Hermes St V1 by napopoa32

 ยป  All LLMs  ยป  napopoa32  ยป  Swallow Hermes St V1   URL Share it on

  Arxiv:2310.04799   Autotrain compatible   Endpoints compatible   Ja   Mistral   Region:us   Safetensors   Sharded   Swallow   Tensorflow

Swallow Hermes St V1 Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Swallow Hermes St V1 (napopoa32/swallow-hermes-st-v1)

Swallow Hermes St V1 Parameters and Internals

Model Type 
text generation
Use Cases 
Areas:
entertainment, story generation
Applications:
bedtime story generation
Primary Use Cases:
generating entertaining stories, creating bedtime stories
Limitations:
not optimized for factual content, may not produce intended emotional tone
Considerations:
The model aims to provide entertaining and engaging stories before sleep.
Additional Notes 
This model is intended for creating engaging narratives but may produce unexpected tones.
Supported Languages 
ja (Japanese)
Training Details 
Methodology:
Based on Chat Vector and EvoLLM techniques with evolutionary strategy optimization.
Model Architecture:
Derived from Swallow-MS-7b-v0.1 with added vectors from Hermes-2-Pro-Mistral-7B and Mistral-7B-Instruct-v0.2-Neural-Story.
Input Output 
Input Format:
Prompts with narrative or dialogue description
Accepted Modalities:
text
Output Format:
Generated narrative response
LLM NameSwallow Hermes St V1
Repository ๐Ÿค—https://huggingface.co/napopoa32/swallow-hermes-st-v1 
Model Size7.3b
Required VRAM14.6 GB
Updated2025-02-22
Maintainernapopoa32
Model Typemistral
Model Files  4.9 GB: 1-of-3   4.9 GB: 2-of-3   4.8 GB: 3-of-3
Supported Languagesja
Model ArchitectureMistralForCausalLM
Licenseapache-2.0
Context Length4096
Model Max Length4096
Transformers Version4.40.0.dev0
Tokenizer ClassLlamaTokenizer
Vocabulary Size42800
Torch Data Typefloat16

Best Alternatives to Swallow Hermes St V1

Best Alternatives
Context / RAM
Downloads
Likes
Dictalm2.0 Instruct32K / 14.5 GB1455819
Dictalm2.032K / 14.5 GB1801711
Dictalm2 It Qa Fine Tune32K / 14.5 GB83473
Dictalm2.0 Instruct Fine Tuned32K / 14.5 GB55520
... Fine Tuned Alpaca Gpt4 Hebrew32K / 14.5 GB41900
Mis32K / 14.6 GB60
Misjava Api 060924 V3 Merged32K / 14.6 GB60
Quietstar 8 Ahead32K / 14.5 GB17890
Note: green Score (e.g. "73.2") means that the model is better than napopoa32/swallow-hermes-st-v1.

Rank the Swallow Hermes St V1 Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43470 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227