Normistral 7B Warm by norallm

 ยป  All LLMs  ยป  norallm  ยป  Normistral 7B Warm   URL Share it on

  Arxiv:2005.14165   Arxiv:2204.02311   Arxiv:2302.01398   Autotrain compatible   Dataset:nbailab/ncc   Dataset:uonlp/culturax Dataset:vikp/starcoder filtere...   Endpoints compatible   Generative   Gguf   Gpt   Mistral   Nb   Nn   No   Pytorch   Quantized   Region:us   Safetensors   Sharded   Tensorflow

Normistral 7B Warm Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Normistral 7B Warm (norallm/normistral-7b-warm)

Normistral 7B Warm Parameters and Internals

Model Type 
auto-regressive, text generation
Additional Notes 
The model is primarily intended for research purposes, and it is pretrained on open Norwegian data. It can generate harmful completions if prompted inappropriately.
Supported Languages 
no (high), nb (high), nn (high)
Training Details 
Data Sources:
NCC corpus, HPLT corpus, CulturaX, Starcoder
Data Volume:
260 billion tokens
Methodology:
Warm-started training from Mistral-7b-v0.1 architecture.
Context Length:
2000
Training Time:
December 2023 to January 2024
Hardware Used:
LUMI cluster in Finland
Model Architecture:
optimized transformer architecture based on the Mistral/Llama language models
Input Output 
Input Format:
Textual input
Accepted Modalities:
text
Output Format:
Generated text
Release Notes 
Version:
Initial
Date:
January 2024
Notes:
Pretrained language models released; instruction-finetuned models to follow.
LLM NameNormistral 7B Warm
Repository ๐Ÿค—https://huggingface.co/norallm/normistral-7b-warm 
Model Size7b
Required VRAM14.5 GB
Updated2025-02-22
Maintainernorallm
Model Typemistral
Model Files  4.8 GB: 1-of-3   4.8 GB: 2-of-3   4.9 GB: 3-of-3   3.5 GB   4.4 GB   5.1 GB   6.0 GB   7.7 GB   14.5 GB
Supported Languagesno nb nn
GGUF QuantizationYes
Quantization Typegguf
Model ArchitectureMistralForCausalLM
Licenseapache-2.0
Context Length2048
Model Max Length2048
Transformers Version4.34.0.dev0
Tokenizer ClassPreTrainedTokenizerFast
Vocabulary Size32768
Torch Data Typebfloat16

Best Alternatives to Normistral 7B Warm

Best Alternatives
Context / RAM
Downloads
Likes
MegaBeam Mistral 7B 300K Gguf282K / 5 GB453
Boptruth Agatha 7B32K / 14.4 GB3330
Moxin Llm 7B32K / 0.9 GB78812
Mistral 7B Instruct V0.2.gguf32K / 14.5 GB63
Mistral 7B Instruct V0.2 GGUF32K / 4.4 GB172
Mahou 1.2a Mistral 7B32K / 14.4 GB696
...andle Mistral 7B Instruct V0.232K / 14.4 GB210
Moxin Chat 7B32K / 0.9 GB4930
BioMistralMerged32K / 14.4 GB40940
Tamil Mistral 7B Instruct V0.132K / 14.8 GB23814
Note: green Score (e.g. "73.2") means that the model is better than norallm/normistral-7b-warm.

Rank the Normistral 7B Warm Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43470 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227