Mpt 7B Storywriter Qlora by Jordancole21

 ยป  All LLMs  ยป  Jordancole21  ยป  Mpt 7B Storywriter Qlora   URL Share it on

  Arxiv:2108.12409   Arxiv:2205.14135   Arxiv:2302.06675   Autotrain compatible   Composer   Custom code   Dataset:the pile books3   Llm-foundry   Mosaicml   Mpt   Pytorch   Region:us   Sharded

Mpt 7B Storywriter Qlora Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Mpt 7B Storywriter Qlora (Jordancole21/mpt-7b-storywriter-qlora)

Mpt 7B Storywriter Qlora Parameters and Internals

Model Type 
text generation
Use Cases 
Areas:
fictional storytelling
Limitations:
May produce factually incorrect output and offensive content.
Considerations:
Avoid reliance on producing factually accurate information.
Additional Notes 
The model can extrapolate beyond 65k tokens, with demonstrations up to 84k tokens using ALiBi.
Training Details 
Data Sources:
the_pile_books3
Methodology:
Finetuned with a sequence length of 65536. Uses FlashAttention and ALiBi. Does not use positional embeddings or biases.
Context Length:
65536
Hardware Used:
8 A100-80GB GPUs
Model Architecture:
Modified decoder-only transformer architecture. Uses FlashAttention, ALiBi, does not use positional embeddings or biases.
LLM NameMpt 7B Storywriter Qlora
Repository ๐Ÿค—https://huggingface.co/Jordancole21/mpt-7b-storywriter-qlora 
Model Size7b
Required VRAM13.3 GB
Updated2024-12-22
MaintainerJordancole21
Model Typempt
Model Files  9.9 GB: 1-of-2   3.4 GB: 2-of-2
Model ArchitectureMPTForCausalLM
Licenseapache-2.0
Model Max Length65536
Transformers Version4.28.1
Tokenizer ClassGPTNeoXTokenizer
Vocabulary Size50432
Torch Data Typebfloat16

Quantized Models of the Mpt 7B Storywriter Qlora

Model
Likes
Downloads
VRAM
Mpt 7B Storywriter Q82246 GB

Best Alternatives to Mpt 7B Storywriter Qlora

Best Alternatives
Context / RAM
Downloads
Likes
Mpt 7B0K / 13.3 GB298221164
Mpt 7B Chat0K / 13.3 GB20372512
Mpt 7B Storywriter0K / 13.3 GB1694824
Mpt 7B Instruct0K / 13.3 GB7997468
Mpt 7B Int8 Ov0K / 0 GB120
Shears Mpt 7B 50 Base0K / 13.3 GB181
Sea Lion 7B Instruct0K / 15 GB48723
Sea Lion 7B0K / 15 GB240236
Mpt 7B0K / 26.5 GB35281
Mpt 7B 8K0K / 13.3 GB156626
Note: green Score (e.g. "73.2") means that the model is better than Jordancole21/mpt-7b-storywriter-qlora.

Rank the Mpt 7B Storywriter Qlora Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 40123 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241217