GPT J 6B 8bit Sharded by ethzanalytics

 ยป  All LLMs  ยป  ethzanalytics  ยป  GPT J 6B 8bit Sharded   URL Share it on

  8-bit   8bit   Autotrain compatible   Compression   Gptj   Pytorch   Quantization   Quantized   Region:us   Safetensors   Sharded   Tensorflow

GPT J 6B 8bit Sharded Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
GPT J 6B 8bit Sharded (ethzanalytics/gpt-j-6B-8bit-sharded)

GPT J 6B 8bit Sharded Parameters and Internals

Model Type 
text-generation, quantization, compression
Additional Notes 
This model version is sharded into smaller parts for loading on low-RAM environments, such as free Colab runtimes. It is intended as an 8-bit quantized version for compression and efficient inference.
Input Output 
Performance Tips:
Patch the model before loading using 8bit weights.
LLM NameGPT J 6B 8bit Sharded
Repository ๐Ÿค—https://huggingface.co/ethzanalytics/gpt-j-6B-8bit-sharded 
Model Size6b
Required VRAM6.3 GB
Updated2025-01-21
Maintainerethzanalytics
Model Typegptj
Model Files  1.0 GB: 1-of-7   1.0 GB: 2-of-7   1.0 GB: 3-of-7   0.9 GB: 4-of-7   1.0 GB: 5-of-7   1.0 GB: 6-of-7   0.4 GB: 7-of-7   1.0 GB: 1-of-7   1.0 GB: 2-of-7   1.0 GB: 3-of-7   0.9 GB: 4-of-7   1.0 GB: 5-of-7   1.0 GB: 6-of-7   0.4 GB: 7-of-7
Quantization Type8bit
Model ArchitectureGPTJForCausalLM
Licenseapache-2.0
Model Max Length2048
Transformers Version4.21.3
Tokenizer ClassGPT2Tokenizer
Beginning of Sentence Token<|endoftext|>
End of Sentence Token<|endoftext|>
Unk Token<|endoftext|>
Vocabulary Size50400
Torch Data Typefloat32
Activation Functiongelu_new
Errorsreplace

Best Alternatives to GPT J 6B 8bit Sharded

Best Alternatives
Context / RAM
Downloads
Likes
Model0K / 6.2 GB160
...oduct NER GPT J 6B 4bit Merged0K / 2.5 GB960
...nese Novel GPT J 6B F16 Marisa0K / 12.2 GB164
Kakaobrain Kogpt 6B 8bit0K / 6.7 GB252
Pygmalion 6b Dev 4bit 128g0K / 4 GB63121
GPT J 6B Skein 4bit 128g0K / 4 GB101
GPT J 6B Alpaca Gpt40K / 24.3 GB1420
Pygmalion 6B 4bit 128g0K / 4 GB113
Dolly Shygmalion 6B 4bit 128g0K / 4 GB215
Pygmalion 6B 4bit 128g0K / 4 GB4039
Note: green Score (e.g. "73.2") means that the model is better than ethzanalytics/gpt-j-6B-8bit-sharded.

Rank the GPT J 6B 8bit Sharded Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 41680 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227