Bertin GPT J 6B by bertin-project

 ยป  All LLMs  ยป  bertin-project  ยป  Bertin GPT J 6B   URL Share it on

  Arxiv:2101.00027   Arxiv:2104.09864   Autotrain compatible   Base model:eleutherai/gpt-j-6b Base model:finetune:eleutherai... Dataset:bertin-project/mc4-es-...   Endpoints compatible   Es   Gptj   Pytorch   Region:us   Safetensors

Bertin GPT J 6B Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Bertin GPT J 6B (bertin-project/bertin-gpt-j-6B)

Bertin GPT J 6B Parameters and Internals

Model Type 
causal-lm
Use Cases 
Areas:
Research, Commercial applications
Primary Use Cases:
Text generation from a prompt
Limitations:
Model may produce inaccurate or biased text, May produce socially unacceptable content
Considerations:
Review and curate output before use
Supported Languages 
Spanish (Highly proficient)
Training Details 
Data Sources:
bertin-project/mc4-es-sampled
Data Volume:
65 billion tokens
Methodology:
Finetuning
Context Length:
2048
Training Time:
~6 months
Hardware Used:
Single TPU v3-8 VM
Model Architecture:
28 layers, d_model=4096, d_ff=16384, n_heads=16, Rotary Position Embedding (RoPE) applied to 64 dimensions of each head
Input Output 
Input Format:
Text format for prompt input
Accepted Modalities:
text
Output Format:
Text format
Performance Tips:
Monitor outputs for bias or inaccuracies
Release Notes 
Version:
v1
Date:
August 25th, 2022
Notes:
Full and half-precision weights at step 1M
Version:
v1beta3
Date:
July 22nd, 2022
Notes:
Full and half-precision weights at step 850k
Version:
v1beta2
Date:
June 6th, 2022
Notes:
Full and half-precision weights at step 616k
Version:
v1beta1
Date:
April 28th, 2022
Notes:
Half-precision weights only at step 408k
LLM NameBertin GPT J 6B
Repository ๐Ÿค—https://huggingface.co/bertin-project/bertin-gpt-j-6B 
Base Model(s)  GPT J 6B   EleutherAI/gpt-j-6b
Model Size6b
Required VRAM24.2 GB
Updated2025-03-18
Maintainerbertin-project
Model Typegptj
Model Files  24.2 GB   24.2 GB
Supported Languageses
Model ArchitectureGPTJForCausalLM
Licenseapache-2.0
Model Max Length2048
Transformers Version4.10.0.dev0
Tokenizer ClassGPT2Tokenizer
Vocabulary Size50400
Torch Data Typefloat32
Activation Functiongelu_new

Best Alternatives to Bertin GPT J 6B

Best Alternatives
Context / RAM
Downloads
Likes
Test GPT J 6B0K / 2.5 GB5500
Deception Normal0K / 12.2 GB70
Deception Filteredpositive0K / 12.2 GB60
Gptj Allenai Toxicity Blackbox0K / 12.2 GB70
...j Allenai Toxicity Explainable0K / 12.2 GB50
Pygmalion 6B0K / 16.3 GB4328745
Pygmalion 6B Roleplay0K / 12.1 GB19292
GPT J 6B0K / 24.2 GB2714221486
Explainable GPT J 6B0K / 24.2 GB50
Gpt4all J0K / 12.2 GB2691297
Note: green Score (e.g. "73.2") means that the model is better than bertin-project/bertin-gpt-j-6B.

Rank the Bertin GPT J 6B Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 45231 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227