Instruct GPT J by crumb

 ยป  All LLMs  ยป  crumb  ยป  Instruct GPT J   URL Share it on

  Arxiv:2106.09685   Alpaca   Dataset:tatsu-lab/alpaca   Dataset:the pile   En   Endpoints compatible   Gptj   Instruct   Lora   Peft   Region:us
Model Card on HF ๐Ÿค—: https://huggingface.co/crumb/Instruct-GPT-J 

Instruct GPT J Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Instruct GPT J (crumb/Instruct-GPT-J)

Instruct GPT J Parameters and Internals

Model Type 
text generation, instruction-following
Use Cases 
Areas:
Research, Instruction-following applications
Applications:
Instruction-based text generation tasks
Primary Use Cases:
Generating responses to instructional prompts
Additional Notes 
This model is a personal project and not officially affiliated with EleutherAI.
Supported Languages 
en (high)
Training Details 
Data Sources:
tatsu-lab/alpaca
Methodology:
Finetuned using LoRA (Low-Rank Adaptation)
Model Architecture:
Based on GPT-J architecture
Input Output 
Input Format:
Instruction and optionally paired input context.
Accepted Modalities:
text
Output Format:
Generated text response to the instruction.
LLM NameInstruct GPT J
Repository ๐Ÿค—https://huggingface.co/crumb/Instruct-GPT-J 
Required VRAM0 GB
Updated2025-02-22
Maintainercrumb
Instruction-BasedYes
Model Files  0.0 GB
Supported Languagesen
Model ArchitectureAutoModel
Is Biasednone
PEFT TypeLORA
LoRA ModelYes
PEFT Target Modulesq_proj|v_proj|k_proj|out_proj
LoRA Alpha32
LoRA Dropout0.05
R Param8

Best Alternatives to Instruct GPT J

Best Alternatives
Context / RAM
Downloads
Likes
Autotrain Umberto Proclama0K / 0.9 GB50
Mamba Python0K / 2 GB130
...hi 3 Mini 4K Instruct Ct2 Int80K / 3.8 GB21
...l 8x7B Instruct V0.1 Llamafile0K /  GB203518
...hin 2.5 Mixtral 8x7b Llamafile0K /  GB3805
CSUMLM0K /  GB370
Vigogne Bloom 7b1 Instruct0K / 0.1 GB04
...a Instruction Fine Tune French0K / 0 GB04
Lora Model0K / 0.1 GB11890
Phi 3.1 EvolKit Lora0K / 0.1 GB01
Note: green Score (e.g. "73.2") means that the model is better than crumb/Instruct-GPT-J.

Rank the Instruct GPT J Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43470 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227