Replit V1 CodeInstruct 3B by teknium

 ยป  All LLMs  ยป  teknium  ยป  Replit V1 CodeInstruct 3B   URL Share it on

  Autotrain compatible   Code   Custom code Dataset:bigcode/the-stack-dedu... Dataset:sahil2801/codealpaca-2... Dataset:teknium/gpteacher-code...   Endpoints compatible   Instruct   Mpt   Pytorch   Region:us   Self-instruct   Sharded

Replit V1 CodeInstruct 3B Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Replit V1 CodeInstruct 3B (teknium/Replit-v1-CodeInstruct-3B)

Replit V1 CodeInstruct 3B Parameters and Internals

Model Type 
code, instruct, self instruct
Additional Notes 
The model might have issues with 'device=auto' in model arguments.
Supported Languages 
code (Advanced), Markdown (Advanced), Java (Intermediate), JavaScript (Intermediate), Python (Intermediate), TypeScript (Intermediate), PHP (Intermediate), SQL (Intermediate), JSX (Basic), reStructuredText (Basic), Rust (Basic), C (Basic), CSS (Basic), Go (Basic), C++ (Basic), HTML (Basic), Vue (Basic), Ruby (Basic), Jupyter Notebook (Basic), R (Basic), Shell (Basic)
Training Details 
Data Sources:
bigcode/the-stack-dedup, sahil2801/CodeAlpaca-20k, teknium/GPTeacher-CodeInstruct
Data Volume:
~25,000 code instruction/response pairs
Methodology:
Model is fine-tuned on Sahil2801's CodeAlpaca & Teknium's GPTeacher Code-Instruct datasets.
Training Time:
1 hour
Hardware Used:
2x A100 80GB
Input Output 
Input Format:
Alpaca format
Performance Tips:
The model needs trust_remote_code=True in the tokenizer and model from_pretrained functions.
LLM NameReplit V1 CodeInstruct 3B
Repository ๐Ÿค—https://huggingface.co/teknium/Replit-v1-CodeInstruct-3B 
Model Size3b
Required VRAM10.4 GB
Updated2025-02-22
Maintainerteknium
Model Typempt
Model Files  10.0 GB: 1-of-2   0.4 GB: 2-of-2   0.0 GB
Supported Languagescode
Model ArchitectureMPTForCausalLM
Licensecc-by-sa-4.0
Model Max Length512
Transformers Version4.29.2
Tokenizer ClassReplitLMTokenizer
Padding Token<|pad|>
Vocabulary Size32769
Torch Data Typefloat32

Quantized Models of the Replit V1 CodeInstruct 3B

Model
Likes
Downloads
VRAM
Replit V1 CodeInstruct 3B Fp16141695 GB

Best Alternatives to Replit V1 CodeInstruct 3B

Best Alternatives
Context / RAM
Downloads
Likes
Replit Code V1.5 3B0K / 6.6 GB90716289
Code Millenials 3B0K / 5.2 GB1321
Sea Lion 3B0K / 6.4 GB51317
Mpt 3B 8K Instruct0K / 6.9 GB43
Replit Code V1 3B0K / 10.4 GB994725
Glaive Function Calling V10K / 10.4 GB11268
...aive Function Calling V2 Small0K / 10.4 GB8914
Evol Replit V10K / 10.4 GB138
Replit V2 CodeInstruct 3B0K / 10.4 GB4672
Replit CodeInstruct V30K / 10.4 GB92
Note: green Score (e.g. "73.2") means that the model is better than teknium/Replit-v1-CodeInstruct-3B.

Rank the Replit V1 CodeInstruct 3B Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43470 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227