Alpaca Cerebras 6.7B by bjoernp

 ยป  All LLMs  ยป  bjoernp  ยป  Alpaca Cerebras 6.7B   URL Share it on

  Arxiv:1910.09700   Dataset:tatsu-lab/alpaca   Dataset:yahma/alpaca-cleaned   En   Endpoints compatible   Lora   Region:us

Alpaca Cerebras 6.7B Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Alpaca Cerebras 6.7B (bjoernp/alpaca-cerebras-6.7B)

Alpaca Cerebras 6.7B Parameters and Internals

Model Type 
Transformer-based Language Model
Additional Notes 
Experiments used a private infrastructure with estimated total emissions of 0.97 kgCO2eq.
Supported Languages 
English (proficient)
Training Details 
Data Sources:
yhama/alpaca-cleaned, tatsu-lab/alpaca
Data Volume:
20 tokens per model parameter (Chinchilla scaling laws)
Methodology:
Pre-training and fine-tuning on Cerebras systems with LoRA adapter
Context Length:
2048
Hardware Used:
Andromeda AI supercomputer with 16 CS-2 wafer scale systems
Model Architecture:
GPT-3 style architecture with LoRA adapter
Input Output 
Input Format:
Byte Pair Encoding
Accepted Modalities:
text
Output Format:
Generated text
LLM NameAlpaca Cerebras 6.7B
Repository ๐Ÿค—https://huggingface.co/bjoernp/alpaca-cerebras-6.7B 
Model Size6.7b
Required VRAM0 GB
Updated2024-12-22
Maintainerbjoernp
Model Files  0.0 GB
Supported Languagesen
Model ArchitectureAutoModel
Licenseapache-2.0
Is Biasednone
PEFT TypeLORA
LoRA ModelYes
PEFT Target Modulesc_attn
LoRA Alpha16
LoRA Dropout0.05
R Param8

Best Alternatives to Alpaca Cerebras 6.7B

Best Alternatives
Context / RAM
Downloads
Likes
Cerebras GPT 6.7B0K / 26.8 GB114865
Vigogne Opt 6.7B Instruct0K / 0 GB02
...pseek Coder 6.7B Instruct GGUF0K / 2.8 GB5895173
Magicoder S DS 6.7B GGUF0K / 2.8 GB89876
...ydecompiler 3.7 6.7B V0.9 GGUF0K / 2.5 GB560
...enCodeInterpreter DS 6.7B GGUF0K / 2.5 GB622
Deepseek Coder 6.7B Base GGUF0K / 2.8 GB96711
Note: green Score (e.g. "73.2") means that the model is better than bjoernp/alpaca-cerebras-6.7B.

Rank the Alpaca Cerebras 6.7B Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 40066 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241217