Sft Pythia 1B Deduped Tldr Preference Sft Trl by saminyeasar

 ยป  All LLMs  ยป  saminyeasar  ยป  Sft Pythia 1B Deduped Tldr Preference Sft Trl   URL Share it on

  Autotrain compatible Base model:eleutherai/pythia-1... Base model:finetune:eleutherai...   Conversational   Endpoints compatible   Generated from trainer   Gpt neox   Region:us   Safetensors   Sft   Trl

Sft Pythia 1B Deduped Tldr Preference Sft Trl Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Sft Pythia 1B Deduped Tldr Preference Sft Trl (saminyeasar/sft-pythia-1b-deduped-tldr-preference-sft-trl)

Sft Pythia 1B Deduped Tldr Preference Sft Trl Parameters and Internals

LLM NameSft Pythia 1B Deduped Tldr Preference Sft Trl
Repository ๐Ÿค—https://huggingface.co/saminyeasar/sft-pythia-1b-deduped-tldr-preference-sft-trl 
Base Model(s)  Pythia 1B Deduped   EleutherAI/pythia-1b-deduped
Model Size1b
Required VRAM2 GB
Updated2025-01-24
Maintainersaminyeasar
Model Typegpt_neox
Model Files  2.0 GB   0.0 GB
Model ArchitectureGPTNeoXForCausalLM
Licenseapache-2.0
Context Length2048
Model Max Length2048
Transformers Version4.44.2
Tokenizer ClassGPTNeoXTokenizer
Padding Token[PAD]
Vocabulary Size50304
Torch Data Typebfloat16

Best Alternatives to Sft Pythia 1B Deduped Tldr Preference Sft Trl

Best Alternatives
Context / RAM
Downloads
Likes
Pythia 2.8B Deduped Rp 710M 4K4K / 11.7 GB131
Pythia 1.4B Deduped Rp 420M 4K4K / 6.1 GB141
Pythia 1.4B Deduped Rp 280M 4K4K / 6.1 GB121
...rAI Pythia 1B Deduped Sft Tldr2K / 4 GB75960
Pythia 1B Deduped Tldr Sft2K / 2 GB8470
Pythia 1B Self Kto Iter02K / 2 GB560
Pythia 1B Kto Iter02K / 2 GB400
Rloo Trial22K / 2 GB80
TinyLM 1B Test22K / 2.7 GB270
Rloo Tldr2K / 2 GB110
Note: green Score (e.g. "73.2") means that the model is better than saminyeasar/sft-pythia-1b-deduped-tldr-preference-sft-trl.

Rank the Sft Pythia 1B Deduped Tldr Preference Sft Trl Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 45429 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227