REPROD DPO Helpfulhelpful Gpt3 Subset 1 Modelgemma2b Maxsteps10000 Bz8 Lr1e 05 by Holarissun

 ยป  All LLMs  ยป  Holarissun  ยป  REPROD DPO Helpfulhelpful Gpt3 Subset 1 Modelgemma2b Maxsteps10000 Bz8 Lr1e 05   URL Share it on

  Adapter Base model:adapter:google/gemm...   Base model:google/gemma-2b   Dpo   Finetuned   Generated from trainer   Lora   Peft   Region:us   Safetensors   Trl

REPROD DPO Helpfulhelpful Gpt3 Subset 1 Modelgemma2b Maxsteps10000 Bz8 Lr1e 05 Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
REPROD DPO Helpfulhelpful Gpt3 Subset 1 Modelgemma2b Maxsteps10000 Bz8 Lr1e 05 (Holarissun/REPROD_dpo_helpfulhelpful_gpt3_subset-1_modelgemma2b_maxsteps10000_bz8_lr1e-05)

REPROD DPO Helpfulhelpful Gpt3 Subset 1 Modelgemma2b Maxsteps10000 Bz8 Lr1e 05 Parameters and Internals

LLM NameREPROD DPO Helpfulhelpful Gpt3 Subset 1 Modelgemma2b Maxsteps10000 Bz8 Lr1e 05
Repository ๐Ÿค—https://huggingface.co/Holarissun/REPROD_dpo_helpfulhelpful_gpt3_subset-1_modelgemma2b_maxsteps10000_bz8_lr1e-05 
Base Model(s)  Gemma 2B   google/gemma-2b
Model Size2b
Required VRAM0 GB
Updated2025-02-22
MaintainerHolarissun
Model Files  0.0 GB   0.0 GB
Model ArchitectureAdapter
Licensegemma
Is Biasednone
Tokenizer ClassGemmaTokenizer
Padding Token<pad>
PEFT TypeLORA
LoRA ModelYes
PEFT Target Modulesv_proj|q_proj
LoRA Alpha32
LoRA Dropout0.05
R Param32

Best Alternatives to REPROD DPO Helpfulhelpful Gpt3 Subset 1 Modelgemma2b Maxsteps10000 Bz8 Lr1e 05

Best Alternatives
Context / RAM
Downloads
Likes
Finetuned Gemma38K / 5.1 GB90
Phi Gemma Nlaf V10K / 0.1 GB50
Phi Gemma Nlaf V00K / 0.1 GB50
Gemma 2B It Nlai P10K / 0 GB60
Ger Lora 3K Checkpoint0K / 0 GB80
German 2B Lora 6K0K / 0 GB60
1 8K Adater Ger0K / 0 GB50
2B Lora Adapter Llama Alpaca0K / 0.1 GB70
Google Gemma 2B 17198825710K / 0 GB70
Google Gemma 2B 17198986620K / 0 GB60
Note: green Score (e.g. "73.2") means that the model is better than Holarissun/REPROD_dpo_helpfulhelpful_gpt3_subset-1_modelgemma2b_maxsteps10000_bz8_lr1e-05.

Rank the REPROD DPO Helpfulhelpful Gpt3 Subset 1 Modelgemma2b Maxsteps10000 Bz8 Lr1e 05 Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43470 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227