T3Q LLM Sft1.0 Dpo1.0 by T3Q-LLM

 ยป  All LLMs  ยป  T3Q-LLM  ยป  T3Q LLM Sft1.0 Dpo1.0   URL Share it on

  Autotrain compatible   Conversational Dataset:maywell/ko ultrafeedba...   Endpoints compatible   Llama   Region:us   Safetensors   Sharded   Tensorflow

T3Q LLM Sft1.0 Dpo1.0 Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
T3Q LLM Sft1.0 Dpo1.0 (T3Q-LLM/T3Q-LLM-sft1.0-dpo1.0)

T3Q LLM Sft1.0 Dpo1.0 Parameters and Internals

Model Type 
text-generation
Use Cases 
Areas:
text-generation
Primary Use Cases:
Korean text generation
Training Details 
Data Sources:
maywell/ko_Ultrafeedback_binarized
Methodology:
Fine-tuned with DPO.
Model Architecture:
Version of T3Q-LLM/T3Q-LLM-solar10.8-sft-v1.0 fine-tuned with DPO.
Input Output 
Input Format:
A chat format with prompts such as 'A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.'
Accepted Modalities:
text
Output Format:
Text responses
LLM NameT3Q LLM Sft1.0 Dpo1.0
Repository ๐Ÿค—https://huggingface.co/T3Q-LLM/T3Q-LLM-sft1.0-dpo1.0 
Model Size10.8b
Required VRAM21.6 GB
Updated2024-12-26
MaintainerT3Q-LLM
Model Typellama
Model Files  4.9 GB: 1-of-5   4.9 GB: 2-of-5   5.0 GB: 3-of-5   4.9 GB: 4-of-5   1.9 GB: 5-of-5
Model ArchitectureLlamaForCausalLM
Licenseapache-2.0
Context Length4096
Model Max Length4096
Transformers Version4.38.2
Tokenizer ClassLlamaTokenizer
Padding Token</s>
Vocabulary Size40960
Torch Data Typebfloat16

Best Alternatives to T3Q LLM Sft1.0 Dpo1.0

Best Alternatives
Context / RAM
Downloads
Likes
...Korean Instruct 10.8B V1.0 32K32K / 21.6 GB7510
EEVE Korean 10.8B V1.0 16K16K / 21.6 GB290
...EVE Korean Instruct 10.8B V1.04K / 21.6 GB22493137
...Judgment Transducer 10.8B V2.04K / 21.6 GB2950
EEVE Korean 10.8B RAFT4K / 43.2 GB47670
T3Q LLM1 CV V1.04K / 21.6 GB92100
T3Q LLM1 CV V2.04K / 21.6 GB47710
EEVE Ver 4.1 Sft4K / 21.6 GB47670
T3Q LLM Solar10.8 Sft V1.04K / 21.6 GB47690
AlgograpV44K / 43.2 GB47640
Note: green Score (e.g. "73.2") means that the model is better than T3Q-LLM/T3Q-LLM-sft1.0-dpo1.0.

Rank the T3Q LLM Sft1.0 Dpo1.0 Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 40248 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241217