Japanese Large Lm 3.6B Instruction Sft 4bit 128g Actorder False by line-corporation

 ยป  All LLMs  ยป  line-corporation  ยป  Japanese Large Lm 3.6B Instruction Sft 4bit 128g Actorder False   URL Share it on

  4bit   Autotrain compatible   Gpt neox   Instruct   Ja   Pytorch   Quantized   Region:us   Safetensors

Japanese Large Lm 3.6B Instruction Sft 4bit 128g Actorder False Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Japanese Large Lm 3.6B Instruction Sft 4bit 128g Actorder False (line-corporation/japanese-large-lm-3.6b-instruction-sft-4bit-128g-actorder_False)

Japanese Large Lm 3.6B Instruction Sft 4bit 128g Actorder False Parameters and Internals

Model Type 
text generation
Supported Languages 
Japanese (fully supported)
LLM NameJapanese Large Lm 3.6B Instruction Sft 4bit 128g Actorder False
Repository ๐Ÿค—https://huggingface.co/line-corporation/japanese-large-lm-3.6b-instruction-sft-4bit-128g-actorder_False 
Model Size3.6b
Required VRAM2.1 GB
Updated2025-02-22
Maintainerline-corporation
Model Typegpt_neox
Instruction-BasedYes
Model Files  2.4 GB   2.1 GB
Supported Languagesja
Quantization Type4bit
Model ArchitectureGPTNeoXForCausalLM
Licenseapache-2.0
Context Length2048
Model Max Length2048
Transformers Version4.33.0
Vocabulary Size51200
Torch Data Typefloat16

Best Alternatives to Japanese Large Lm 3.6B Instruction Sft 4bit 128g Actorder False

Best Alternatives
Context / RAM
Downloads
Likes
...tion Sft 8bit 1g Actorder True2K / 2.8 GB843
...rrowSmartPlus 3.6B Instruction2K / 14.3 GB51
...rtPlus 3.6B Instant Sft JHSVer2K / 14.3 GB91
...T Neox 3.6B Instruction Sft V22K / 7.4 GB5463826
... Large Lm 3.6B Instruction Sft2K / 7.2 GB89025
... GPT Neox 3.6B Instruction Ppo2K / 7.4 GB258770
... GPT Neox 3.6B Instruction Sft2K / 7.4 GB900101
Note: green Score (e.g. "73.2") means that the model is better than line-corporation/japanese-large-lm-3.6b-instruction-sft-4bit-128g-actorder_False.

Rank the Japanese Large Lm 3.6B Instruction Sft 4bit 128g Actorder False Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43470 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227