Ov Opt 350M 8bit Kv Cache by vuiseng9

 ยป  All LLMs  ยป  vuiseng9  ยป  Ov Opt 350M 8bit Kv Cache   URL Share it on

  8bit   Autotrain compatible   Dataset:wikitext   Endpoints compatible   Generated from trainer   Model-index   Openvino   Opt   Pytorch   Quantized   Region:us

Ov Opt 350M 8bit Kv Cache Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Ov Opt 350M 8bit Kv Cache (vuiseng9/ov-opt-350m-8bit-kv-cache)

Ov Opt 350M 8bit Kv Cache Parameters and Internals

Model Type 
text-generation
Use Cases 
Considerations:
More information needed
Additional Notes 
## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 1 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - training_steps: 1 ### Training results ### Framework versions - Transformers 4.30.2 - Pytorch 2.0.1+cu117 - Datasets 2.13.1 - Tokenizers 0.13.3
Training Details 
Data Sources:
wikitext wikitext-2-raw-v1
Methodology:
fine-tuning
LLM NameOv Opt 350M 8bit Kv Cache
Repository ๐Ÿค—https://huggingface.co/vuiseng9/ov-opt-350m-8bit-kv-cache 
Model Size350m
Required VRAM0.4 GB
Updated2025-03-15
Maintainervuiseng9
Model Typeopt
Model Files  0.4 GB   1.3 GB   0.0 GB
Quantization Type8bit
Model ArchitectureOPTForCausalLM
Licenseother
Context Length2048
Model Max Length2048
Transformers Version4.30.2
Tokenizer ClassGPT2Tokenizer
Beginning of Sentence Token</s>
End of Sentence Token</s>
Unk Token</s>
Vocabulary Size50272
Torch Data Typefloat32
Activation Functionrelu
Errorsreplace

Best Alternatives to Ov Opt 350M 8bit Kv Cache

Best Alternatives
Context / RAM
Downloads
Likes
Fbopt 350M 8bit2K / 0.4 GB41530
Opt Mini Dataset 02K / 0.7 GB590
Facebook Opt 350M SFT Korz142K / 0.7 GB440
Temp Model Sft2K / 1.3 GB260
Gpt350 Chat S V0 12K / 0.7 GB710
Gpt350 Chat S V02K / 0.7 GB680
Dadjokes Tuned Opt2K / 1.3 GB802
Opt 350M2K / 0.7 GB497507139
Sunday2K / 1.3 GB1560
Remedycure2K / 1.3 GB1550
Note: green Score (e.g. "73.2") means that the model is better than vuiseng9/ov-opt-350m-8bit-kv-cache.

Rank the Ov Opt 350M 8bit Kv Cache Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 45097 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227