SuperThoughts CoT 14B 16K O1 QwQ by Pinkstack

 ยป  All LLMs  ยป  Pinkstack  ยป  SuperThoughts CoT 14B 16K O1 QwQ   URL Share it on

  Arxiv:2412.08905   4bit   Autotrain compatible Base model:finetune:microsoft/...   Base model:microsoft/phi-4   Chat   Code   Conversational   Cot Dataset:amphora/qwq-longcot-13...   En   Endpoints compatible   Llama   Math   Model-index   Multilingual   Phi   Phi3   Pytorch   Quantized   Reasoning   Region:us   Safetensors   Sharded   Tensorflow

SuperThoughts CoT 14B 16K O1 QwQ Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
SuperThoughts CoT 14B 16K O1 QwQ (Pinkstack/SuperThoughts-CoT-14B-16k-o1-QwQ)

SuperThoughts CoT 14B 16K O1 QwQ Parameters and Internals

LLM NameSuperThoughts CoT 14B 16K O1 QwQ
Repository ๐Ÿค—https://huggingface.co/Pinkstack/SuperThoughts-CoT-14B-16k-o1-QwQ 
Base Model(s)  Phi 4   microsoft/phi-4
Model Size14b
Required VRAM29.4 GB
Updated2025-02-05
MaintainerPinkstack
Model Typellama
Model Files  4.9 GB: 1-of-6   5.0 GB: 2-of-6   4.9 GB: 3-of-6   5.0 GB: 4-of-6   5.0 GB: 5-of-6   4.6 GB: 6-of-6
Supported Languagesen
Quantization Type4bit
Model ArchitectureLlamaForCausalLM
Licensemit
Context Length16384
Model Max Length16384
Transformers Version4.47.1
Tokenizer ClassGPT2Tokenizer
Padding Token<|dummy_87|>
Vocabulary Size100352
Torch Data Typefloat16

Best Alternatives to SuperThoughts CoT 14B 16K O1 QwQ

Best Alternatives
Context / RAM
Downloads
Likes
...hts CoT 14B 16K O1 QwQ PyTorch16K / 29.4 GB01
CausalLM 14B EXL28K / 8.5 GB53
...Qwen2.5llamaify 14B V23.1 200K195K / 29.7 GB51790
...Qwen2.5llamaify 14B V23.3 200K195K / 29.7 GB105
GeM2 Llamion 14B LongChat195K / 29 GB33914
Openbuddy Zero 14B V22.3 32K32K / 28 GB11601
Phi 4 ReasoningRP16K / 29.5 GB1002
GZA 14B Sce16K / 29.5 GB480
Phi 4 Super16K / 29.5 GB1218
Blaze 14B XElite16K / 29.4 GB577
Note: green Score (e.g. "73.2") means that the model is better than Pinkstack/SuperThoughts-CoT-14B-16k-o1-QwQ.

Rank the SuperThoughts CoT 14B 16K O1 QwQ Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 42577 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227