Fox 1 1.6B Instruct V0.1 by tensoropera

 ยป  All LLMs  ยป  tensoropera  ยป  Fox 1 1.6B Instruct V0.1   URL Share it on

  Arxiv:2411.05281   Autotrain compatible Base model:finetune:tensoroper... Base model:tensoropera/fox-1-1...   Conversational   En   Endpoints compatible   Instruct   Llama   Region:us   Safetensors

Fox 1 1.6B Instruct V0.1 Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Fox 1 1.6B Instruct V0.1 (tensoropera/Fox-1-1.6B-Instruct-v0.1)

Fox 1 1.6B Instruct V0.1 Parameters and Internals

Model Type 
instruction tuned, decoder-only, transformer-based, small language model
Additional Notes 
This model is an instruction tuned version and requires alignment for production use. A chat version will be released soon.
Supported Languages 
en (high)
Training Details 
Data Sources:
text, code data
Data Volume:
3 trillion tokens
Methodology:
3-stage data curriculum
Context Length:
8000
Hardware Used:
8*H100 GPUs
Model Architecture:
Grouped Query Attention (GQA) with 4 key-value heads and 16 attention heads
LLM NameFox 1 1.6B Instruct V0.1
Repository ๐Ÿค—https://huggingface.co/tensoropera/Fox-1-1.6B-Instruct-v0.1 
Base Model(s)  tensoropera/Fox-1-1.6B   tensoropera/Fox-1-1.6B
Model Size1.6b
Required VRAM3.3 GB
Updated2025-02-05
Maintainertensoropera
Model Typellama
Instruction-BasedYes
Model Files  3.3 GB
Supported Languagesen
Model ArchitectureLlamaForCausalLM
Licenseapache-2.0
Context Length8192
Model Max Length8192
Transformers Version4.39.3
Tokenizer ClassGemmaTokenizer
Padding Token<pad>
Vocabulary Size256000
Torch Data Typebfloat16

Best Alternatives to Fox 1 1.6B Instruct V0.1

Best Alternatives
Context / RAM
Downloads
Likes
1.5 Pints 2K V0.116K / 3.1 GB34516
1.5 Pints 16K V0.116K / 3.1 GB3214
Note: green Score (e.g. "73.2") means that the model is better than tensoropera/Fox-1-1.6B-Instruct-v0.1.

Rank the Fox 1 1.6B Instruct V0.1 Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 42565 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227