TinyStoriesChinese 110M by fzmnm

 ยป  All LLMs  ยป  fzmnm  ยป  TinyStoriesChinese 110M   URL Share it on

  Arxiv:2203.15556   Autotrain compatible Dataset:adam89/tinystorieschin...   Endpoints compatible   Llama   Pytorch   Region:us   Zh

TinyStoriesChinese 110M Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
TinyStoriesChinese 110M (fzmnm/TinyStoriesChinese-110M)

TinyStoriesChinese 110M Parameters and Internals

Model Type 
text generation
Use Cases 
Areas:
Educational purposes, Machine learning training for beginners
Applications:
Generation of short Chinese stories
Primary Use Cases:
Text generation for educational tools
Limitations:
Consistency and depth of logical capabilities vary
Considerations:
Focused on simple, narrowly defined tasks; ideal for experimentation
Additional Notes 
Demonstrates potential of smaller models to produce coherent text without extensive computational resources.
Supported Languages 
zh (Proficient)
Training Details 
Data Sources:
adam89/TinyStoriesChinese
Data Volume:
2G tokens
Methodology:
Standard llama2 format Transformer with 12 layers, 12 attention heads, and a hidden size of 768. The model utilizes a 1024 context window with Relative Positional Encodings (RoPE) and a vocabulary size of 5,000.
Context Length:
1024
Training Time:
87 hours
Hardware Used:
NVIDIA RTX 2080 Super with 8 GB RAM
Model Architecture:
Standard llama2 format Transformer
Input Output 
Input Format:
Text prompt
Accepted Modalities:
text
Output Format:
Generated story text
LLM NameTinyStoriesChinese 110M
Repository ๐Ÿค—https://huggingface.co/fzmnm/TinyStoriesChinese-110M 
Model Size110m
Required VRAM0.4 GB
Updated2025-02-15
Maintainerfzmnm
Model Typellama
Model Files  0.4 GB
Supported Languageszh
Model ArchitectureLlamaForCausalLM
Licensecc
Context Length1024
Model Max Length1024
Transformers Version4.41.2
Tokenizer ClassLlamaTokenizer
Vocabulary Size5000

Best Alternatives to TinyStoriesChinese 110M

Best Alternatives
Context / RAM
Downloads
Likes
BigWeave V20 110B4K / 220 GB470
HelpingAI 110M1K / 0.2 GB5082
Tinybooks 110M Idiot Iter10001K / 0.4 GB1350
Llama 110M1K / 0.2 GB90
Tinyllama 110M1K / 0.2 GB9684
Tinyllama 110M Fp321K / 0.4 GB1820
Note: green Score (e.g. "73.2") means that the model is better than fzmnm/TinyStoriesChinese-110M.

Rank the TinyStoriesChinese 110M Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43137 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227