Smol Llama 101M GQA Python by BEE-spoke-data

 ยป  All LLMs  ยป  BEE-spoke-data  ยป  Smol Llama 101M GQA Python   URL Share it on

  Autotrain compatible   Codegen Dataset:bee-spoke-data/pypi cl...   En   Endpoints compatible   Llama   Markdown   Python   Region:us   Safetensors   Smol llama

Smol Llama 101M GQA Python Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Smol Llama 101M GQA Python (BEE-spoke-data/smol_llama-101M-GQA-python)

Smol Llama 101M GQA Python Parameters and Internals

Model Type 
code generation, text generation
Use Cases 
Areas:
research, education
Applications:
Python code generation, Markdown generation
Primary Use Cases:
generating basic Python functions, creating simple scripting utilities
Limitations:
struggles with complex reasoning and planning tasks
Considerations:
Use with understanding that it may contain bugs and is best suited for educational or experimental purposes.
Additional Notes 
The model is part of an experiment to explore code generation capabilities of smaller models. May create useful utilities but is not optimized for more complex tasks.
Supported Languages 
en (basic level)
Training Details 
Data Sources:
BEE-spoke-data/pypi_clean-deduped
Methodology:
general pre-trained checkpoint with additional Python-related tokens added to vocab
Training Time:
+1 epoch
Model Architecture:
similar to base model architecture with new tokens
Input Output 
Input Format:
text prompts, specifically Python code snippets or descriptions
Accepted Modalities:
text
Output Format:
generated code snippets in Python or README markdown
Performance Tips:
May require specific configurations in tokenizer settings (e.g., use_fast=False).
LLM NameSmol Llama 101M GQA Python
Repository ๐Ÿค—https://huggingface.co/BEE-spoke-data/smol_llama-101M-GQA-python 
Model Size101m
Required VRAM0.4 GB
Updated2025-02-05
MaintainerBEE-spoke-data
Model Typellama
Model Files  0.4 GB
Supported Languagesen
Model ArchitectureLlamaForCausalLM
Licenseapache-2.0
Context Length1024
Model Max Length1024
Transformers Version4.33.3
Tokenizer ClassLlamaTokenizer
Beginning of Sentence Token<s>
End of Sentence Token</s>
Unk Token<unk>
Vocabulary Size32128
Torch Data Typefloat32

Best Alternatives to Smol Llama 101M GQA Python

Best Alternatives
Context / RAM
Downloads
Likes
Smol Llama 101M GQA1K / 0.4 GB382228
Smol Llama 101M Chat V11K / 0.4 GB2319
...ephyr Smol Llama 100M Sft Full1K / 0.2 GB12961
...Llama 101M Midjourney Messages1K / 0.4 GB131
Note: green Score (e.g. "73.2") means that the model is better than BEE-spoke-data/smol_llama-101M-GQA-python.

Rank the Smol Llama 101M GQA Python Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 42625 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227