TeenyTinyLlama 460M Chat AWQ by nicholasKluge

 ยป  All LLMs  ยป  nicholasKluge  ยป  TeenyTinyLlama 460M Chat AWQ   URL Share it on

  Arxiv:2401.16640   4-bit   Alignment   Assistant   Autotrain compatible   Awq Base model:nicholaskluge/teeny... Base model:quantized:nicholask...   Co2 eq emissions   Conversation   Conversational Dataset:nicholaskluge/instruct...   Instruct   Llama   Pt   Quantized   Region:us   Safetensors

TeenyTinyLlama 460M Chat AWQ Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
TeenyTinyLlama 460M Chat AWQ (nicholasKluge/TeenyTinyLlama-460m-Chat-awq)

TeenyTinyLlama 460M Chat AWQ Parameters and Internals

Model Type 
text generation, conversation, assistant
Use Cases 
Areas:
Research on low-resource language modeling
Applications:
Scientific experiments, Fine-tuning for deployment
Primary Use Cases:
Understand challenges related to developing language models in low-resource languages
Limitations:
Not intended for deployment, Not suitable for human-facing interactions, Not suitable for translation or generating text in other languages
Considerations:
Users are warned of risks, biases, and the need for human moderation.
Additional Notes 
Use CAUTION in deploying as it inherits general language model biases and behaviors.
Supported Languages 
Brazilian Portuguese (high)
Training Details 
Data Sources:
nicholasKluge/instruct-aira-dataset-v2
Data Volume:
unknown
Methodology:
fine-tuning
Hardware Used:
NVIDIA A100-SXM4-40GB
Model Architecture:
quantized version using AutoAWQ
Responsible Ai Considerations 
Fairness:
The model can produce toxic content, i.e., harmful, offensive, or detrimental to individuals, groups, or communities.
Transparency:
Users are encouraged to perform their risk analysis on these models.
Accountability:
Developers of TeenyTinyLlama
Mitigation Strategies:
Hasn't been outlined specifically.
Input Output 
Input Format:
Uses special tokens to demarcate user queries and model responses.
Accepted Modalities:
text
Output Format:
Generates text in response to queries.
Performance Tips:
Ensure human moderation in interactive applications.
LLM NameTeenyTinyLlama 460M Chat AWQ
Repository ๐Ÿค—https://huggingface.co/nicholasKluge/TeenyTinyLlama-460m-Chat-awq 
Base Model(s)  TeenyTinyLlama 460M   nicholasKluge/TeenyTinyLlama-460m
Model Size460m
Required VRAM0.3 GB
Updated2025-02-05
MaintainernicholasKluge
Model Typellama
Instruction-BasedYes
Model Files  0.3 GB
Supported Languagespt
AWQ QuantizationYes
Quantization Typeawq
Model ArchitectureLlamaForCausalLM
Licenseapache-2.0
Context Length2048
Model Max Length2048
Transformers Version4.35.2
Tokenizer ClassLlamaTokenizer
Padding Token<pad>
Vocabulary Size32002
Torch Data Typefloat16

Best Alternatives to TeenyTinyLlama 460M Chat AWQ

Best Alternatives
Context / RAM
Downloads
Likes
TeenyTinyLlama 460M AWQ2K / 0.3 GB921
TeenyTinyLlama 460M2K / 1.9 GB131911
TeenyTinyLlama 460M Chat2K / 0 GB4123
...60M Experimental Ptbr Instruct2K / 0.9 GB213
Note: green Score (e.g. "73.2") means that the model is better than nicholasKluge/TeenyTinyLlama-460m-Chat-awq.

Rank the TeenyTinyLlama 460M Chat AWQ Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 42625 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227