Huskylm 2.5 8B by darkcloudai

 ยป  All LLMs  ยป  darkcloudai  ยป  Huskylm 2.5 8B   URL Share it on

  Autotrain compatible Base model:finetune:meta-llama... Base model:meta-llama/meta-lla...   Conversational   Darkcloudai   Dataset:ai2 arc Dataset:allenai/ultrafeedback ... Dataset:argilla/distilabel-int... Dataset:b-mc2/sql-create-conte... Dataset:bluemoon-fandom-1-1-rp...   Dataset:boolq   Dataset:cakiki/rosetta-code   Dataset:camel-ai/biology   Dataset:camel-ai/chemistry   Dataset:camel-ai/math   Dataset:camel-ai/physics   Dataset:codeparrot/apps   Dataset:darkcloudai-generaldpo Dataset:darkcloudai-smallmodel... Dataset:darkcloudai-unalignmen... Dataset:darkcloudai-webdriver-...   Dataset:facebook/belebele Dataset:glaiveai/glaive-functi... Dataset:grimulkan/limarp-augme... Dataset:jondurbin/airoboros-3.... Dataset:jondurbin/cinematika-v... Dataset:jondurbin/contextual-d... Dataset:jondurbin/gutenberg-dp...   Dataset:jondurbin/py-dpo-v0.1 Dataset:jondurbin/truthy-dpo-v...   Dataset:kingbri/pippa-sharegpt   Dataset:ldjnr/capybara   Dataset:lmsys/lmsys-chat-1m Dataset:mattpscott/airoboros-s... Dataset:migtissera/synthia-v1.... Dataset:muennighoff/natural-in...   Dataset:open-orca/slimorca   Dataset:openbookqa Dataset:parisneo/lollms aware ...   Dataset:piqa   Dataset:ropes   Dataset:squad v2   Dataset:tiger-lab/mathinstruct Dataset:unalignment/toxic-dpo-... Dataset:vezora/tested-22k-pyth... Dataset:whiterabbitneo/wrn-cha... Dataset:whiterabbitneo/wrn-cha...   Dataset:winogrande Dataset:wizardlm/wizardlm evol...   En   Endpoints compatible   Huskylm   Instruct   Llama   Llama-3   Region:us   Safetensors   Sharded   Tensorflow

Huskylm 2.5 8B Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Huskylm 2.5 8B (darkcloudai/huskylm-2.5-8b)

Huskylm 2.5 8B Parameters and Internals

Model Type 
text generation, chat
Use Cases 
Areas:
General-purpose chat, Assistance with writing and coding, Live chat agent, In-game NPC interactions
Applications:
Conversational agents, Programming assistance, Customer support, Gaming
Primary Use Cases:
Interactive dialogue, Text generation
Limitations:
Potential for biases, Limited context window of 8192 tokens, Possibility of errors
Considerations:
Users should verify critical information and be aware of potential biases.
Additional Notes 
Ongoing development with plans to increase context window to 128k tokens.
Supported Languages 
English (High), Chinese (High), Spanish (High), French (High), Polish (High), Others (High)
Training Details 
Data Sources:
darkcloudai-smallmodel-frontieredition, darkcloudai-webdriver-redditcrawl-2023, darkcloudai-unalignment-truthfulness, darkcloudai-generaldpo, ai2_arc, allenai/ultrafeedback_binarized_cleaned, argilla/distilabel-intel-orca-dpo-pairs, jondurbin/airoboros-3.2, codeparrot/apps, facebook/belebele, bluemoon-fandom-1-1-rp-cleaned, boolq, camel-ai/biology, camel-ai/chemistry, camel-ai/math, camel-ai/physics, jondurbin/contextual-dpo-v0.1, jondurbin/gutenberg-dpo-v0.1, jondurbin/py-dpo-v0.1, jondurbin/truthy-dpo-v0.1, LDJnr/Capybara, jondurbin/cinematika-v0.1, WizardLM/WizardLM_evol_instruct_70k, glaiveai/glaive-function-calling-v2, grimulkan/LimaRP-augmented, lmsys/lmsys-chat-1m, ParisNeo/lollms_aware_dataset, TIGER-Lab/MathInstruct, Muennighoff/natural-instructions, openbookqa, kingbri/PIPPA-shareGPT, piqa, Vezora/Tested-22k-Python-Alpaca, ropes, cakiki/rosetta-code, Open-Orca/SlimOrca, b-mc2/sql-create-context, squad_v2, mattpscott/airoboros-summarization, migtissera/Synthia-v1.3, unalignment/toxic-dpo-v0.2, WhiteRabbitNeo/WRN-Chapter-1, WhiteRabbitNeo/WRN-Chapter-2, winogrande
Methodology:
Direct Preference Optimization (DPO) and Reinforcement Learning from Artificial Intelligence Feedback (RLAIF) with traditional SFT (Supervised Fine-Tuning)
Context Length:
8192
Model Architecture:
Llama 3
Input Output 
Input Format:
Text prompts with a context window of up to 8192 tokens.
Accepted Modalities:
text
Output Format:
Text
Performance Tips:
For best results, ensure prompts are clear and within context length limits.
LLM NameHuskylm 2.5 8B
Repository ๐Ÿค—https://huggingface.co/darkcloudai/huskylm-2.5-8b 
Base Model(s)  Meta Llama 3 8B Instruct   meta-llama/Meta-Llama-3-8B-Instruct
Model Size8b
Required VRAM16.1 GB
Updated2025-02-22
Maintainerdarkcloudai
Model Typellama
Instruction-BasedYes
Model Files  5.0 GB: 1-of-4   5.0 GB: 2-of-4   4.9 GB: 3-of-4   1.2 GB: 4-of-4
Supported Languagesen
Model ArchitectureLlamaForCausalLM
Licensellama3
Context Length8192
Model Max Length8192
Transformers Version4.41.0
Tokenizer ClassPreTrainedTokenizerFast
Vocabulary Size128256
Torch Data Typebfloat16

Quantized Models of the Huskylm 2.5 8B

Model
Likes
Downloads
VRAM
Huskylm 2.5 8B AWQ0805 GB

Best Alternatives to Huskylm 2.5 8B

Best Alternatives
Context / RAM
Downloads
Likes
...a 3 8B Instruct Gradient 1048K1024K / 16.1 GB3927680
Mpasila Viking 8B1024K / 16.1 GB840
Hel V2 8B DARK FICTION1024K / 16.1 GB220
161024K / 16.1 GB1690
...di95 LewdStorytellerMix 8B 64K1024K / 16.1 GB692
Because Im Bored Nsfw11024K / 16.1 GB361
121024K / 16.1 GB600
MrRoboto ProLong 8B V4b1024K / 16.1 GB1070
MrRoboto ProLong 8B V1a1024K / 16.1 GB1080
MrRoboto ProLong 8B V2a1024K / 16.1 GB1020
Note: green Score (e.g. "73.2") means that the model is better than darkcloudai/huskylm-2.5-8b.

Rank the Huskylm 2.5 8B Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43470 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227