Cosmosage V2 by Tijmen2

 ยป  All LLMs  ยป  Tijmen2  ยป  Cosmosage V2   URL Share it on

  4bit   Autotrain compatible Base model:finetune:mistralai/... Base model:mistralai/mistral-7...   Conversational   Cosmology   Dataset:teknium/openhermes-2.5   En   Endpoints compatible   Gptq   Mistral   Physics   Pytorch   Quantized   Region:us   Safetensors
Model Card on HF ๐Ÿค—: https://huggingface.co/Tijmen2/cosmosage_v2 

Cosmosage V2 Benchmarks

Cosmosage V2 (Tijmen2/cosmosage_v2)

Cosmosage V2 Parameters and Internals

Model Type 
text generation
Additional Notes 
The model is designed as a natural-language cosmology assistant but continues to have reliability issues in ensuring factually accurate responses.
Training Details 
Data Sources:
thousands of papers and textbooks
Methodology:
first underwent continued pretraining on papers and textbooks, then fine-tuned on synthetically-generated question-answer pairs
Hardware Used:
4xA100 (80 GB) at the Center for Computational Astrophysics (CfCA), National Astronomical Observatory of Japan (NAOJ)
Input Output 
Input Format:
inst chat template with U+2581 Lower One Eighth Block Unicode Character to separate sections
Output Format:
text in Q&A format
LLM NameCosmosage V2
Repository ๐Ÿค—https://huggingface.co/Tijmen2/cosmosage_v2 
Base Model(s)  mistralai/Mistral-7B-v0.1   mistralai/Mistral-7B-v0.1
Model Size7b
Required VRAM4.2 GB
Updated2025-02-22
MaintainerTijmen2
Model Typemistral
Model Files  4.2 GB   7.7 GB   14.5 GB   14.5 GB
Supported Languagesen
GPTQ QuantizationYes
Quantization Typegptq|4bit|8bit
Model ArchitectureMistralForCausalLM
Licensemit
Context Length32768
Model Max Length32768
Transformers Version4.38.0.dev0
Tokenizer ClassLlamaTokenizer
Padding Token</s>
Vocabulary Size32000
Torch Data Typebfloat16

Best Alternatives to Cosmosage V2

Best Alternatives
Context / RAM
Downloads
Likes
Mistral 7B Instruct V0.3 GPTQ32K / 4.2 GB640870
Mistral 7B Instruct V0.2 GPTQ32K / 4.2 GB11675850
...ral 7B Instruct V0.3 GPTQ 4bit32K / 4.2 GB128318
...ephyr 7B Beta Channelwise Gptq32K / 4 GB108670
...istral 7B Pruned50 GPTQ Marlin32K / 4 GB760
...lai Mistral 7B V0.1 4 Bit Gptq32K / 4.2 GB1410
...stral 7B Instruct V0.3 GPTQ 8B32K / 7.7 GB2261
...l Neural Chat 7B V3.8 Bit Gptq32K / 7.7 GB810
...lai Mistral 7B V0.1 8 Bit Gptq32K / 7.7 GB810
...l Neural Chat 7B V3.4 Bit Gptq32K / 4.2 GB800
Note: green Score (e.g. "73.2") means that the model is better than Tijmen2/cosmosage_v2.

Rank the Cosmosage V2 Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 43470 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241227