A newer version of this model is available: FlameF0X/SnowflakeCore-G0-Release-2.5

SnowflakeCore-G0-Release

This is the initial release of the pre-trained model SnowflakeCore-G0 series language models, trained on the DialogMLM-50K dataset with optimized memory usage.

SUPPORT ME

You can support me via https://ko-fi.com/flamef0x

Model details

  • Architecture: SnowflakeCore
  • Hidden size: 384
  • Number of attention heads: 6
  • Number of layers: 4
  • Feed-forward dimension: 768
  • Maximum sequence length: 384
  • Vocabulary size: 30522

HuggingFace Transformers Compatibility

This model is fully compatible with the HuggingFace Transformers library. You can load it using:

from transformers import AutoConfig, AutoModel, AutoTokenizer

tokenizer = AutoTokenizer.from_pretrained("FlameF0X/SnowflakeCore-G0-Release")
config = AutoConfig.from_pretrained("FlameF0X/SnowflakeCore-G0-Release")
model = AutoModel.from_pretrained("FlameF0X/SnowflakeCore-G0-Release")

Memory Optimization Techniques

  • Mixed precision training
  • Gradient accumulation (4 steps)
  • Fused QKV projection
  • Pre-norm architecture
  • Weight tying between embedding and output layers
  • Half-precision model storage

The model weights are stored in both PyTorch (.bin) and safetensors format for improved security, loading efficiency, and compatibility.

Downloads last month
6
Safetensors
Model size
16.6M params
Tensor type
F16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train FlameF0X/SnowflakeCore-G0-Release

Space using FlameF0X/SnowflakeCore-G0-Release 1

Collection including FlameF0X/SnowflakeCore-G0-Release