NxMobileLM-1.5B-SFT

Model Description

NxMobileLM-1.5B-SFT is a fine-tuned version of the base model Qwen2.5-1.5B, optimized for mobile and edge applications. This model has been trained on proprietary instruction datasets curated to enhance performance in natural language understanding and generation tasks tailored to specific applications.

Key Features:

  • Base Model: Qwen2.5-1.5B
  • Parameter Count: 1.5 billion
  • Fine-tuning Objective: Supervised fine-tuning (SFT) on instruction datasets.
  • Specialization: Lightweight and efficient performance for mobile environments.
  • Multilingual Support: Designed to handle multiple languages effectively, enabling robust cross-lingual capabilities for diverse applications.

Model Details

Training Data

The model was fine-tuned using a proprietary dataset designed for diverse instruction-following tasks, including question answering, summarization, and dialogue. The dataset emphasizes:

  • Multi-domain generalization
  • Task-specific instruction understanding
  • Multilingual Coverage: Training data includes samples from several major languages to enhance cross-lingual understanding.

Training Configuration

  • Framework: PyTorch
  • Optimizer: AdamW
  • Learning Rate: 5e-5
  • Batch Size: 128
  • Epochs: 3
  • Mixed Precision: FP16

Evaluation

The model was evaluated on a variety of benchmarks, demonstrating:

  • High Accuracy: Achieves strong performance across general natural language tasks.
  • Efficiency: Optimized for low-latency inference on edge devices.
  • Multilingual Competence: Strong performance across multiple languages, making it suitable for global applications.

Performance Comparison

Open-LLM Leaderboard

On January 15, 2025, NxMobileLM-1.5B-SFT was ranked among the top 10 edge device models with fewer than 3 billion parameters and achieved the first rank for models with under 2 billion parameters, according to the OpenLLM leaderboard. image/png

P-MMEval

To evaluate the multilingual capabilities of the model, We conducted evaluations on several benchmarks across three languages: English (en), Japanese (ja), and Vietnamese (vi). For detailed benchmark information, refer to P-MMEval.

Benchmark Llama-3.2-1B-Instruct SmolLM2-1.7B-Instruct Qwen2.5-1.5B-Instruct NxMobileLM-1.5B-SFT
mifeval-en 44.79 43.75 50 57.29
mifeval-ja 22.92 23.96 29.17 30.21
mifeval-vi 30.21 25 28.12 46.88
mmmlu-EN-US 35.25 42.5 45.5 45.25
mmmlu-JA-JP 31.5 26.25 36.00 41.00
mmmlu-VI-VT 22.75 22.25 39.00 38.00
xnli-en 35.83 35.83 59.17 66.67
xnli-ja 34.17 35.83 52.5 57.5
xnli-vi 37.5 34.17 45.83 55.83
Average 32.21 31.61 42.93 48.07

LightEval

The table below compares NxMobileLM-1.5B-SFT with other instruction-tuned models using various benchmarks. Results were obtained using the lighteval evaluation framework and are referenced from Hugging Face TB:

Metric SmolLM2-1.7B-Instruct Llama-1B-Instruct Qwen2.5-1.5B-Instruct SmolLM1-1.7B-Instruct NxMobileLM-1.5B-SFT
IFEval (Average prompt/inst) 56.7 53.5 47.4 23.1 64.2
HellaSwag 66.1 56.1 60.9 55.5 63.57
ARC (Average) 51.7 41.6 46.2 43.7 45.21
PIQA 74.4 72.3 73.2 71.6 72.91
MMLU-Pro (MCF) 19.3 12.7 24.2 11.7 15.43
BBH (3-shot) 32.2 27.6 35.3 25.7 31.44
GSM8K (5-shot) 48.2 26.8 42.8 4.62 59.51
Average 49.8 41.5 47.1 33.7 50.3

Limitations

While NxMobileLM-1.5B-SFT excels in many areas, it may not perform well on tasks outside the scope of the fine-tuned dataset. Biases inherent in the training data may also affect outcomes.

Intended Use

NxMobileLM-1.5B-SFT is designed for use in:

  • Mobile virtual assistants
  • Real-time language-based applications
  • Compact edge AI solutions
  • Multilingual Scenarios: Supporting applications that require cross-lingual communication and understanding.

Misuse Warning: The model is not intended for use in generating harmful, biased, or illegal content.

How to Use

Here is a sample code snippet to load and use the model:

from transformers import AutoModelForCausalLM, AutoTokenizer

# Load the model and tokenizer
model_name = "NTQAI/NxMobileLM-1.5B-SFT"
model = AutoModelForCausalLM.from_pretrained(model_name)
tokenizer = AutoTokenizer.from_pretrained(model_name)

# Example usage
inputs = tokenizer("What is the capital of Vietnam?", return_tensors="pt")
outputs = model.generate(**inputs, max_length=50)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Citation

If you use this model in your research, please cite it as:

@misc{NxMobileLM-1.5B-SFT,
  title={NxMobileLM-1.5B-SFT},
  author={NTQAI},
  year={2025},
  url={https://huggingface.co/NTQAI/NxMobileLM-1.5B-SFT},
}

License

This model is licensed under MIT.

Contact

For questions or issues, please contact us via website: https://ntq.ai

Downloads last month
64
Safetensors
Model size
1.54B params
Tensor type
FP16
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for NTQAI/NxMobileLM-1.5B-SFT

Base model

Qwen/Qwen2.5-1.5B
Finetuned
(41)
this model
Quantizations
2 models