You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

Chat Model

This is a custom chat model fine-tuned for conversational AI. The model is based on LLaMA architecture and is specifically designed for Arabic and English conversations.

Model Details

  • Architecture: LLaMA
  • Task: Text Generation
  • Language: Arabic/English
  • License: MIT
  • Model Size: Large
  • Training Data: Custom conversational data
  • Optimization: Quantized (int8)

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("AMRALMughira/chat-model")
tokenizer = AutoTokenizer.from_pretrained("AMRALMughira/chat-model")

# Example input
input_text = "مرحبا كيف حالك؟"
inputs = tokenizer(input_text, return_tensors="pt")
outputs = model.generate(
    **inputs,
    max_length=256,
    temperature=0.7,
    do_sample=True,
    top_p=0.95
)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)

Inference API

This model is compatible with Hugging Face's Inference API. You can use it with the following endpoint:

POST https://api-inference.huggingface.co/models/AMRALMughira/chat-model

Model Performance

  • Optimized for conversational tasks
  • Supports both Arabic and English
  • Fast response times
  • High-quality responses

Requirements

  • PyTorch
  • Transformers
  • CUDA (optional for GPU acceleration)
Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support