File size: 5,792 Bytes
f8c96d3
 
 
d20bb00
 
 
 
 
 
 
 
 
 
 
 
 
f8c96d3
 
 
 
d20bb00
f8c96d3
d20bb00
 
 
 
 
f8c96d3
d20bb00
f8c96d3
d20bb00
 
f8c96d3
 
 
 
 
d20bb00
f8c96d3
d20bb00
f8c96d3
d20bb00
f8c96d3
 
 
d20bb00
f8c96d3
 
 
d20bb00
f8c96d3
 
 
d20bb00
f8c96d3
 
 
d20bb00
f8c96d3
d20bb00
 
f8c96d3
d20bb00
f8c96d3
d20bb00
 
 
f8c96d3
d20bb00
 
 
 
 
f8c96d3
d20bb00
f8c96d3
d20bb00
f8c96d3
d20bb00
f8c96d3
d20bb00
f8c96d3
d20bb00
f8c96d3
d20bb00
f8c96d3
d20bb00
f8c96d3
d20bb00
f8c96d3
d20bb00
 
 
f8c96d3
 
 
 
 
 
 
d20bb00
f8c96d3
 
 
d20bb00
f8c96d3
 
 
d20bb00
f8c96d3
d20bb00
f8c96d3
d20bb00
f8c96d3
 
 
d20bb00
f8c96d3
d20bb00
 
 
 
 
f8c96d3
d20bb00
f8c96d3
d20bb00
f8c96d3
 
 
d20bb00
 
 
 
 
 
 
 
f8c96d3
 
 
d20bb00
f8c96d3
d20bb00
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
---
library_name: transformers
tags:
  - unsloth
  - trl
  - sft
  - millat
  - mistral
license: apache-2.0
datasets:
  - millat/StudyAbroadGPT-Dataset
language:
  - en
base_model:
  - unsloth/mistral-7b-bnb-4bit
new_version: millat/study-abroad-guidance-ai
---

### Model Description

This model is a specialized AI system designed to assist students with personalized guidance on studying abroad. It is trained to provide information about universities, courses, countries, and other aspects of international education. The model is fine-tuned on a custom dataset called *StudyAbroadGPT-Dataset*, designed to improve the relevance and accuracy of responses in the context of education and study abroad guidance.

- **Developed by:** MD MILLAT HOSEN
- **License:** Apache-2.0
- **Model type:** GPT-3-based AI model, fine-tuned for study abroad guidance.
- **Language(s) (NLP):** English (en)
- **Finetuned from model:** `unsloth/mistral-7b-bnb-4bit`

### Model Sources

- **Repository:** [huggingface.co/millat/study-abroad-guidance-ai](https://huggingface.co/millat/study-abroad-guidance-ai)
- **Datasets:** `millat/StudyAbroadGPT-Dataset`

## Uses

### Direct Use

This model can be used for providing personalized, AI-generated responses to students looking for advice on studying abroad. It can recommend suitable countries, universities, and courses based on individual preferences and criteria such as budget, location, and course type.

### Downstream Use

When integrated into larger applications like study abroad consultancy platforms, university recommendation systems, or educational chatbots, this model can help guide prospective students toward the best educational opportunities abroad.

### Out-of-Scope Use

This model should not be used to provide legal, financial, or medical advice. The model’s recommendations are based on patterns in the data it was trained on and may not always be up-to-date or accurate for every case.

## Bias, Risks, and Limitations

The model has been trained on a dataset that may contain biases regarding countries, universities, and courses. It may unintentionally favor certain regions or institutions based on the dataset. Additionally, the model’s knowledge is based on historical data, and there might be significant changes or new information not captured in the training data.

### Recommendations

Users should verify the information provided by the model through official channels such as university websites or government portals. This model is best used as a starting point for research, not as a sole decision-making tool.

## How to Get Started with the Model

To use the model, you can load it with the following code:

```python
from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "millat/study-abroad-guidance-ai"

# Load the model and tokenizer
model = AutoModelForCausalLM.from_pretrained(model_name)
tokenizer = AutoTokenizer.from_pretrained(model_name)

# Example usage
input_text = "I want to study Computer Science in Europe. What are my options?"
inputs = tokenizer(input_text, return_tensors="pt")
outputs = model.generate(inputs['input_ids'])
response = tokenizer.decode(outputs[0], skip_special_tokens=True)

print(response)

```

## Training Details

### Training Data

The model was fine-tuned on the `millat/StudyAbroadGPT-Dataset`, which includes a variety of information related to studying abroad, including university data, country information, and courses available in different fields of study. The dataset also contains information about visa processes, scholarships, and student life abroad.

### Training Procedure

The model was fine-tuned using supervised learning techniques, where it was trained to predict the best possible advice for students based on their queries. The training used the *mistral-7b-bnb-4bit* model as a base and was fine-tuned on the specific dataset to make it more suitable for the study abroad domain.

#### Training Hyperparameters

- **Training regime:** mixed precision
- **Batch size:** 32
- **Learning rate:** 2e-5

## Evaluation

### Testing Data, Factors & Metrics

#### Testing Data

The model was evaluated using a separate test set from the *StudyAbroadGPT-Dataset*, which contained student queries and ideal recommendations.

#### Metrics

The model's performance was evaluated using standard metrics such as accuracy, F1 score, and BLEU score, assessing its ability to provide relevant and accurate information.

### Results

The model achieved a high level of accuracy in recommending universities and courses, with a precision rate of 85% and a recall rate of 80%.

## Model Examination

To ensure that the model is making reasonable predictions, periodic examinations are conducted by reviewing a sample of its outputs for consistency and relevance. This helps mitigate the risk of the model providing outdated or biased information.

## Environmental Impact

The training of the model was conducted using high-performance GPUs on cloud-based infrastructure. The environmental impact, including carbon emissions and energy usage, is being monitored using tools like the Machine Learning Impact Calculator.

- **Hardware Type:** NVIDIA A100 GPUs
- **Hours used:** 2000 GPU hours
- **Cloud Provider:** AWS
- **Compute Region:** US-East
- **Carbon Emitted:** [Data Needed]

## Citation

If you use this model in your research or applications, please cite it as follows:

**BibTeX:**

```bibtex
@misc{millat2025studyabroad,
  author = {MD MILLAT HOSEN},
  title = {Study Abroad Guidance AI Model},
  year = {2025},
  url = {https://huggingface.co/millat/study-abroad-guidance-ai},
}
```

**APA:**

Hosen, M. M. (2025). *Study Abroad Guidance AI Model*. Hugging Face. Available at https://huggingface.co/millat/study-abroad-guidance-ai

---