File size: 1,679 Bytes
6c5eb8a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
03afb93
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
---
license: apache-2.0
language:
- en
base_model:
- TinyLlama/TinyLlama-1.1B-Chat-v1.0
library_name: transformers
tags:
- text-generation-inference
---
Tiny Llama Project Guide
This repository provides a comprehensive guide for students and researchers to experiment with the TinyLlama-1.1B-Chat-v1.0 model, an open-source language model developed by the TinyLlama organization. The goal is to enable accessible AI experimentation without any fees or personal information requirements.
Model Details

Model: TinyLlama-1.1B-Chat-v1.0
Source: Hugging Face - TinyLlama/TinyLlama-1.1B-Chat-v1.0
Organization: TinyLlama
Description: A lightweight, efficient 1.1B parameter model optimized for chat and text generation tasks, suitable for low-resource environments like laptops with 16GB RAM.
License: Refer to the model's official Hugging Face page for licensing details (typically Apache 2.0).

Resources

Code: Includes scripts for downloading the model, fine-tuning, and running a Flask-based chat UI.
Dataset: A small JSON dataset for fine-tuning tests.
Loss Plot: Training loss plot from fine-tuning (loss_plot.png).


Usage
This repository provides:

A Flask app for local inference with a user-friendly chat interface.
Fine-tuning scripts using LoRA for efficient training.
Detailed setup instructions in document.txt.

Note: Model weights are not included in this repository. Users must download them from the official Hugging Face repository using their access token.
Attribution
This project uses the TinyLlama-1.1B-Chat-v1.0 model by the TinyLlama organization. All credits for the model go to the original authors. For more details, visit the TinyLlama Hugging Face page.