|
--- |
|
license: apache-2.0 |
|
language: |
|
- en |
|
base_model: |
|
- TinyLlama/TinyLlama-1.1B-Chat-v1.0 |
|
library_name: transformers |
|
tags: |
|
- text-generation-inference |
|
--- |
|
Tiny Llama Project Guide |
|
This repository provides a comprehensive guide for students and researchers to experiment with the TinyLlama-1.1B-Chat-v1.0 model, an open-source language model developed by the TinyLlama organization. The goal is to enable accessible AI experimentation without any fees or personal information requirements. |
|
Model Details |
|
|
|
Model: TinyLlama-1.1B-Chat-v1.0 |
|
Source: Hugging Face - TinyLlama/TinyLlama-1.1B-Chat-v1.0 |
|
Organization: TinyLlama |
|
Description: A lightweight, efficient 1.1B parameter model optimized for chat and text generation tasks, suitable for low-resource environments like laptops with 16GB RAM. |
|
License: Refer to the model's official Hugging Face page for licensing details (typically Apache 2.0). |
|
|
|
Resources |
|
|
|
Code: Includes scripts for downloading the model, fine-tuning, and running a Flask-based chat UI. |
|
Dataset: A small JSON dataset for fine-tuning tests. |
|
Loss Plot: Training loss plot from fine-tuning (loss_plot.png). |
|
|
|
|
|
Usage |
|
This repository provides: |
|
|
|
A Flask app for local inference with a user-friendly chat interface. |
|
Fine-tuning scripts using LoRA for efficient training. |
|
Detailed setup instructions in document.txt. |
|
|
|
Note: Model weights are not included in this repository. Users must download them from the official Hugging Face repository using their access token. |
|
Attribution |
|
This project uses the TinyLlama-1.1B-Chat-v1.0 model by the TinyLlama organization. All credits for the model go to the original authors. For more details, visit the TinyLlama Hugging Face page. |