File size: 1,679 Bytes
6c5eb8a 03afb93 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 |
---
license: apache-2.0
language:
- en
base_model:
- TinyLlama/TinyLlama-1.1B-Chat-v1.0
library_name: transformers
tags:
- text-generation-inference
---
Tiny Llama Project Guide
This repository provides a comprehensive guide for students and researchers to experiment with the TinyLlama-1.1B-Chat-v1.0 model, an open-source language model developed by the TinyLlama organization. The goal is to enable accessible AI experimentation without any fees or personal information requirements.
Model Details
Model: TinyLlama-1.1B-Chat-v1.0
Source: Hugging Face - TinyLlama/TinyLlama-1.1B-Chat-v1.0
Organization: TinyLlama
Description: A lightweight, efficient 1.1B parameter model optimized for chat and text generation tasks, suitable for low-resource environments like laptops with 16GB RAM.
License: Refer to the model's official Hugging Face page for licensing details (typically Apache 2.0).
Resources
Code: Includes scripts for downloading the model, fine-tuning, and running a Flask-based chat UI.
Dataset: A small JSON dataset for fine-tuning tests.
Loss Plot: Training loss plot from fine-tuning (loss_plot.png).
Usage
This repository provides:
A Flask app for local inference with a user-friendly chat interface.
Fine-tuning scripts using LoRA for efficient training.
Detailed setup instructions in document.txt.
Note: Model weights are not included in this repository. Users must download them from the official Hugging Face repository using their access token.
Attribution
This project uses the TinyLlama-1.1B-Chat-v1.0 model by the TinyLlama organization. All credits for the model go to the original authors. For more details, visit the TinyLlama Hugging Face page. |