Cristian Sas
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -16,272 +16,91 @@ metrics:
|
|
16 |
- character
|
17 |
- code_eval
|
18 |
---
|
19 |
-
|
20 |
|
21 |
-
|
22 |
-
LLMLit is a high-performance, multilingual large language model (LLM) fine-tuned from Meta's Llama 3.1 8B Instruct model. Designed for both English and Romanian NLP tasks, LLMLit leverages advanced instruction-following capabilities to provide accurate, context-aware, and efficient results across diverse applications.
|
23 |
-
|
24 |
-
## Model Details
|
25 |
|
26 |
-
|
27 |
-
|
28 |
|
29 |
-
|
30 |
-
|
31 |
-
- **Shared by:** LLMLit Community
|
32 |
-
- **Model type:** Large Language Model (Instruction-tuned)
|
33 |
-
- **Languages:** English (en), Romanian (ro)
|
34 |
-
- **License:** MIT
|
35 |
-
- **Fine-tuned from model:** meta-llama/Llama-3.1-8B-Instruct
|
36 |
|
37 |
-
|
38 |
-
- **Repository:** [GitHub Repository Link](https://github.com/PyThaGoAI/LLMLit)
|
39 |
-
- **Paper:** [To be published]
|
40 |
-
- **Demo:** [Coming Soon)
|
41 |
|
42 |
-
##
|
|
|
43 |
|
44 |
-
|
45 |
-
LLMLit
|
46 |
-
-
|
47 |
-
|
48 |
-
- Summarizing articles, reports, or documents
|
49 |
-
- Answering complex questions with context sensitivity
|
50 |
|
51 |
-
|
52 |
-
|
53 |
-
|
54 |
-
-
|
55 |
-
|
56 |
-
|
|
|
|
|
|
|
57 |
|
58 |
-
|
59 |
-
LLMLit is not suitable for:
|
60 |
-
- Malicious or unethical applications, such as spreading misinformation
|
61 |
-
- Highly sensitive or critical decision-making without human oversight
|
62 |
-
- Tasks requiring real-time, low-latency performance in constrained environments
|
63 |
|
64 |
-
##
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
65 |
|
66 |
-
|
67 |
-
- LLMLit inherits biases present in the training data. It may produce outputs that reflect societal or cultural biases.
|
68 |
|
69 |
-
|
70 |
-
|
71 |
-
|
|
|
|
|
|
|
72 |
|
73 |
-
|
74 |
-
|
75 |
-
-
|
76 |
|
77 |
-
|
78 |
-
- Always review model outputs for accuracy, especially in sensitive applications.
|
79 |
-
- Fine-tune or customize for domain-specific tasks to minimize risks.
|
80 |
|
81 |
-
##
|
82 |
-
|
83 |
|
84 |
```python
|
85 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
86 |
|
87 |
-
#
|
88 |
model = AutoModelForCausalLM.from_pretrained("llmlit/LLMLit-0.2-8B-Instruct")
|
89 |
tokenizer = AutoTokenizer.from_pretrained("llmlit/LLMLit-0.2-8B-Instruct")
|
90 |
|
91 |
-
#
|
92 |
inputs = tokenizer("Your prompt here", return_tensors="pt")
|
93 |
outputs = model.generate(**inputs, max_length=100)
|
94 |
-
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|
95 |
```
|
96 |
|
97 |
-
|
98 |
-
|
99 |
-
### Training Data
|
100 |
-
LLMLit is fine-tuned on a diverse dataset containing bilingual (English and Romanian) content, ensuring both linguistic accuracy and cultural relevance.
|
101 |
-
|
102 |
-
### Training Procedure
|
103 |
-
#### Preprocessing
|
104 |
-
- Data was filtered for high-quality, instruction-based examples.
|
105 |
-
- Augmentation techniques were used to balance linguistic domains.
|
106 |
-
|
107 |
-
#### Training Hyperparameters
|
108 |
-
- **Training regime:** Mixed precision (fp16)
|
109 |
-
- **Batch size:** 512
|
110 |
-
- **Epochs:** 3
|
111 |
-
- **Learning rate:** 2e-5
|
112 |
-
|
113 |
-
#### Speeds, Sizes, Times
|
114 |
-
- **Checkpoint size:** ~16GB
|
115 |
-
- **Training time:** Approx. 1 week on 8 A100 GPUs
|
116 |
-
|
117 |
-
## Evaluation
|
118 |
-
|
119 |
-
### Testing Data, Factors & Metrics
|
120 |
-
#### Testing Data
|
121 |
-
Evaluation was conducted on multilingual benchmarks, such as:
|
122 |
-
- FLORES-101 (Translation accuracy)
|
123 |
-
- HELM (Instruction-following capabilities)
|
124 |
-
|
125 |
-
#### Factors
|
126 |
-
Evaluation considered:
|
127 |
-
- Linguistic fluency
|
128 |
-
- Instruction adherence
|
129 |
-
- Contextual understanding
|
130 |
-
|
131 |
-
#### Metrics
|
132 |
-
- BLEU for translation tasks
|
133 |
-
- ROUGE-L for summarization
|
134 |
-
- Human evaluation scores for instruction tasks
|
135 |
-
|
136 |
-
### Results
|
137 |
-
LLMLit achieves state-of-the-art performance on instruction-following tasks for English and Romanian, with BLEU scores surpassing comparable models.
|
138 |
-
|
139 |
-
#### Summary
|
140 |
-
LLMLit excels in bilingual NLP tasks, offering robust performance across diverse domains while maintaining instruction adherence and linguistic accuracy.
|
141 |
-
|
142 |
-
## Model Examination
|
143 |
-
Efforts to interpret the model include:
|
144 |
-
- Attention visualization
|
145 |
-
- Prompt engineering guides
|
146 |
-
- Bias audits
|
147 |
-
|
148 |
-
## Environmental Impact
|
149 |
-
Training LLMLit resulted in estimated emissions of ~200 kg CO2eq. Carbon offsets were purchased to mitigate environmental impact. Future optimizations aim to reduce energy consumption.
|
150 |
-
|
151 |
-
|
152 |
-
|
153 |
-
LLMLit – The Cutting-Edge AI Assistant for Performance Analysis and Prediction
|
154 |
-
|
155 |
-
LLMLIT – Coming Soon
|
156 |
-
The premium version of the LLama 3 model, LLMLIT, sets a new standard in AI through advanced customization and innovative technologies, offering ideal solutions for a wide range of industries and applications.
|
157 |
-
|
158 |
-
This version integrates Retrieval-Augmented Generation (RAG) capabilities, ensuring fast and accurate access to relevant and specialized information.
|
159 |
-
|
160 |
-
Key Features of LLMLIT:
|
161 |
-
|
162 |
-
Advanced Customization: Tailored to meet the specific needs of each user, delivering optimized solutions.
|
163 |
-
|
164 |
-
Enhanced RAG Integration: Support for multiple domains and complex data sources.
|
165 |
-
|
166 |
-
Innovative Frontend and Backend:Frontend: Intuitive, customizable interfaces with user-centric interactions.
|
167 |
-
|
168 |
-
Backend: Top-tier performance, rapid data processing, and efficient task management.
|
169 |
-
|
170 |
-
Extensive Community Integrations
|
171 |
-
LLMLIT supports a wide range of platforms and applications, offering unparalleled flexibility:
|
172 |
-
|
173 |
-
Web & Desktop: Open WebUI, HTML UI, Ollama GUI, LMstudio, MindMac, Ollama Spring.
|
174 |
-
|
175 |
-
Mobile: Native apps such as Enchanted, macAI, Ollama Telegram Bot, and Ollama RAG Chatbot.
|
176 |
-
|
177 |
-
CLI & Terminal: Advanced plugins for Emacs, Vim, and tools like ShellOracle and typechat-cli.
|
178 |
-
|
179 |
-
Extensions & Plugins: Raycast Extensions, Obsidian Plugins, Ollama for Discord, and more.
|
180 |
-
|
181 |
-
Package Managers: Integration with Pacman, Gentoo, Nix, and Flox.
|
182 |
-
|
183 |
-
Enterprise Solutions & Advanced AI
|
184 |
-
|
185 |
-
AI Frameworks and Chatbot UI: Hollama, Saddle, big-AGI, Cheshire Cat, Amica.
|
186 |
-
|
187 |
-
Backend RAG Integration: LangChain, LangChainGo, Haystack, and Semantic Kernel.
|
188 |
-
|
189 |
-
Developer Support: VSCode extensions, QodeAssist for Qt Creator, and Ollama support for multiple programming languages (Java, Python, C++, etc.).
|
190 |
-
|
191 |
-
Team and Multi-Agent Applications: AnythingLLM, crewAI, and BrainSoup.
|
192 |
-
|
193 |
-
Cross-Platform Performance
|
194 |
-
LLMLIT delivers advanced interoperability:
|
195 |
-
|
196 |
-
MacOS Native: OllamaSwift, macAI, and support for Apple Vision Pro.
|
197 |
-
|
198 |
-
Windows/Linux: Docker-native and containerized apps like ARGO and StreamDeploy.
|
199 |
-
|
200 |
-
Mobile Applications: Ollama Telegram Bot, Ollama Discord Bot, and Ollama RAG Chatbot.
|
201 |
-
|
202 |
-
Open Ecosystem: Integration with popular platforms such as Google Mesop, Firebase, and SAP ABAP.
|
203 |
-
|
204 |
-
The Future of AI is Here 🚀
|
205 |
-
LLMLIT revolutionizes how we work with large language models, offering a scalable, powerful, and adaptable platform ready to meet the most demanding needs with innovation, flexibility, and superior performance.
|
206 |
-
|
207 |
-
|
208 |
-
|
209 |
-
🌟 Themes and Agents: Revolutionizing AI-driven Applications
|
210 |
-
|
211 |
-
The integration of AI-powered technologies into development tools is rapidly transforming how applications are built and deployed. With LLMLit as the core engine, this suite of tools offers groundbreaking possibilities, from low-code app building to advanced conversational agents.
|
212 |
-
|
213 |
-
AI-Driven Development in Your Terminal 🚀
|
214 |
-
|
215 |
-
Design full-stack web applications with AI-powered capabilities directly from your terminal. This environment is built for large, real-world tasks, allowing developers to prompt, run, edit, and deploy web apps with seamless integration into your workflow.
|
216 |
-
|
217 |
-
|
218 |
-
|
219 |
-
Low-Code App Builder for RAG and Multi-Agent AI Applications 🔧
|
220 |
-
|
221 |
-
Python-based and agnostic to any model, API, or database, this platform simplifies the development of complex AI-driven applications, including Retrieval-Augmented Generation (RAG) and multi-agent AI systems.
|
222 |
-
|
223 |
-
It empowers developers to create powerful apps without needing extensive coding knowledge, making it ideal for businesses and researchers who want to implement sophisticated AI without the overhead.
|
224 |
-
|
225 |
-
Open Voice OS and HiveMind Installer 🔊
|
226 |
-
|
227 |
-
Ansible-based: Set up and configure your Open Voice OS and HiveMind installer effortlessly with an intuitive, text-based user interface.
|
228 |
-
|
229 |
-
This provides an easy installation and setup process, especially designed for voice-enabled applications and environments that rely on multi-agent systems.
|
230 |
-
|
231 |
-
Generative UI: AI-Powered Search Engine 🔍
|
232 |
-
|
233 |
-
Harness the power of a generative UI for your search engines. This AI-powered tool offers contextual searches and adaptive results, providing users with an efficient and intelligent way to explore content and data. It can be embedded in various systems like websites or apps to improve user experience.
|
234 |
-
|
235 |
-
|
236 |
-
|
237 |
-
🔧 LitSeek-R1: Distilled Version
|
238 |
-
|
239 |
-
A lighter, distilled version of the powerful LLMLit model, LitSeek-R1 maintains the same robust capabilities but with optimized performance for faster, more efficient responses. Perfect for applications requiring speed and low-latency operations.
|
240 |
-
|
241 |
-
|
242 |
-
|
243 |
-
Agent and Application Themes
|
244 |
-
|
245 |
-
🌐 LitAgentWeb-ui
|
246 |
-
|
247 |
-
Direct Interaction with LLMLit: No complex installations required! This theme allows users to interact with LLMLit through a simple, intuitive web interface, making it ideal for applications that need to be accessed directly from a browser.
|
248 |
-
|
249 |
-
Whether you're building a customer support system or a virtual assistant, AgentWeb-ui provides a fast and simple experience.
|
250 |
-
|
251 |
-
💬 LibreChat
|
252 |
-
|
253 |
-
Multi-AI Integration: LibreChat is a powerful chatbot that integrates multiple AI models, including LLMLit. It provides a natural conversational experience, whether for customer support or virtual assistance.
|
254 |
-
|
255 |
-
Its flexibility and scalability make it suitable for a wide range of applications, helping businesses manage automated communication at scale.
|
256 |
-
|
257 |
-
🖥️ LITflow
|
258 |
-
|
259 |
-
Low-Code Platform for Custom Apps: Langflow is a low-code solution for creating custom applications that integrate seamlessly with LLMLit.
|
260 |
-
|
261 |
-
It excels in building RAG-based applications, combining search and content generation to deliver smarter, faster solutions for complex environments. It's perfect for anyone looking to integrate advanced AI into their applications without the complexity of traditional development.
|
262 |
-
|
263 |
-
📱 NextChat
|
264 |
-
|
265 |
-
Cross-Platform Conversational App: NextChat is a fast, easy-to-use application that enables smooth, interactive conversations across different platforms.
|
266 |
-
|
267 |
-
By integrating LLMLit, NextChat offers a dynamic interactive experience, ideal for users who want to have effective conversations with AI, whether for personal use, business, or service centers.
|
268 |
-
|
269 |
-
🗣️ VoiceLit
|
270 |
-
|
271 |
-
Voice Interaction Capabilities: Extend LLMLit’s abilities into the voice realm with VoiceLit. This extension brings AI-driven voice support to your applications, whether they’re for personal assistants or service centers.
|
272 |
-
|
273 |
-
It enhances accessibility and interactivity, making it essential for creating voice-enabled AI applications.
|
274 |
-
|
275 |
-
🌍 Web-llm-chat
|
276 |
-
|
277 |
-
Run LLMLit Directly in the Browser: With Web-llm-chat, users can run LLMLit directly in their browser, bypassing the need for servers.
|
278 |
-
|
279 |
-
This ensures maximum privacy and speed, offering a confidential and fast interaction experience. It’s perfect for applications where confidentiality and performance are of utmost importance.
|
280 |
-
|
281 |
-
The Future of AI Interaction 🌐💡
|
282 |
-
|
283 |
-
These themes and agents open up a wide array of possibilities, allowing businesses, developers, and individuals to easily integrate LLMLit into their systems. Whether it's building a simple chatbot or a highly sophisticated voice-enabled app, LLMLit offers the flexibility and power to transform the way we interact with AI technology. 🔥
|
284 |
|
|
|
285 |
|
286 |
|
287 |
|
|
|
16 |
- character
|
17 |
- code_eval
|
18 |
---
|
19 |
+
Am reformulat și aranjat textul pentru a fi mai clar, bine structurat și ușor de citit. Vezi varianta îmbunătățită mai jos:
|
20 |
|
21 |
+
---
|
|
|
|
|
|
|
22 |
|
23 |
+
# **LLMLit – Model Card**
|
24 |
+
📌 *High-performance multilingual LLM for English & Romanian NLP tasks*
|
25 |
|
26 |
+
🔗 [LLMLit on Hugging Face](https://huggingface.co/LLMLit)
|
27 |
+
🔗 [LitSeekR1 on Hugging Face](https://huggingface.co/LLMLit/LitSeekR1)
|
|
|
|
|
|
|
|
|
|
|
28 |
|
29 |
+
---
|
|
|
|
|
|
|
30 |
|
31 |
+
## **🔍 Quick Summary**
|
32 |
+
**LLMLit** este un model de limbaj mare (LLM) performant, multilingv, optimizat din **Meta’s Llama 3.1 8B Instruct**. Este conceput pentru **task-uri NLP în limba engleză și română**, având capacități avansate de **urmărire a instrucțiunilor, înțelegere contextuală și generare de conținut precis**.
|
33 |
|
34 |
+
## **📌 Model Details**
|
35 |
+
🔹 **Descriere:** LLMLit poate fi utilizat pentru **generare de conținut, sumarizare, răspuns la întrebări și multe altele**.
|
36 |
+
🔹 **Fine-tuning:** Modelul a fost antrenat pentru **adherarea la instrucțiuni de înaltă calitate și o mai bună înțelegere a contextului**.
|
37 |
+
🔹 **Utilizatori țintă:** Dezvoltatori, cercetători și companii care au nevoie de **soluții NLP fiabile**.
|
|
|
|
|
38 |
|
39 |
+
| Caracteristici | Detalii |
|
40 |
+
|----------------|---------|
|
41 |
+
| 🏢 **Dezvoltat de** | LLMLit Development Team |
|
42 |
+
| 💰 **Finanțare** | Contribuții open-source & sponsori privați |
|
43 |
+
| 🌍 **Limbaje** | Engleză (en), Română (ro) |
|
44 |
+
| 🏷 **Licență** | MIT |
|
45 |
+
| 🔗 **Model de bază** | `meta-llama/Llama-3.1-8B-Instruct` |
|
46 |
+
| 📂 **Resurse** | [GitHub Repository](#) / Paper: *To be published* |
|
47 |
+
| 🚀 **Demo** | *Coming Soon* |
|
48 |
|
49 |
+
---
|
|
|
|
|
|
|
|
|
50 |
|
51 |
+
## **💡 Utilizări principale**
|
52 |
+
### ✅ **Utilizare directă**
|
53 |
+
LLMLit poate fi aplicat la:
|
54 |
+
✔️ Generarea de răspunsuri asemănătoare celor umane
|
55 |
+
✔️ Traducere între **engleză și română**
|
56 |
+
✔️ Sumarizarea articolelor, rapoartelor și documentelor
|
57 |
+
✔️ Răspuns la întrebări complexe cu sensibilitate la context
|
58 |
+
|
59 |
+
### 🚀 **Utilizare avansată (fine-tuning & integrare)**
|
60 |
+
LLMLit poate fi optimizat pentru:
|
61 |
+
🗨️ **Chatboți & asistenți virtuali**
|
62 |
+
📚 **Instrumente educaționale bilingve**
|
63 |
+
⚖️ **Analiza documentelor legale/medicale**
|
64 |
+
🛒 **Automatizare în e-commerce & suport clienți**
|
65 |
+
|
66 |
+
### ❌ **Utilizări nerecomandate**
|
67 |
+
⛔ Aplicații neetice (dezinformare, manipulare)
|
68 |
+
⛔ Luarea deciziilor critice fără supervizare umană
|
69 |
+
⛔ Task-uri care necesită **performanță în timp real**
|
70 |
|
71 |
+
---
|
|
|
72 |
|
73 |
+
## **⚠️ Bias, Riscuri și Limitări**
|
74 |
+
🔍 **Bias:** Modelul poate reflecta bias-urile existente în datele de antrenament.
|
75 |
+
⚠️ **Riscuri:** Poate genera informații inexacte sau neconforme.
|
76 |
+
📌 **Limitări:**
|
77 |
+
- Performanța depinde de **calitatea prompturilor**.
|
78 |
+
- Înțelegere limitată a domeniilor **foarte tehnice sau de nișă**.
|
79 |
|
80 |
+
🔹 **Recomandări:**
|
81 |
+
✔️ Revizuirea output-ului pentru **aplicații sensibile**.
|
82 |
+
✔️ Fine-tuning pentru sarcini specifice pentru **minimizarea riscurilor**.
|
83 |
|
84 |
+
---
|
|
|
|
|
85 |
|
86 |
+
## **🚀 Cum să începi cu LLMLit**
|
87 |
+
Pentru a utiliza LLMLit, instalează librăriile necesare și încarcă modelul:
|
88 |
|
89 |
```python
|
90 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
91 |
|
92 |
+
# Încarcă modelul și tokenizer-ul
|
93 |
model = AutoModelForCausalLM.from_pretrained("llmlit/LLMLit-0.2-8B-Instruct")
|
94 |
tokenizer = AutoTokenizer.from_pretrained("llmlit/LLMLit-0.2-8B-Instruct")
|
95 |
|
96 |
+
# Generează text
|
97 |
inputs = tokenizer("Your prompt here", return_tensors="pt")
|
98 |
outputs = model.generate(**inputs, max_length=100)
|
|
|
99 |
```
|
100 |
|
101 |
+
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
102 |
|
103 |
+
🔗 **Mai multe detalii:** [LLMLit on Hugging Face](https://huggingface.co/LLMLit) 🚀
|
104 |
|
105 |
|
106 |
|