ptrdvn commited on
Commit
7749d67
·
verified ·
1 Parent(s): 17ef67a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +45 -0
README.md CHANGED
@@ -1,6 +1,51 @@
 
 
 
 
 
 
 
 
1
  # Kurage
2
 
3
  <p align="center">
4
  <img src="https://cdn-uploads.huggingface.co/production/uploads/64b63f8ad57e02621dc93c8b/_SkPhhsg40juscfv9dU4v.jpeg" alt="An anime image of a pink and blue jellyfish surrounded by bubbles" width=500 style="border: 5px solid #3d3c3c"/>
5
  </p>
6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - ja
5
+ pipeline_tag: text-generation
6
+ tags:
7
+ - RAG
8
+ ---
9
  # Kurage
10
 
11
  <p align="center">
12
  <img src="https://cdn-uploads.huggingface.co/production/uploads/64b63f8ad57e02621dc93c8b/_SkPhhsg40juscfv9dU4v.jpeg" alt="An anime image of a pink and blue jellyfish surrounded by bubbles" width=500 style="border: 5px solid #3d3c3c"/>
13
  </p>
14
 
15
+ Kurage is a multipurpose RAG model from [Lightblue](https://huggingface.co/lightblue).
16
+
17
+ This version of the model has been trained to perform RAG in Japanese.
18
+
19
+ # Features / How to use
20
+
21
+ * **Multi-chunk RAG**
22
+
23
+ This model can take multiple contexts and a question as input, and it will first output the references of the relevant contexts before outputting an answer to the question.
24
+
25
+ * **Single-chunk RAG**
26
+
27
+ This model can also take a single context and a question as input, and it will determine whether it can answer the question based on the context, outputting an answer if it can. This allows for parallel computing of multiple contexts at the same time.
28
+
29
+ * **Answer extension**
30
+
31
+ By default, this model is trained to output the shortest possible answer to a question. However, if you require a longer answer, you can prompt the model to write a longer answer by writing " <<Long>>" after your question.
32
+
33
+ * **Multilinguality**
34
+
35
+ We have trained our model to be able to answer questions in Japanese based on texts in other languages too!
36
+
37
+ * **Q&A generation**
38
+
39
+ This model can also generate questions and answers based on a piece of text. This can be useful for pre-indexing a database or fine-tuning IR models that will then be used for RAG.
40
+
41
+
42
+ # Training data
43
+
44
+ We trained on chunks sourced from the documents in [MADLAD-400](https://huggingface.co/datasets/allenai/MADLAD-400) dataset that
45
+ had been evaluated to contain a higher amount of educational information according to a state-of-the-art LLM.
46
+
47
+ We took chunks of size 250 tokens, 500 tokens, and 1000 tokens randomly for each document.
48
+
49
+ We then used these chunks to generate questions and answers based on this text using a state-of-the-art LLM.
50
+
51
+ Finally, we selected negatives for each chunk using the similarity from the dense embeddings of the [BAAI/bge-m3](https://huggingface.co/BAAI/bge-m3) model.