athoillah commited on
Commit
11ad5ce
·
verified ·
1 Parent(s): 11d1074

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -14,14 +14,14 @@ tags:
14
  - text
15
  ---
16
 
17
- # LaBahasa 11B
18
 
19
  ## Model Information
20
  LaBahasa 11B is a multimodal LLM that combines text, audio, and image processing capabilities. Built upon OpenAI's Whisper and Meta's Llama 3.2 architectures, this model has been specifically optimized for Indonesian language understanding while maintaining English capability. The model was trained on 9 billion high quality bilingual dataset comprising Indonesian and English speech and text data.
21
 
22
  **Model Architecture**: LaBahasa 11B uses a feed-forward network to project audio embeddings from Whisper Large encoder to Llama's input embeddings, combined with image/text inputs to enable multimodal understanding.
23
 
24
- **Model Developer**: Meeting.AI and LintasArta
25
 
26
  ## Intended Use
27
  This model is intended for various NLP tasks that require text/audio/image understanding and generating Indonesian language.
@@ -103,9 +103,9 @@ print(processor.decode(outputs[0][input_len:]))
103
  **Training Time**: 25 hours
104
 
105
  ### Training Data
106
- LaBahasa 11B was trained on an extensive 9 billion high quality bilingual dataset comprising Indonesian and English speech and text data.
107
 
108
  ### Training Procedure
109
- LaBahasa 11B was trained on customized training methodology modifications to enhance:
110
  * Image input processing capabilities through integration with Llama 3.2's vision features
111
  * Indonesian language understanding and generation
 
14
  - text
15
  ---
16
 
17
+ # LABahasa 11B
18
 
19
  ## Model Information
20
  LaBahasa 11B is a multimodal LLM that combines text, audio, and image processing capabilities. Built upon OpenAI's Whisper and Meta's Llama 3.2 architectures, this model has been specifically optimized for Indonesian language understanding while maintaining English capability. The model was trained on 9 billion high quality bilingual dataset comprising Indonesian and English speech and text data.
21
 
22
  **Model Architecture**: LaBahasa 11B uses a feed-forward network to project audio embeddings from Whisper Large encoder to Llama's input embeddings, combined with image/text inputs to enable multimodal understanding.
23
 
24
+ **Model Developer**: Meeting.AI and Lintasarta
25
 
26
  ## Intended Use
27
  This model is intended for various NLP tasks that require text/audio/image understanding and generating Indonesian language.
 
103
  **Training Time**: 25 hours
104
 
105
  ### Training Data
106
+ LABahasa 11B was trained on an extensive 9 billion high quality bilingual dataset comprising Indonesian and English speech and text data.
107
 
108
  ### Training Procedure
109
+ LABahasa 11B was trained on customized training methodology modifications to enhance:
110
  * Image input processing capabilities through integration with Llama 3.2's vision features
111
  * Indonesian language understanding and generation