Update modeling_llama3.py
Browse files- modeling_llama3.py +3 -3
modeling_llama3.py
CHANGED
@@ -403,8 +403,8 @@ class Llama3ForCausalLM(MllamaPreTrainedModel, GenerationMixin):
|
|
403 |
attentions=outputs.attentions,
|
404 |
)
|
405 |
|
406 |
-
AutoModelForCausalLM.register(MllamaTextConfig, Llama3ForCausalLM)
|
407 |
-
transformers.Llama3ForCausalLM = Llama3ForCausalLM
|
408 |
|
409 |
class Llama3ForConditionalGeneration(MllamaPreTrainedModel, GenerationMixin):
|
410 |
config_class = Llama3Config
|
@@ -420,7 +420,7 @@ class Llama3ForConditionalGeneration(MllamaPreTrainedModel, GenerationMixin):
|
|
420 |
self.pad_token_id = self.config.pad_token_id if self.config.pad_token_id is not None else -1
|
421 |
|
422 |
self.vision_model = MllamaVisionModel._from_config(config.vision_config)
|
423 |
-
self.language_model =
|
424 |
self.embed_tokens = Llama3Embedding(config.audio_config, config.text_config)
|
425 |
self.multi_modal_projector = nn.Linear(
|
426 |
config.vision_config.vision_output_dim,
|
|
|
403 |
attentions=outputs.attentions,
|
404 |
)
|
405 |
|
406 |
+
#AutoModelForCausalLM.register(MllamaTextConfig, Llama3ForCausalLM)
|
407 |
+
#transformers.Llama3ForCausalLM = Llama3ForCausalLM
|
408 |
|
409 |
class Llama3ForConditionalGeneration(MllamaPreTrainedModel, GenerationMixin):
|
410 |
config_class = Llama3Config
|
|
|
420 |
self.pad_token_id = self.config.pad_token_id if self.config.pad_token_id is not None else -1
|
421 |
|
422 |
self.vision_model = MllamaVisionModel._from_config(config.vision_config)
|
423 |
+
self.language_model = MllamaForCausalLM._from_config(config.text_config)
|
424 |
self.embed_tokens = Llama3Embedding(config.audio_config, config.text_config)
|
425 |
self.multi_modal_projector = nn.Linear(
|
426 |
config.vision_config.vision_output_dim,
|