Update README.md
Browse filesremoved download requirement
README.md
CHANGED
|
@@ -80,7 +80,7 @@ from optimum.neuron import NeuronModelForCausalLM
|
|
| 80 |
#num_cores should be changed based on the instance. inf2.24xlarge has 6 neuron processors (they have two cores each) so 12 total
|
| 81 |
input_shapes = {"batch_size": 1, "sequence_length": 4096}
|
| 82 |
compiler_args = {"num_cores": 8, "auto_cast_type": 'fp16'}
|
| 83 |
-
model = NeuronModelForCausalLM.from_pretrained("SOLAR-10.7B-v1.0", export=True, **compiler_args, **input_shapes)
|
| 84 |
model.save_pretrained("SOLAR-10.7B-v1.0-neuron-24xlarge-2.16-8core-4096")
|
| 85 |
|
| 86 |
from transformers import AutoTokenizer
|
|
|
|
| 80 |
#num_cores should be changed based on the instance. inf2.24xlarge has 6 neuron processors (they have two cores each) so 12 total
|
| 81 |
input_shapes = {"batch_size": 1, "sequence_length": 4096}
|
| 82 |
compiler_args = {"num_cores": 8, "auto_cast_type": 'fp16'}
|
| 83 |
+
model = NeuronModelForCausalLM.from_pretrained("upstage/SOLAR-10.7B-v1.0", export=True, **compiler_args, **input_shapes)
|
| 84 |
model.save_pretrained("SOLAR-10.7B-v1.0-neuron-24xlarge-2.16-8core-4096")
|
| 85 |
|
| 86 |
from transformers import AutoTokenizer
|