Upload model trained with Unsloth
Browse filesUpload model trained with Unsloth 2x faster
- adapter_config.json +5 -5
- adapter_model.safetensors +2 -2
    	
        adapter_config.json
    CHANGED
    
    | @@ -1,7 +1,7 @@ | |
| 1 | 
             
            {
         | 
| 2 | 
             
              "alpha_pattern": {},
         | 
| 3 | 
             
              "auto_mapping": null,
         | 
| 4 | 
            -
              "base_model_name_or_path": " | 
| 5 | 
             
              "bias": "none",
         | 
| 6 | 
             
              "fan_in_fan_out": false,
         | 
| 7 | 
             
              "inference_mode": true,
         | 
| @@ -23,13 +23,13 @@ | |
| 23 | 
             
              "rank_pattern": {},
         | 
| 24 | 
             
              "revision": null,
         | 
| 25 | 
             
              "target_modules": [
         | 
| 26 | 
            -
                " | 
| 27 | 
            -
                "down_proj",
         | 
| 28 | 
             
                "gate_proj",
         | 
| 29 | 
            -
                "v_proj",
         | 
| 30 | 
             
                "k_proj",
         | 
|  | |
| 31 | 
             
                "up_proj",
         | 
| 32 | 
            -
                " | 
|  | |
| 33 | 
             
              ],
         | 
| 34 | 
             
              "task_type": "CAUSAL_LM",
         | 
| 35 | 
             
              "use_dora": false,
         | 
|  | |
| 1 | 
             
            {
         | 
| 2 | 
             
              "alpha_pattern": {},
         | 
| 3 | 
             
              "auto_mapping": null,
         | 
| 4 | 
            +
              "base_model_name_or_path": "meta-llama/Llama-3.1-8B",
         | 
| 5 | 
             
              "bias": "none",
         | 
| 6 | 
             
              "fan_in_fan_out": false,
         | 
| 7 | 
             
              "inference_mode": true,
         | 
|  | |
| 23 | 
             
              "rank_pattern": {},
         | 
| 24 | 
             
              "revision": null,
         | 
| 25 | 
             
              "target_modules": [
         | 
| 26 | 
            +
                "q_proj",
         | 
|  | |
| 27 | 
             
                "gate_proj",
         | 
|  | |
| 28 | 
             
                "k_proj",
         | 
| 29 | 
            +
                "down_proj",
         | 
| 30 | 
             
                "up_proj",
         | 
| 31 | 
            +
                "o_proj",
         | 
| 32 | 
            +
                "v_proj"
         | 
| 33 | 
             
              ],
         | 
| 34 | 
             
              "task_type": "CAUSAL_LM",
         | 
| 35 | 
             
              "use_dora": false,
         | 
    	
        adapter_model.safetensors
    CHANGED
    
    | @@ -1,3 +1,3 @@ | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            -
            oid sha256: | 
| 3 | 
            -
            size  | 
|  | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:a58284a68812f9364019838993e28cad230631fc5ca587cf711df81ae90e958d
         | 
| 3 | 
            +
            size 3443585096
         |