Spaces:
				
			
			
	
			
			
					
		Running
		
			on 
			
			CPU Upgrade
	
	
	
			
			
	
	
	
	
		
		
					
		Running
		
			on 
			
			CPU Upgrade
	fix(hchat): update internal endpoint (#1716)
Browse files- chart/env/prod.yaml +5 -5
- scripts/updateLocalEnv.ts +1 -1
    	
        chart/env/prod.yaml
    CHANGED
    
    | @@ -158,7 +158,7 @@ envVars: | |
| 158 | 
             
                    "endpoints": [
         | 
| 159 | 
             
                      {
         | 
| 160 | 
             
                        "type": "openai",
         | 
| 161 | 
            -
                        "baseURL": "https:// | 
| 162 | 
             
                      }
         | 
| 163 | 
             
                    ]
         | 
| 164 | 
             
                  },
         | 
| @@ -192,7 +192,7 @@ envVars: | |
| 192 | 
             
                    "endpoints": [
         | 
| 193 | 
             
                      {
         | 
| 194 | 
             
                        "type": "openai",
         | 
| 195 | 
            -
                        "baseURL": "https:// | 
| 196 | 
             
                      }
         | 
| 197 | 
             
                    ]
         | 
| 198 | 
             
                  },
         | 
| @@ -260,7 +260,7 @@ envVars: | |
| 260 | 
             
                    "endpoints": [
         | 
| 261 | 
             
                      {
         | 
| 262 | 
             
                        "type": "openai",
         | 
| 263 | 
            -
                        "baseURL": "https:// | 
| 264 | 
             
                      }
         | 
| 265 | 
             
                    ]
         | 
| 266 | 
             
                  },
         | 
| @@ -279,7 +279,7 @@ envVars: | |
| 279 | 
             
                    "endpoints": [
         | 
| 280 | 
             
                      {
         | 
| 281 | 
             
                        "type": "openai",
         | 
| 282 | 
            -
                        "baseURL": "https:// | 
| 283 | 
             
                        "multimodal": {
         | 
| 284 | 
             
                          "image": {
         | 
| 285 | 
             
                            "maxSizeInMB": 10,
         | 
| @@ -596,7 +596,7 @@ envVars: | |
| 596 | 
             
                ]
         | 
| 597 | 
             
              HF_ORG_ADMIN: '644171cfbd0c97265298aa99'
         | 
| 598 | 
             
              HF_ORG_EARLY_ACCESS: '5e67bd5b1009063689407478'
         | 
| 599 | 
            -
              HF_API_ROOT: 'https:// | 
| 600 | 
             
            infisical:
         | 
| 601 | 
             
              enabled: true
         | 
| 602 | 
             
              env: "prod-us-east-1"
         | 
|  | |
| 158 | 
             
                    "endpoints": [
         | 
| 159 | 
             
                      {
         | 
| 160 | 
             
                        "type": "openai",
         | 
| 161 | 
            +
                        "baseURL": "https://internal.api-inference.huggingface.co/models/deepseek-ai/DeepSeek-R1-Distill-Qwen-32B/v1"
         | 
| 162 | 
             
                      }
         | 
| 163 | 
             
                    ]
         | 
| 164 | 
             
                  },
         | 
|  | |
| 192 | 
             
                    "endpoints": [
         | 
| 193 | 
             
                      {
         | 
| 194 | 
             
                        "type": "openai",
         | 
| 195 | 
            +
                        "baseURL": "https://internal.api-inference.huggingface.co/models/nvidia/Llama-3.1-Nemotron-70B-Instruct-HF/v1"
         | 
| 196 | 
             
                      }
         | 
| 197 | 
             
                    ]
         | 
| 198 | 
             
                  },
         | 
|  | |
| 260 | 
             
                    "endpoints": [
         | 
| 261 | 
             
                      {
         | 
| 262 | 
             
                        "type": "openai",
         | 
| 263 | 
            +
                        "baseURL": "https://internal.api-inference.huggingface.co/models/Qwen/Qwen2.5-Coder-32B-Instruct/v1"
         | 
| 264 | 
             
                      }
         | 
| 265 | 
             
                    ]
         | 
| 266 | 
             
                  },
         | 
|  | |
| 279 | 
             
                    "endpoints": [
         | 
| 280 | 
             
                      {
         | 
| 281 | 
             
                        "type": "openai",
         | 
| 282 | 
            +
                        "baseURL": "https://internal.api-inference.huggingface.co/models/meta-llama/Llama-3.2-11B-Vision-Instruct/v1",
         | 
| 283 | 
             
                        "multimodal": {
         | 
| 284 | 
             
                          "image": {
         | 
| 285 | 
             
                            "maxSizeInMB": 10,
         | 
|  | |
| 596 | 
             
                ]
         | 
| 597 | 
             
              HF_ORG_ADMIN: '644171cfbd0c97265298aa99'
         | 
| 598 | 
             
              HF_ORG_EARLY_ACCESS: '5e67bd5b1009063689407478'
         | 
| 599 | 
            +
              HF_API_ROOT: 'https://internal.api-inference.huggingface.co/models'
         | 
| 600 | 
             
            infisical:
         | 
| 601 | 
             
              enabled: true
         | 
| 602 | 
             
              env: "prod-us-east-1"
         | 
    	
        scripts/updateLocalEnv.ts
    CHANGED
    
    | @@ -26,7 +26,7 @@ let full_config = `${PUBLIC_CONFIG}\n${SECRET_CONFIG}`; | |
| 26 |  | 
| 27 | 
             
            // replace the internal proxy url with the public endpoint
         | 
| 28 | 
             
            full_config = full_config.replaceAll(
         | 
| 29 | 
            -
            	"https:// | 
| 30 | 
             
            	"https://api-inference.huggingface.co"
         | 
| 31 | 
             
            );
         | 
| 32 |  | 
|  | |
| 26 |  | 
| 27 | 
             
            // replace the internal proxy url with the public endpoint
         | 
| 28 | 
             
            full_config = full_config.replaceAll(
         | 
| 29 | 
            +
            	"https://internal.api-inference.huggingface.co",
         | 
| 30 | 
             
            	"https://api-inference.huggingface.co"
         | 
| 31 | 
             
            );
         | 
| 32 |  | 

