Spaces:
				
			
			
	
			
			
		Runtime error
		
	
	
	
			
			
	
	
	
	
		
		
		Runtime error
		
	
		zR
		
	commited on
		
		
					Commit 
							
							Β·
						
						5f7b72c
	
1
								Parent(s):
							
							83a2412
								
add suggested_hardware
Browse files
    	
        README.md
    CHANGED
    
    | 
         @@ -5,6 +5,7 @@ colorFrom: yellow 
     | 
|
| 5 | 
         
             
            colorTo: purple
         
     | 
| 6 | 
         
             
            sdk: gradio
         
     | 
| 7 | 
         
             
            sdk_version: 4.41.0
         
     | 
| 
         | 
|
| 8 | 
         
             
            app_port: 7860
         
     | 
| 9 | 
         
             
            app_file: app.py
         
     | 
| 10 | 
         
             
            models:
         
     | 
| 
         | 
|
| 5 | 
         
             
            colorTo: purple
         
     | 
| 6 | 
         
             
            sdk: gradio
         
     | 
| 7 | 
         
             
            sdk_version: 4.41.0
         
     | 
| 8 | 
         
            +
            suggested_hardware: a100-large
         
     | 
| 9 | 
         
             
            app_port: 7860
         
     | 
| 10 | 
         
             
            app_file: app.py
         
     | 
| 11 | 
         
             
            models:
         
     | 
    	
        app.py
    CHANGED
    
    | 
         @@ -29,7 +29,7 @@ class StopOnTokens(StoppingCriteria): 
     | 
|
| 29 | 
         
             
                    return False
         
     | 
| 30 | 
         | 
| 31 | 
         | 
| 32 | 
         
            -
            @spaces.GPU(duration= 
     | 
| 33 | 
         
             
            def predict(history, prompt, max_length, top_p, temperature):
         
     | 
| 34 | 
         
             
                stop = StopOnTokens()
         
     | 
| 35 | 
         
             
                messages = []
         
     | 
| 
         @@ -83,9 +83,9 @@ with gr.Blocks() as demo: 
     | 
|
| 83 | 
         
             
                    </div>
         
     | 
| 84 | 
         
             
                    <div style="text-align: center; font-size: 15px; font-weight: bold; color: red; margin-bottom: 20px;">
         
     | 
| 85 | 
         
             
                        β οΈ Due to the limitations of Huggingface ZERO GPUs, in order to output 10K characters in one go, 
         
     | 
| 86 | 
         
            -
                        we need to request a 5 
     | 
| 87 | 
         
             
                        This will result in you only being able to use it once every 4 hours. 
         
     | 
| 88 | 
         
            -
                        If you plan to use it long-term, please consider deploying the model yourself.
         
     | 
| 89 | 
         
             
                    </div>
         
     | 
| 90 | 
         
             
                    """
         
     | 
| 91 | 
         
             
                )
         
     | 
| 
         | 
|
| 29 | 
         
             
                    return False
         
     | 
| 30 | 
         | 
| 31 | 
         | 
| 32 | 
         
            +
            @spaces.GPU(duration=280)
         
     | 
| 33 | 
         
             
            def predict(history, prompt, max_length, top_p, temperature):
         
     | 
| 34 | 
         
             
                stop = StopOnTokens()
         
     | 
| 35 | 
         
             
                messages = []
         
     | 
| 
         | 
|
| 83 | 
         
             
                    </div>
         
     | 
| 84 | 
         
             
                    <div style="text-align: center; font-size: 15px; font-weight: bold; color: red; margin-bottom: 20px;">
         
     | 
| 85 | 
         
             
                        β οΈ Due to the limitations of Huggingface ZERO GPUs, in order to output 10K characters in one go, 
         
     | 
| 86 | 
         
            +
                        we need to request a 4-5 minute quota each time. 
         
     | 
| 87 | 
         
             
                        This will result in you only being able to use it once every 4 hours. 
         
     | 
| 88 | 
         
            +
                        If you plan to use it long-term, please consider deploying the model or fork this space yourself.
         
     | 
| 89 | 
         
             
                    </div>
         
     | 
| 90 | 
         
             
                    """
         
     | 
| 91 | 
         
             
                )
         
     |