|  | @@ -534,12 +534,14 @@ class OllamaLargeLanguageModel(LargeLanguageModel):
 | 
	
		
			
				|  |  |                  ),
 | 
	
		
			
				|  |  |                  ParameterRule(
 | 
	
		
			
				|  |  |                      name='num_gpu',
 | 
	
		
			
				|  |  | -                    label=I18nObject(en_US="Num GPU"),
 | 
	
		
			
				|  |  | +                    label=I18nObject(en_US="GPU Layers"),
 | 
	
		
			
				|  |  |                      type=ParameterType.INT,
 | 
	
		
			
				|  |  | -                    help=I18nObject(en_US="The number of layers to send to the GPU(s). "
 | 
	
		
			
				|  |  | -                                          "On macOS it defaults to 1 to enable metal support, 0 to disable."),
 | 
	
		
			
				|  |  | -                    min=0,
 | 
	
		
			
				|  |  | -                    max=1
 | 
	
		
			
				|  |  | +                    help=I18nObject(en_US="The number of layers to offload to the GPU(s). "
 | 
	
		
			
				|  |  | +                                          "On macOS it defaults to 1 to enable metal support, 0 to disable."
 | 
	
		
			
				|  |  | +                                          "As long as a model fits into one gpu it stays in one. "
 | 
	
		
			
				|  |  | +                                          "It does not set the number of GPU(s). "),
 | 
	
		
			
				|  |  | +                    min=-1,
 | 
	
		
			
				|  |  | +                    default=1
 | 
	
		
			
				|  |  |                  ),
 | 
	
		
			
				|  |  |                  ParameterRule(
 | 
	
		
			
				|  |  |                      name='num_thread',
 |