|  | @@ -27,7 +27,6 @@ class AppMode(enum.Enum):
 | 
	
		
			
				|  |  |  class PromptTransform:
 | 
	
		
			
				|  |  |      def get_prompt(self,
 | 
	
		
			
				|  |  |                     app_mode: str,
 | 
	
		
			
				|  |  | -                   app_model_config: AppModelConfig,
 | 
	
		
			
				|  |  |                     pre_prompt: str,
 | 
	
		
			
				|  |  |                     inputs: dict,
 | 
	
		
			
				|  |  |                     query: str,
 | 
	
	
		
			
				|  | @@ -36,10 +35,9 @@ class PromptTransform:
 | 
	
		
			
				|  |  |                     memory: Optional[BaseChatMemory],
 | 
	
		
			
				|  |  |                     model_instance: BaseLLM) -> \
 | 
	
		
			
				|  |  |              Tuple[List[PromptMessage], Optional[List[str]]]:
 | 
	
		
			
				|  |  | -        model_mode = app_model_config.model_dict['mode']
 | 
	
		
			
				|  |  |  
 | 
	
		
			
				|  |  |          app_mode_enum = AppMode(app_mode)
 | 
	
		
			
				|  |  | -        model_mode_enum = ModelMode(model_mode)
 | 
	
		
			
				|  |  | +        model_mode_enum = model_instance.model_mode
 | 
	
		
			
				|  |  |  
 | 
	
		
			
				|  |  |          prompt_rules = self._read_prompt_rules_from_file(self._prompt_file_name(app_mode, model_instance))
 | 
	
		
			
				|  |  |  
 |