| 
					
				 | 
			
			
				@@ -45,7 +45,7 @@ class LLMGenerator: 
			 | 
		
	
		
			
				 | 
				 | 
			
			
				         prompt = CONVERSATION_SUMMARY_PROMPT 
			 | 
		
	
		
			
				 | 
				 | 
			
			
				         prompt_with_empty_context = prompt.format(context='') 
			 | 
		
	
		
			
				 | 
				 | 
			
			
				         prompt_tokens = TokenCalculator.get_num_tokens(model, prompt_with_empty_context) 
			 | 
		
	
		
			
				 | 
				 | 
			
			
				-        rest_tokens = llm_constant.max_context_token_length[model] - prompt_tokens - max_tokens 
			 | 
		
	
		
			
				 | 
				 | 
			
			
				+        rest_tokens = llm_constant.max_context_token_length[model] - prompt_tokens - max_tokens - 1 
			 | 
		
	
		
			
				 | 
				 | 
			
			
				  
			 | 
		
	
		
			
				 | 
				 | 
			
			
				         context = '' 
			 | 
		
	
		
			
				 | 
				 | 
			
			
				         for message in messages: 
			 | 
		
	
	
		
			
				| 
					
				 | 
			
			
				@@ -56,6 +56,9 @@ class LLMGenerator: 
			 | 
		
	
		
			
				 | 
				 | 
			
			
				             if rest_tokens - TokenCalculator.get_num_tokens(model, context + message_qa_text) > 0: 
			 | 
		
	
		
			
				 | 
				 | 
			
			
				                 context += message_qa_text 
			 | 
		
	
		
			
				 | 
				 | 
			
			
				  
			 | 
		
	
		
			
				 | 
				 | 
			
			
				+        if not context: 
			 | 
		
	
		
			
				 | 
				 | 
			
			
				+            return '[message too long, no summary]' 
			 | 
		
	
		
			
				 | 
				 | 
			
			
				+ 
			 | 
		
	
		
			
				 | 
				 | 
			
			
				         prompt = prompt.format(context=context) 
			 | 
		
	
		
			
				 | 
				 | 
			
			
				  
			 | 
		
	
		
			
				 | 
				 | 
			
			
				         llm: StreamableOpenAI = LLMBuilder.to_llm( 
			 |