|  | @@ -45,7 +45,7 @@ class LLMGenerator:
 | 
	
		
			
				|  |  |          prompt = CONVERSATION_SUMMARY_PROMPT
 | 
	
		
			
				|  |  |          prompt_with_empty_context = prompt.format(context='')
 | 
	
		
			
				|  |  |          prompt_tokens = TokenCalculator.get_num_tokens(model, prompt_with_empty_context)
 | 
	
		
			
				|  |  | -        rest_tokens = llm_constant.max_context_token_length[model] - prompt_tokens - max_tokens
 | 
	
		
			
				|  |  | +        rest_tokens = llm_constant.max_context_token_length[model] - prompt_tokens - max_tokens - 1
 | 
	
		
			
				|  |  |  
 | 
	
		
			
				|  |  |          context = ''
 | 
	
		
			
				|  |  |          for message in messages:
 | 
	
	
		
			
				|  | @@ -56,6 +56,9 @@ class LLMGenerator:
 | 
	
		
			
				|  |  |              if rest_tokens - TokenCalculator.get_num_tokens(model, context + message_qa_text) > 0:
 | 
	
		
			
				|  |  |                  context += message_qa_text
 | 
	
		
			
				|  |  |  
 | 
	
		
			
				|  |  | +        if not context:
 | 
	
		
			
				|  |  | +            return '[message too long, no summary]'
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  |          prompt = prompt.format(context=context)
 | 
	
		
			
				|  |  |  
 | 
	
		
			
				|  |  |          llm: StreamableOpenAI = LLMBuilder.to_llm(
 |