diff --git a/zhenxun/services/llm/adapters/gemini.py b/zhenxun/services/llm/adapters/gemini.py index 109569d2..720bdbe4 100644 --- a/zhenxun/services/llm/adapters/gemini.py +++ b/zhenxun/services/llm/adapters/gemini.py @@ -119,8 +119,7 @@ class GeminiAdapter(BaseAdapter): system_instruction_parts = [{"text": msg.content}] elif isinstance(msg.content, list): system_instruction_parts = [ - await converter.convert_part(part) - for part in msg.content + await converter.convert_part(part) for part in msg.content ] continue diff --git a/zhenxun/services/llm/api.py b/zhenxun/services/llm/api.py index 3bb45897..d56d9a91 100644 --- a/zhenxun/services/llm/api.py +++ b/zhenxun/services/llm/api.py @@ -302,7 +302,6 @@ async def _generate_image_from_message( messages = await normalize_to_llm_messages(message) async with await get_model_instance(model) as model_instance: - response = await model_instance.generate_response(messages, config=config) if not response.images: diff --git a/zhenxun/services/llm/session.py b/zhenxun/services/llm/session.py index c7df1476..660c2952 100644 --- a/zhenxun/services/llm/session.py +++ b/zhenxun/services/llm/session.py @@ -587,9 +587,7 @@ class AI: final_config = LLMGenerationConfig() if max_validation_retries is None: - max_validation_retries = ( - get_llm_config().client_settings.structured_retries - ) + max_validation_retries = get_llm_config().client_settings.structured_retries resolved_model_name = self._resolve_model_name(model or self.config.model)