From cd2db86a95f5f48c132b452fd8ba01239d002857 Mon Sep 17 00:00:00 2001 From: Maciej Pienczyn Date: Fri, 16 Jan 2026 20:37:09 +0100 Subject: [PATCH] =?UTF-8?q?fix(zopk):=20Naprawiono=20b=C5=82=C4=85d=20syst?= =?UTF-8?q?em=5Fprompt=20w=20AI=20extraction?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Problem: GeminiService.generate_text() nie obsługuje parametru system_prompt Rozwiązanie: Połączono system_prompt z user_prompt w jeden full_prompt Dotyczy: ekstrakcji faktów, encji i relacji z artykułów ZOPK Co-Authored-By: Claude Opus 4.5 --- zopk_knowledge_service.py | 8 +++++--- 1 file changed, 5 insertions(+), 3 deletions(-) diff --git a/zopk_knowledge_service.py b/zopk_knowledge_service.py index 1b82788..32a4afe 100644 --- a/zopk_knowledge_service.py +++ b/zopk_knowledge_service.py @@ -380,15 +380,17 @@ class ZOPKKnowledgeService: Returns parsed JSON or None on error. """ try: - prompt = EXTRACTION_USER_PROMPT.format( + user_prompt = EXTRACTION_USER_PROMPT.format( chunk_text=chunk.content, source_name=source_name, published_date=published_date ) + # Combine system prompt with user prompt (Gemini doesn't support separate system_prompt) + full_prompt = f"{EXTRACTION_SYSTEM_PROMPT}\n\n---\n\n{user_prompt}" + response = self.gemini.generate_text( - prompt=prompt, - system_prompt=EXTRACTION_SYSTEM_PROMPT, + prompt=full_prompt, temperature=0.1, # Low temperature for consistency max_tokens=2000, user_id=self.user_id,