nordabiz/audit_ai_service.py
Maciej Pienczyn 5030b71beb
Some checks are pending
NordaBiz Tests / Unit & Integration Tests (push) Waiting to run
NordaBiz Tests / E2E Tests (Playwright) (push) Blocked by required conditions
NordaBiz Tests / Smoke Tests (Production) (push) Blocked by required conditions
NordaBiz Tests / Send Failure Notification (push) Blocked by required conditions
chore: update Author to Maciej Pienczyn, InPi sp. z o.o. across all files
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
2026-04-10 08:20:47 +02:00

1394 lines
58 KiB
Python

"""
Audit AI Service
=================
Centralny serwis AI do analizy wyników audytów i generowania
priorytetowanych akcji z treścią gotową do wdrożenia.
Obsługiwane typy audytów:
- SEO (PageSpeed, on-page, technical, local SEO)
- GBP (Google Business Profile completeness)
- Social Media (presence across platforms)
Używa Gemini API (via gemini_service.py) do generowania analiz.
Author: Maciej Pienczyn, InPi sp. z o.o.
Created: 2026-02-07
"""
import hashlib
import json
import logging
import re
import requests
from datetime import datetime, timedelta
from html import unescape
from database import (
SessionLocal, Company, CompanyWebsiteAnalysis, CompanySocialMedia,
CompanyCitation, AuditAction, AuditAICache
)
from youtube_service import YouTubeService
from crux_service import CrUXService
from benchmark_service import BenchmarkService
logger = logging.getLogger(__name__)
# Cache expiry: 7 days
CACHE_EXPIRY_DAYS = 7
# Gemini Structured Output schema for audit analysis responses.
# Uses Gemini API schema format (STRING, INTEGER, OBJECT, ARRAY).
# This enforces valid JSON output from the model, eliminating manual parsing.
AUDIT_ANALYSIS_SCHEMA = {
'type': 'OBJECT',
'required': ['summary', 'actions'],
'properties': {
'summary': {
'type': 'STRING',
},
'actions': {
'type': 'ARRAY',
'items': {
'type': 'OBJECT',
'required': ['action_type', 'title', 'description', 'priority', 'impact_score', 'effort_score', 'platform'],
'properties': {
'action_type': {'type': 'STRING'},
'title': {'type': 'STRING'},
'description': {'type': 'STRING'},
'priority': {
'type': 'STRING',
'enum': ['critical', 'high', 'medium', 'low'],
},
'impact_score': {'type': 'INTEGER'},
'effort_score': {'type': 'INTEGER'},
'platform': {'type': 'STRING'},
}
}
}
}
}
def _get_gemini_service():
"""Get the initialized Gemini service instance."""
from gemini_service import get_gemini_service
service = get_gemini_service()
if not service:
raise RuntimeError("Gemini service not initialized")
return service
def _hash_data(data: dict) -> str:
"""Generate SHA256 hash of audit data for cache invalidation."""
serialized = json.dumps(data, sort_keys=True, default=str)
return hashlib.sha256(serialized.encode()).hexdigest()
# ============================================================
# SEO AUDIT DATA COLLECTION
# ============================================================
def _collect_seo_data(db, company) -> dict:
"""Collect SEO audit data for AI analysis."""
analysis = db.query(CompanyWebsiteAnalysis).filter(
CompanyWebsiteAnalysis.company_id == company.id
).order_by(CompanyWebsiteAnalysis.seo_audited_at.desc()).first()
if not analysis or not analysis.seo_audited_at:
return {}
citations = db.query(CompanyCitation).filter(
CompanyCitation.company_id == company.id
).all()
# Security headers check (live request)
security_headers = {}
if company.website:
try:
resp = requests.head(company.website, timeout=5, allow_redirects=True)
headers = resp.headers
security_headers = {
'has_hsts': 'strict-transport-security' in headers,
'has_csp': 'content-security-policy' in headers,
'has_x_frame_options': 'x-frame-options' in headers,
'has_x_content_type_options': 'x-content-type-options' in headers,
'security_headers_count': sum([
'strict-transport-security' in headers,
'content-security-policy' in headers,
'x-frame-options' in headers,
'x-content-type-options' in headers,
]),
}
except Exception:
pass
# Image format analysis (from existing data if available)
image_formats = {}
if company.website:
try:
resp = requests.get(company.website, timeout=10, allow_redirects=True)
if resp.status_code == 200:
img_srcs = re.findall(r'<img[^>]+src=["\']([^"\']+)["\']', resp.text, re.IGNORECASE)
webp_count = sum(1 for s in img_srcs if '.webp' in s.lower())
avif_count = sum(1 for s in img_srcs if '.avif' in s.lower())
svg_count = sum(1 for s in img_srcs if '.svg' in s.lower())
modern_count = webp_count + avif_count + svg_count
legacy_count = len(img_srcs) - modern_count
image_formats = {
'total_images_found': len(img_srcs),
'webp_count': webp_count,
'avif_count': avif_count,
'svg_count': svg_count,
'modern_format_count': modern_count,
'legacy_format_count': legacy_count,
'modern_format_ratio': round(modern_count / len(img_srcs) * 100, 1) if img_srcs else 0,
}
except Exception:
pass
# CrUX field data (real user metrics from Chrome)
crux_data = {}
if company.website:
try:
crux = CrUXService()
crux_data = crux.get_field_data(company.website) or {}
except Exception as e:
logger.warning(f"CrUX error for {company.website}: {e}")
# Search Console data via OAuth (if available)
search_console_data = {}
try:
from oauth_service import OAuthService
from search_console_service import SearchConsoleService
oauth = OAuthService()
gsc_token = oauth.get_valid_token(db, company.id, 'google', 'search_console')
if gsc_token and company.website:
gsc = SearchConsoleService(gsc_token)
analytics = gsc.get_search_analytics(company.website, days=28)
if analytics:
search_console_data = analytics
logger.info(f"Search Console data for company {company.id}: {analytics.get('clicks', 0)} clicks")
except ImportError:
pass
except Exception as e:
logger.warning(f"Search Console data collection failed: {e}")
# Persist live-collected data to DB for dashboard display
try:
if security_headers:
for key, val in security_headers.items():
setattr(analysis, key, val)
if image_formats:
analysis.modern_image_count = image_formats.get('modern_format_count')
analysis.legacy_image_count = image_formats.get('legacy_format_count')
analysis.modern_image_ratio = image_formats.get('modern_format_ratio')
if crux_data:
analysis.crux_lcp_ms = crux_data.get('crux_lcp_ms')
analysis.crux_inp_ms = crux_data.get('crux_inp_ms')
analysis.crux_cls = crux_data.get('crux_cls')
analysis.crux_fcp_ms = crux_data.get('crux_fcp_ms')
analysis.crux_ttfb_ms = crux_data.get('crux_ttfb_ms')
analysis.crux_lcp_good_pct = crux_data.get('crux_lcp_ms_good_pct')
analysis.crux_inp_good_pct = crux_data.get('crux_inp_ms_good_pct')
analysis.crux_period_end = crux_data.get('crux_period_end')
if search_console_data:
analysis.gsc_clicks = search_console_data.get('clicks')
analysis.gsc_impressions = search_console_data.get('impressions')
analysis.gsc_ctr = search_console_data.get('ctr')
analysis.gsc_avg_position = search_console_data.get('position')
analysis.gsc_top_queries = search_console_data.get('top_queries', [])
analysis.gsc_period_days = search_console_data.get('period_days', 28)
db.commit()
except Exception as e:
logger.warning(f"Failed to persist live metrics for {company.name}: {e}")
db.rollback()
return {
'company_name': company.name,
'company_category': company.category.name if company.category else None,
'website': company.website,
'city': company.address_city,
# PageSpeed scores
'seo_score': analysis.pagespeed_seo_score,
'performance_score': analysis.pagespeed_performance_score,
'accessibility_score': analysis.pagespeed_accessibility_score,
'best_practices_score': analysis.pagespeed_best_practices_score,
# On-page
'meta_title': unescape(analysis.meta_title or ''),
'meta_description': unescape(analysis.meta_description or ''),
'h1_count': analysis.h1_count,
'h1_text': unescape(analysis.h1_text or ''),
'h2_count': analysis.h2_count,
'h3_count': analysis.h3_count,
'total_images': analysis.total_images,
'images_without_alt': analysis.images_without_alt,
# Technical
'has_ssl': analysis.has_ssl,
'has_sitemap': analysis.has_sitemap,
'has_robots_txt': analysis.has_robots_txt,
'has_canonical': analysis.has_canonical,
'is_indexable': analysis.is_indexable,
'is_mobile_friendly': getattr(analysis, 'is_mobile_friendly', None),
'load_time_ms': analysis.load_time_ms,
# Structured data
'has_structured_data': analysis.has_structured_data,
'structured_data_types': analysis.structured_data_types,
'has_local_business_schema': analysis.has_local_business_schema,
# Social/analytics
'has_og_tags': analysis.has_og_tags,
'has_twitter_cards': analysis.has_twitter_cards,
'has_google_analytics': analysis.has_google_analytics,
'has_google_tag_manager': analysis.has_google_tag_manager,
# Local SEO
'local_seo_score': analysis.local_seo_score,
'has_google_maps_embed': analysis.has_google_maps_embed,
'has_local_keywords': analysis.has_local_keywords,
'nap_on_website': analysis.nap_on_website,
# Core Web Vitals
'lcp_ms': analysis.largest_contentful_paint_ms,
'inp_ms': analysis.interaction_to_next_paint_ms,
'cls': float(analysis.cumulative_layout_shift) if analysis.cumulative_layout_shift else None,
# Additional performance metrics (10 missing metrics)
'fcp_ms': getattr(analysis, 'first_contentful_paint_ms', None),
'ttfb_ms': getattr(analysis, 'time_to_first_byte_ms', None),
'tbt_ms': getattr(analysis, 'total_blocking_time_ms', None),
'speed_index': getattr(analysis, 'speed_index_ms', None),
'meta_title_length': len(analysis.meta_title or ''),
'meta_description_length': len(analysis.meta_description or ''),
'html_lang': analysis.html_lang,
'local_business_schema_fields': analysis.local_business_schema_fields,
# Content
'content_freshness_score': analysis.content_freshness_score,
'word_count_homepage': analysis.word_count_homepage,
# Links
'internal_links_count': analysis.internal_links_count,
'external_links_count': analysis.external_links_count,
'broken_links_count': analysis.broken_links_count,
# Citations
'citations_count': len(citations),
'citations_found': len([c for c in citations if c.status == 'found']),
# Security headers
'has_hsts': security_headers.get('has_hsts', None),
'has_csp': security_headers.get('has_csp', None),
'has_x_frame_options': security_headers.get('has_x_frame_options', None),
'has_x_content_type_options': security_headers.get('has_x_content_type_options', None),
'security_headers_count': security_headers.get('security_headers_count', None),
# Image formats
'modern_format_ratio': image_formats.get('modern_format_ratio', None),
'webp_count': image_formats.get('webp_count', None),
'legacy_image_count': image_formats.get('legacy_format_count', None),
# CrUX field data (real user metrics)
'crux_lcp_ms': crux_data.get('crux_lcp_ms'),
'crux_inp_ms': crux_data.get('crux_inp_ms'),
'crux_cls': crux_data.get('crux_cls'),
'crux_fcp_ms': crux_data.get('crux_fcp_ms'),
'crux_ttfb_ms': crux_data.get('crux_ttfb_ms'),
'crux_lcp_good_pct': crux_data.get('crux_lcp_ms_good_pct'),
'crux_inp_good_pct': crux_data.get('crux_inp_ms_good_pct'),
'crux_period_end': crux_data.get('crux_period_end'),
# Search Console (OAuth)
'gsc_clicks': search_console_data.get('clicks'),
'gsc_impressions': search_console_data.get('impressions'),
'gsc_ctr': search_console_data.get('ctr'),
'gsc_avg_position': search_console_data.get('position'),
'gsc_top_queries': search_console_data.get('top_queries', []),
'gsc_period_days': search_console_data.get('period_days', 28),
}
def _collect_gbp_data(db, company) -> dict:
"""Collect GBP audit data for AI analysis."""
try:
from gbp_audit_service import get_company_audit
audit = get_company_audit(db, company.id)
except ImportError:
audit = None
if not audit:
return {}
# Get Places API (New) enriched data from CompanyWebsiteAnalysis
analysis = db.query(CompanyWebsiteAnalysis).filter(
CompanyWebsiteAnalysis.company_id == company.id
).order_by(CompanyWebsiteAnalysis.analyzed_at.desc()).first()
# Build descriptive photo status for AI context
photo_count = audit.photo_count or 0
if photo_count == 0:
photo_status = "Brak zdjęć w profilu"
elif photo_count == 1:
photo_status = "1 zdjęcie (prawdopodobnie logo)"
elif photo_count == 2:
photo_status = f"2 zdjęcia (prawdopodobnie logo i zdjęcie w tle)"
elif photo_count < 10:
photo_status = f"{photo_count} zdjęć (logo i zdjęcie w tle prawdopodobnie ustawione, ale mało zdjęć dodatkowych)"
else:
photo_status = f"{photo_count} zdjęć (dobra ilość)"
# AI-enhanced review sentiment (if available)
review_sentiment = audit.review_sentiment
return {
'company_name': company.name,
'company_category': company.category.name if company.category else None,
'city': company.address_city,
'completeness_score': audit.completeness_score,
# Field presence
'has_name': audit.has_name,
'has_address': audit.has_address,
'has_phone': audit.has_phone,
'has_website': audit.has_website,
'has_hours': audit.has_hours,
'has_categories': audit.has_categories,
'has_photos': audit.has_photos,
'has_description': audit.has_description,
'has_services': audit.has_services,
'has_reviews': audit.has_reviews,
# Reviews
'review_count': audit.review_count,
'average_rating': float(audit.average_rating) if audit.average_rating else None,
'reviews_with_response': audit.reviews_with_response,
'reviews_without_response': audit.reviews_without_response,
'review_response_rate': float(audit.review_response_rate) if audit.review_response_rate else None,
'review_keywords': audit.review_keywords, # Top keywords from reviews (already collected)
'review_sentiment': review_sentiment, # AI-enhanced sentiment analysis
# Activity
'has_posts': audit.has_posts,
'posts_count_30d': audit.posts_count_30d,
'has_products': audit.has_products,
'has_qa': audit.has_qa,
# Photos
'photo_count': photo_count,
'logo_present': audit.logo_present,
'cover_photo_present': audit.cover_photo_present,
'photo_status': photo_status,
# NAP
'nap_consistent': audit.nap_consistent,
'nap_issues': audit.nap_issues,
# Keywords
'description_keywords': audit.description_keywords, # Already collected during audit
# Places API (New) enriched data
'primary_type': getattr(analysis, 'google_primary_type', None) if analysis else None,
'editorial_summary': getattr(analysis, 'google_editorial_summary', None) if analysis else None,
'price_level': getattr(analysis, 'google_price_level', None) if analysis else None,
'attributes': getattr(analysis, 'google_attributes', None) if analysis else None,
'photos_metadata': getattr(analysis, 'google_photos_metadata', None) if analysis else None,
}
def _collect_social_data(db, company) -> dict:
"""Collect social media audit data for AI analysis."""
profiles = db.query(CompanySocialMedia).filter(
CompanySocialMedia.company_id == company.id
).all()
all_platforms = ['facebook', 'instagram', 'linkedin', 'youtube', 'twitter', 'tiktok']
profiles_dict = {}
for p in profiles:
profiles_dict[p.platform] = {
'url': p.url,
'is_valid': p.is_valid,
'followers_count': p.followers_count,
'has_bio': p.has_bio,
'has_profile_photo': p.has_profile_photo,
'has_cover_photo': p.has_cover_photo,
'posts_count_30d': p.posts_count_30d,
'last_post_date': str(p.last_post_date) if p.last_post_date else None,
'posting_frequency_score': p.posting_frequency_score,
'engagement_rate': float(p.engagement_rate) if p.engagement_rate else None,
'profile_completeness_score': p.profile_completeness_score,
'page_name': getattr(p, 'page_name', None),
}
present = [p for p in all_platforms if p in profiles_dict]
missing = [p for p in all_platforms if p not in profiles_dict]
# Fetch YouTube API data if profile exists
youtube_data = None
if 'youtube' in profiles_dict and profiles_dict['youtube'].get('url'):
try:
yt_service = YouTubeService()
channel_id = yt_service.extract_channel_id_from_url(profiles_dict['youtube']['url'])
if channel_id:
youtube_data = yt_service.get_channel_stats(channel_id)
if youtube_data:
profiles_dict['youtube']['subscriber_count'] = youtube_data.get('subscriber_count')
profiles_dict['youtube']['view_count'] = youtube_data.get('view_count')
profiles_dict['youtube']['video_count'] = youtube_data.get('video_count')
except Exception as e:
logger.warning(f"YouTube API error: {e}")
return {
'company_name': company.name,
'company_category': company.category.name if company.category else None,
'city': company.address_city,
'platforms_present': present,
'platforms_missing': missing,
'profiles': profiles_dict,
'total_platforms': len(all_platforms),
'platforms_found': len(present),
'score': int((len(present) / len(all_platforms)) * 100) if all_platforms else 0,
}
# ============================================================
# GEMINI PROMPTS
# ============================================================
def _build_seo_prompt(data: dict) -> str:
"""Build Gemini prompt for SEO audit analysis."""
# Benchmark comparison
benchmark_section = ""
benchmarks = data.get('_benchmarks')
if benchmarks and benchmarks.get('seo'):
bm = benchmarks['seo']
cat = benchmarks.get('category_name', '?')
benchmark_section = f"""
BENCHMARK (średnia w kategorii "{cat}", {benchmarks.get('category_company_count', '?')} firm):
- Średni wynik SEO: {bm.get('avg_seo_score', '?')}/100
- Średnia wydajność: {bm.get('avg_performance_score', '?')}/100
- Średni czas ładowania: {bm.get('avg_load_time_ms', '?')} ms
Porównaj wyniki tej firmy ze średnią kategorii w analizie."""
return f"""Jesteś ekspertem SEO analizującym stronę internetową lokalnej firmy w Polsce.
DANE FIRMY:
- Nazwa: {data.get('company_name', 'N/A')}
- Branża: {data.get('company_category', 'N/A')}
- Miasto: {data.get('city', 'N/A')}
- Strona: {data.get('website', 'N/A')}
WYNIKI AUDYTU SEO:
- Wynik SEO (PageSpeed): {data.get('seo_score', 'brak')}/100
- Wydajność: {data.get('performance_score', 'brak')}/100
- Dostępność: {data.get('accessibility_score', 'brak')}/100
- Best Practices: {data.get('best_practices_score', 'brak')}/100
Core Web Vitals (lab data z PageSpeed):
- LCP: {data.get('lcp_ms', 'brak')} ms
- INP: {data.get('inp_ms', 'brak')} ms (zastąpił FID w marcu 2024)
- CLS: {data.get('cls', 'brak')}
CrUX Field Data (dane od realnych użytkowników Chrome):
- LCP (field): {data.get('crux_lcp_ms', 'brak danych')} ms ({data.get('crux_lcp_good_pct', '?')}% dobrych)
- INP (field): {data.get('crux_inp_ms', 'brak danych')} ms ({data.get('crux_inp_good_pct', '?')}% dobrych)
- CLS (field): {data.get('crux_cls', 'brak danych')}
- FCP (field): {data.get('crux_fcp_ms', 'brak danych')} ms
- TTFB (field): {data.get('crux_ttfb_ms', 'brak danych')} ms
- Okres pomiarowy: do {data.get('crux_period_end', 'brak')}
UWAGA: "brak danych" oznacza, że strona nie ma wystarczającego ruchu z Chrome do raportowania CrUX.
Dodatkowe metryki wydajności (lab data):
- FCP: {data.get('fcp_ms', 'brak')} ms
- TTFB: {data.get('ttfb_ms', 'brak')} ms
- TBT: {data.get('tbt_ms', 'brak')} ms
- Speed Index: {data.get('speed_index', 'brak')} ms
- Czas ładowania: {data.get('load_time_ms', 'brak')} ms
On-Page SEO:
- Meta title: {data.get('meta_title', 'brak')} (długość: {data.get('meta_title_length', '?')} znaków, optymalna: 50-60)
- Meta description: {'tak' if data.get('meta_description') else 'BRAK'} (długość: {data.get('meta_description_length', '?')} znaków, optymalna: 150-160)
- H1: {data.get('h1_count', 0)} (treść: {data.get('h1_text', 'brak')})
- H2: {data.get('h2_count', 0)}, H3: {data.get('h3_count', 0)}
- Obrazy: {data.get('total_images', 0)} (bez alt: {data.get('images_without_alt', 0)})
- Linki wewnętrzne: {data.get('internal_links_count', 0)}, zewnętrzne: {data.get('external_links_count', 0)}, uszkodzone: {data.get('broken_links_count', 0)}
Technical SEO:
- SSL: {'tak' if data.get('has_ssl') else 'NIE'}
- Sitemap: {'tak' if data.get('has_sitemap') else 'NIE'}
- Robots.txt: {'tak' if data.get('has_robots_txt') else 'NIE'}
- Canonical: {'tak' if data.get('has_canonical') else 'NIE'}
- Indeksowalna: {'tak' if data.get('is_indexable') else 'NIE'}
- Mobile-friendly: {'tak' if data.get('is_mobile_friendly') else 'NIE/brak danych'}
Security Headers:
- HSTS: {'tak' if data.get('has_hsts') else 'NIE' if data.get('has_hsts') is not None else 'brak danych'}
- CSP: {'tak' if data.get('has_csp') else 'NIE' if data.get('has_csp') is not None else 'brak danych'}
- X-Frame-Options: {'tak' if data.get('has_x_frame_options') else 'NIE' if data.get('has_x_frame_options') is not None else 'brak danych'}
- X-Content-Type-Options: {'tak' if data.get('has_x_content_type_options') else 'NIE' if data.get('has_x_content_type_options') is not None else 'brak danych'}
- Nagłówki bezpieczeństwa: {data.get('security_headers_count', '?')}/4
Dane strukturalne:
- Schema.org: {'tak' if data.get('has_structured_data') else 'NIE'} (typy: {data.get('structured_data_types', [])})
- LocalBusiness Schema: {'tak' if data.get('has_local_business_schema') else 'NIE'}
- Pola LocalBusiness Schema: {data.get('local_business_schema_fields', 'brak danych')}
- Język strony (html lang): {data.get('html_lang', 'brak')}
Search Console (dane z Google Search Console, ostatnie {data.get('gsc_period_days', 28)} dni):
- Kliknięcia: {data.get('gsc_clicks', 'brak danych (wymaga połączenia OAuth)')}
- Wyświetlenia: {data.get('gsc_impressions', 'brak danych')}
- CTR: {data.get('gsc_ctr', 'brak danych')}%
- Średnia pozycja: {data.get('gsc_avg_position', 'brak danych')}
- Top zapytania: {', '.join(q.get('query', '') for q in (data.get('gsc_top_queries') or [])[:5]) or 'brak danych'}
Social & Analytics:
- Open Graph: {'tak' if data.get('has_og_tags') else 'NIE'}
- Twitter Cards: {'tak' if data.get('has_twitter_cards') else 'NIE'}
- Google Analytics: {'tak' if data.get('has_google_analytics') else 'NIE'}
- GTM: {'tak' if data.get('has_google_tag_manager') else 'NIE'}
Local SEO (wynik: {data.get('local_seo_score', 'brak')}/100):
- Mapa Google: {'tak' if data.get('has_google_maps_embed') else 'NIE'}
- Lokalne słowa kluczowe: {'tak' if data.get('has_local_keywords') else 'NIE'}
- NAP na stronie: {'tak' if data.get('nap_on_website') else 'NIE'}
- Cytacje: {data.get('citations_found', 0)}/{data.get('citations_count', 0)} znalezionych
Treść:
- Świeżość: {data.get('content_freshness_score', 'brak')}/100
- Słów na stronie głównej: {data.get('word_count_homepage', 'brak')}
Formaty obrazów:
- Nowoczesne (WebP/AVIF/SVG): {data.get('modern_format_ratio', '?')}% ({data.get('webp_count', 0)} WebP)
- Legacy (JPEG/PNG): {data.get('legacy_image_count', '?')} obrazów
{benchmark_section}
ZADANIE:
Przygotuj analizę w formacie JSON z dwoma kluczami:
1. "summary" - krótki akapit (2-4 zdania) podsumowujący stan SEO strony, co jest dobrze, a co wymaga poprawy. Pisz bezpośrednio do właściciela firmy, po polsku.
2. "actions" - lista od 3 do 8 priorytetowanych akcji do podjęcia. Każda akcja to obiekt:
{{
"action_type": "typ akcji z listy: generate_schema_org, generate_meta_description, suggest_heading_fix, generate_alt_texts, seo_roadmap, add_analytics, add_sitemap, fix_ssl, add_og_tags, improve_performance, add_local_keywords, add_nap, fix_broken_links, improve_security_headers, optimize_images",
"title": "krótki tytuł po polsku",
"description": "opis co trzeba zrobić i dlaczego, 1-2 zdania",
"priority": "critical/high/medium/low",
"impact_score": 1-10,
"effort_score": 1-10,
"platform": "website"
}}
Priorytetyzuj wg: impact_score / effort_score (wyższy stosunek = wyższy priorytet).
NIE sugeruj akcji dla rzeczy, które firma już ma poprawnie.
Odpowiedz WYŁĄCZNIE poprawnym JSON-em, bez markdown, bez komentarzy."""
def _build_gbp_prompt(data: dict) -> str:
"""Build Gemini prompt for GBP audit analysis."""
# Benchmark comparison
gbp_benchmark_section = ""
benchmarks = data.get('_benchmarks')
if benchmarks and benchmarks.get('gbp'):
bm = benchmarks['gbp']
cat = benchmarks.get('category_name', '?')
gbp_benchmark_section = f"""
BENCHMARK (średnia w kategorii "{cat}", {benchmarks.get('category_company_count', '?')} firm):
- Średnia kompletność GBP: {bm.get('avg_completeness_score', '?')}/100
- Średnia ocena: {bm.get('avg_rating', '?')}/5
- Średnia liczba opinii: {bm.get('avg_review_count', '?')}
- Średnia liczba zdjęć: {bm.get('avg_photo_count', '?')}
Porównaj wyniki tej firmy ze średnią kategorii w analizie."""
# Build review keywords line (if available)
review_keywords_line = ""
if data.get('review_keywords'):
review_keywords_line = f"\n- Top słowa kluczowe z opinii: {', '.join(data.get('review_keywords', []))}"
# Build AI sentiment section
sentiment_section = ""
sentiment = data.get('review_sentiment')
if sentiment and isinstance(sentiment, dict):
pos = sentiment.get('positive', 0)
neu = sentiment.get('neutral', 0)
neg = sentiment.get('negative', 0)
total = pos + neu + neg
if total > 0:
sentiment_section = f"\n- Sentyment opinii: {pos} pozytywnych, {neu} neutralnych, {neg} negatywnych"
# Build description keywords section
description_keywords_section = "\nSłowa kluczowe w opisie:\n"
if data.get('description_keywords'):
description_keywords_section += f"- Znalezione: {', '.join(data.get('description_keywords', []))}"
else:
description_keywords_section += "- Brak danych"
# Build attributes section
attributes_section = ""
attrs = data.get('attributes')
if attrs and isinstance(attrs, dict):
parts = []
if attrs.get('payment'):
payment_items = [k.replace('_', ' ') for k, v in attrs['payment'].items() if v]
if payment_items:
parts.append(f" Płatności: {', '.join(payment_items)}")
if attrs.get('parking'):
parking_items = [k.replace('_', ' ') for k, v in attrs['parking'].items() if v]
if parking_items:
parts.append(f" Parking: {', '.join(parking_items)}")
if attrs.get('accessibility'):
acc_items = [k.replace('_', ' ') for k, v in attrs['accessibility'].items() if v]
if acc_items:
parts.append(f" Dostępność: {', '.join(acc_items)}")
if attrs.get('service'):
svc_items = [k.replace('_', ' ') for k, v in attrs['service'].items() if v]
if svc_items:
parts.append(f" Usługi: {', '.join(svc_items)}")
if attrs.get('amenities'):
amen_items = [k.replace('_', ' ') for k, v in attrs['amenities'].items() if v]
if amen_items:
parts.append(f" Udogodnienia: {', '.join(amen_items)}")
if attrs.get('food_and_drink'):
food_items = [k for k, v in attrs['food_and_drink'].items() if v]
if food_items:
parts.append(f" Jedzenie/napoje: {', '.join(food_items)}")
if parts:
attributes_section = "\n\nAtrybuty biznesu (z Google):\n" + "\n".join(parts)
# Build primary type and editorial summary
primary_type_line = ""
if data.get('primary_type'):
primary_type_line = f"\n- Typ główny (Google): {data.get('primary_type')}"
editorial_line = ""
if data.get('editorial_summary'):
editorial_line = f"\n- Opis Google: {data.get('editorial_summary')}"
price_level_line = ""
if data.get('price_level'):
price_level_line = f"\n- Poziom cenowy: {data.get('price_level')}"
return f"""Jesteś ekspertem Google Business Profile analizującym wizytówkę lokalnej firmy w Polsce.
DANE FIRMY:
- Nazwa: {data.get('company_name', 'N/A')}
- Branża: {data.get('company_category', 'N/A')}{primary_type_line}{editorial_line}{price_level_line}
- Miasto: {data.get('city', 'N/A')}
WYNIKI AUDYTU GBP (kompletność: {data.get('completeness_score', 'brak')}/100):
- Nazwa: {'' if data.get('has_name') else ''}
- Adres: {'' if data.get('has_address') else ''}
- Telefon: {'' if data.get('has_phone') else ''}
- Strona WWW: {'' if data.get('has_website') else ''}
- Godziny otwarcia: {'' if data.get('has_hours') else ''}
- Kategorie: {'' if data.get('has_categories') else ''}
- Zdjęcia: {'' if data.get('has_photos') else ''} ({data.get('photo_count', 0)} zdjęć)
- Opis: {'' if data.get('has_description') else ''}
- Usługi: {'' if data.get('has_services') else ''}
- Logo: {'✓ (wykryte heurystycznie)' if data.get('logo_present') else ''}
- Zdjęcie w tle: {'✓ (wykryte heurystycznie)' if data.get('cover_photo_present') else ''}
- Status zdjęć: {data.get('photo_status', 'brak danych')}
UWAGA: Logo i zdjęcie w tle są wykrywane heurystycznie na podstawie liczby zdjęć (1+ = logo, 2+ = cover). Nie traktuj tego jako pewnik - formułuj rekomendacje dot. zdjęć ostrożnie, np. "sprawdź czy logo jest ustawione" zamiast "BRAKUJE logo".
Opinie:
- Liczba opinii: {data.get('review_count', 0)}
- Średnia ocena: {data.get('average_rating', 'brak')}
- Z odpowiedzią: {data.get('reviews_with_response', 0)}
- Bez odpowiedzi: {data.get('reviews_without_response', 0)}
- Wskaźnik odpowiedzi: {data.get('review_response_rate', 'brak')}%{review_keywords_line}{sentiment_section}
Aktywność (UWAGA: te pola wymagają autoryzacji OAuth i są obecnie niedostępne):
- Posty: {('✓ (' + str(data.get('posts_count_30d', 0)) + ' w 30 dni)') if data.get('has_posts') else '[dane niedostępne bez autoryzacji OAuth]'}
- Produkty: {'' if data.get('has_products') else '[dane niedostępne bez autoryzacji OAuth]'}
- Pytania i odpowiedzi: {'' if data.get('has_qa') else '[dane niedostępne bez autoryzacji OAuth]'}
NAP:
- Spójność NAP: {'' if data.get('nap_consistent') else ''}
- Problemy NAP: {data.get('nap_issues', 'brak')}{attributes_section}
{description_keywords_section}
{gbp_benchmark_section}
ZADANIE:
Przygotuj analizę w formacie JSON z dwoma kluczami:
1. "summary" - krótki akapit (2-4 zdania) podsumowujący stan wizytówki Google, co jest dobrze, a co wymaga poprawy. Pisz bezpośrednio do właściciela firmy, po polsku.
2. "actions" - lista od 3 do 8 priorytetowanych akcji. Każda akcja:
{{
"action_type": "typ z listy: generate_gbp_description, generate_gbp_post, respond_to_review, suggest_categories, gbp_improvement_plan, add_photos, add_hours, add_services, add_products",
"title": "krótki tytuł po polsku",
"description": "opis co trzeba zrobić i dlaczego",
"priority": "critical/high/medium/low",
"impact_score": 1-10,
"effort_score": 1-10,
"platform": "google"
}}
NIE sugeruj akcji dla pól, które firma już ma poprawnie uzupełnione.
Odpowiedz WYŁĄCZNIE poprawnym JSON-em, bez markdown, bez komentarzy."""
def _build_social_prompt(data: dict) -> str:
"""Build Gemini prompt for social media audit analysis."""
# Benchmark comparison
social_benchmark_section = ""
benchmarks = data.get('_benchmarks')
if benchmarks and benchmarks.get('social'):
bm = benchmarks['social']
cat = benchmarks.get('category_name', '?')
social_benchmark_section = f"""
BENCHMARK (średnia w kategorii "{cat}", {benchmarks.get('category_company_count', '?')} firm):
- Średnia liczba platform: {bm.get('avg_platform_count', '?')}
- Średnia liczba obserwujących: {bm.get('avg_followers', '?')}
- Średnia kompletność profili: {bm.get('avg_completeness', '?')}%
Porównaj wyniki tej firmy ze średnią kategorii w analizie."""
profiles_info = ""
engagement_rates = []
page_names = []
for platform, info in data.get('profiles', {}).items():
profiles_info += f"\n {platform}: followers={info.get('followers_count', '?')}, "
profiles_info += f"bio={'' if info.get('has_bio') else ''}, "
profiles_info += f"photo={'' if info.get('has_profile_photo') else ''}, "
profiles_info += f"posty_30d={info.get('posts_count_30d', '?')}, "
profiles_info += f"kompletność={info.get('profile_completeness_score', '?')}%"
profiles_info += f", freq_score={info.get('posting_frequency_score', '?')}/10"
profiles_info += f", engagement={info.get('engagement_rate', '?')}%"
profiles_info += f", nazwa='{info.get('page_name', '?')}'"
if info.get('last_post_date'):
profiles_info += f", ost.post={info.get('last_post_date')}"
# YouTube metrics from API
if platform == 'youtube' and info.get('video_count') is not None:
profiles_info += f", filmy={info.get('video_count')}"
profiles_info += f", wyświetlenia={info.get('view_count', '?')}"
profiles_info += f", subskrybenci={info.get('subscriber_count', '?')}"
# Collect engagement rates for average calculation
if info.get('engagement_rate'):
engagement_rates.append(info.get('engagement_rate'))
# Collect page names for consistency check
if info.get('page_name'):
page_names.append(info.get('page_name'))
# Calculate average engagement
avg_engagement = round(sum(engagement_rates) / len(engagement_rates), 2) if engagement_rates else 0
# Check name consistency (simple check: all names should be similar)
consistent = len(set(page_names)) <= 1 if page_names else True
return f"""Jesteś ekspertem social media analizującym obecność lokalnej firmy w Polsce w mediach społecznościowych.
DANE FIRMY:
- Nazwa: {data.get('company_name', 'N/A')}
- Branża: {data.get('company_category', 'N/A')}
- Miasto: {data.get('city', 'N/A')}
OBECNOŚĆ W SOCIAL MEDIA (wynik: {data.get('score', 0)}/100):
- Platformy znalezione ({data.get('platforms_found', 0)}/{data.get('total_platforms', 6)}): {', '.join(data.get('platforms_present', []))}
- Platformy brakujące: {', '.join(data.get('platforms_missing', []))}
Szczegóły profili:{profiles_info or ' brak profili'}
DODATKOWE METRYKI:
- Średni engagement rate: {avg_engagement}% (szacunkowy, bez API)
- Spójność nazwy: {'TAK' if consistent else 'NIE — różne nazwy na platformach'}
{social_benchmark_section}
ZADANIE:
Przygotuj analizę w formacie JSON z dwoma kluczami:
1. "summary" - krótki akapit (2-4 zdania) podsumowujący obecność firmy w social media. Pisz po polsku, do właściciela firmy.
2. "actions" - lista od 3 do 8 priorytetowanych akcji. Każda akcja:
{{
"action_type": "typ z listy: generate_social_post, generate_bio, content_calendar, content_strategy, create_profile, improve_profile, increase_engagement",
"title": "krótki tytuł po polsku",
"description": "opis co trzeba zrobić i dlaczego",
"priority": "critical/high/medium/low",
"impact_score": 1-10,
"effort_score": 1-10,
"platform": "facebook/instagram/linkedin/youtube/twitter/tiktok/all"
}}
Dla firm lokalnych priorytetyzuj: Facebook > Instagram > LinkedIn > reszta.
NIE sugeruj tworzenia profili na platformach nieistotnych dla branży.
Odpowiedz WYŁĄCZNIE poprawnym JSON-em, bez markdown, bez komentarzy."""
# ============================================================
# CONTENT GENERATION PROMPTS
# ============================================================
CONTENT_PROMPTS = {
'generate_schema_org': """Wygeneruj kompletny JSON-LD Schema.org LocalBusiness dla firmy:
- Nazwa: {company_name}
- Branża: {category}
- Adres: {address}
- Miasto: {city}
- Telefon: {phone}
- Strona: {website}
- Email: {email}
Wygeneruj WYŁĄCZNIE poprawny tag <script type="application/ld+json"> z JSON-LD. Bez komentarzy.""",
'generate_meta_description': """Napisz meta description (150-160 znaków) dla strony firmy:
- Nazwa: {company_name}
- Branża: {category}
- Miasto: {city}
- Usługi: {services}
Opis powinien zawierać lokalne słowa kluczowe, CTA i zachęcać do kliknięcia.
Odpowiedz WYŁĄCZNIE tekstem meta description, bez cudzysłowów, bez komentarzy.""",
'suggest_heading_fix': """Zaproponuj poprawioną strukturę nagłówków (H1, H2, H3) dla strony firmy:
- Nazwa: {company_name}
- Branża: {category}
- Obecny H1: {h1_text}
- Liczba H1: {h1_count}, H2: {h2_count}, H3: {h3_count}
Zaproponuj strukturę z jednym H1 i logiczną hierarchią H2/H3. Po polsku.
Format odpowiedzi - lista nagłówków z poziomami, np.:
H1: ...
H2: ...
H3: ...
H2: ...""",
'generate_alt_texts': """Zaproponuj teksty alternatywne (alt) dla obrazów na stronie firmy:
- Nazwa: {company_name}
- Branża: {category}
- Liczba obrazów bez alt: {images_without_alt}
Napisz {images_without_alt} propozycji tekstów alt (max 125 znaków każdy).
Uwzględnij lokalne słowa kluczowe i opis kontekstowy. Po polsku.
Format: po jednym alt tekście na linię.""",
'generate_gbp_description': """Napisz opis firmy na Google Business Profile (max 750 znaków):
- Nazwa: {company_name}
- Branża: {category}
- Miasto: {city}
- Usługi: {services}
Opis powinien:
- Zawierać lokalne słowa kluczowe (miasto, region)
- Wymieniać główne usługi/produkty
- Zawierać CTA (zachętę do kontaktu)
- Być profesjonalny ale przyjazny
Odpowiedz WYŁĄCZNIE tekstem opisu, po polsku.""",
'generate_gbp_post': """Napisz post na Google Business Profile dla firmy:
- Nazwa: {company_name}
- Branża: {category}
- Miasto: {city}
Post powinien:
- Mieć 150-300 słów
- Zawierać CTA (np. "Zadzwoń", "Odwiedź stronę")
- Być angażujący i profesjonalny
- Dotyczyć aktualnej oferty/promocji/wydarzenia
Odpowiedz WYŁĄCZNIE tekstem postu, po polsku.""",
'respond_to_review': """Napisz profesjonalną odpowiedź na opinię Google:
- Firma: {company_name}
- Ocena: {review_rating}/5
- Treść opinii: {review_text}
Odpowiedź powinna:
- Podziękować za opinię
- Odnieść się do konkretnych punktów
- {'Przeprosić i zaproponować rozwiązanie' if int(str({review_rating}).replace('None','3')) <= 3 else 'Zachęcić do ponownych odwiedzin'}
- Być profesjonalna i osobista
- Max 200 słów
Odpowiedz WYŁĄCZNIE tekstem odpowiedzi, po polsku.""",
'generate_social_post': """Napisz post na {platform} dla firmy:
- Nazwa: {company_name}
- Branża: {category}
- Miasto: {city}
Wymagania dla {platform}:
- Facebook: 100-250 słów, może zawierać emoji, CTA
- Instagram: 100-150 słów, 5-10 hashtagów, emoji
- LinkedIn: 150-300 słów, profesjonalny ton, 3-5 hashtagów
- Twitter: max 280 znaków, 2-3 hashtagi
Odpowiedz WYŁĄCZNIE tekstem postu z hashtagami, po polsku.""",
'generate_bio': """Napisz bio/opis profilu na {platform} dla firmy:
- Nazwa: {company_name}
- Branża: {category}
- Miasto: {city}
- Strona: {website}
Limity znaków:
- Facebook: 255 znaków
- Instagram: 150 znaków
- LinkedIn: 2000 znaków (opis firmy)
- Twitter: 160 znaków
Odpowiedz WYŁĄCZNIE tekstem bio, po polsku.""",
'content_calendar': """Przygotuj kalendarz treści na tydzień (pon-pt) dla firmy w social media:
- Nazwa: {company_name}
- Branża: {category}
- Miasto: {city}
- Platformy: {platforms}
Dla każdego dnia podaj:
- Dzień tygodnia
- Platforma
- Temat postu
- Krótki zarys treści (1-2 zdania)
- Sugerowany typ: tekst/zdjęcie/wideo/karuzela
Format JSON array:
[{{"day": "Poniedziałek", "platform": "...", "topic": "...", "outline": "...", "content_type": "..."}}]
Odpowiedz WYŁĄCZNIE JSON-em, po polsku.""",
'content_strategy': """Przygotuj krótką strategię obecności w social media dla firmy:
- Nazwa: {company_name}
- Branża: {category}
- Miasto: {city}
- Obecne platformy: {platforms_present}
- Brakujące platformy: {platforms_missing}
Strategia powinna zawierać:
1. Rekomendowane platformy (z uzasadnieniem)
2. Częstotliwość publikacji per platforma
3. Typy treści per platforma
4. Ton komunikacji
5. Cele na 3 miesiące
Max 500 słów, po polsku. Pisz bezpośrednio do właściciela firmy.""",
}
# ============================================================
# JSON PARSE FAILURE TRACKING
# ============================================================
def _log_json_parse_failure(company_id: int, audit_type: str, details: str):
"""Log JSON parse failure for monitoring structured output reliability."""
try:
from database import SessionLocal, AIUsageLog
db = SessionLocal()
try:
log = AIUsageLog(
request_type='audit_json_parse_failure',
model='gemini-3.1-pro-preview',
tokens_input=0,
tokens_output=0,
cost_cents=0,
company_id=company_id,
related_entity_type=f'{audit_type}_audit',
success=False,
error_message=f'JSON parse failed: {details[:200]}',
)
db.add(log)
db.commit()
logger.warning(f"JSON parse failure logged for company {company_id}, audit_type={audit_type}")
finally:
db.close()
except Exception as e:
logger.error(f"Failed to log JSON parse failure: {e}")
# ============================================================
# MAIN SERVICE FUNCTIONS
# ============================================================
def generate_analysis(company_id: int, audit_type: str, user_id: int = None, force: bool = False) -> dict:
"""
Generate AI analysis for an audit.
Returns cached version if available and data hasn't changed.
Args:
company_id: Company ID
audit_type: 'seo', 'gbp', or 'social'
user_id: Current user ID for cost tracking
force: Force regeneration even if cache is valid
Returns:
dict with 'summary' and 'actions' keys
"""
db = SessionLocal()
try:
company = db.query(Company).filter_by(id=company_id, status='active').first()
if not company:
return {'error': 'Firma nie znaleziona'}
# Collect audit data
collectors = {
'seo': _collect_seo_data,
'gbp': _collect_gbp_data,
'social': _collect_social_data,
}
collector = collectors.get(audit_type)
if not collector:
return {'error': f'Nieznany typ audytu: {audit_type}'}
data = collector(db, company)
if not data:
return {'error': f'Brak danych audytu {audit_type} dla tej firmy'}
# Add benchmark data for AI context
try:
benchmarks = BenchmarkService(db).get_benchmarks(company.id)
if benchmarks:
data['_benchmarks'] = benchmarks
except Exception as e:
logger.warning(f"Benchmark loading failed: {e}")
# Exclude volatile fields from hash to improve cache hit rate
hash_data = {k: v for k, v in data.items() if k not in ('citations_count', 'citations_found', '_benchmarks')}
data_hash = _hash_data(hash_data)
# Check cache
if not force:
cache = db.query(AuditAICache).filter_by(
company_id=company_id,
audit_type=audit_type
).first()
if cache:
logger.info(f"Cache check: stored_hash={cache.audit_data_hash[:12]}... current_hash={data_hash[:12]}... match={cache.audit_data_hash == data_hash}")
if cache and cache.audit_data_hash == data_hash and cache.expires_at and cache.expires_at > datetime.now():
logger.info(f"AI analysis cache hit for company {company_id} audit_type={audit_type}")
result = {
'summary': cache.analysis_summary,
'actions': cache.actions_json or [],
'cached': True,
'generated_at': cache.generated_at.isoformat() if cache.generated_at else None,
}
if cache.previous_summary or cache.previous_actions_json:
result['previous'] = {
'summary': cache.previous_summary,
'actions': cache.previous_actions_json or [],
'generated_at': cache.previous_generated_at.isoformat() if cache.previous_generated_at else None,
}
return result
# Build prompt
prompt_builders = {
'seo': _build_seo_prompt,
'gbp': _build_gbp_prompt,
'social': _build_social_prompt,
}
prompt = prompt_builders[audit_type](data)
# Call Gemini with structured output schema
gemini = _get_gemini_service()
json_parse_failed = False
response_text = gemini.generate_text(
prompt=prompt,
temperature=0.3,
thinking_level='low',
feature='audit_analysis',
user_id=user_id,
company_id=company_id,
related_entity_type=f'{audit_type}_audit',
model='3-pro',
response_schema=AUDIT_ANALYSIS_SCHEMA,
)
if not response_text:
return {'error': 'Gemini nie zwrócił odpowiedzi'}
# Parse JSON response — structured output should be valid JSON directly
try:
result = json.loads(response_text)
except json.JSONDecodeError:
# Structured output failed — fallback to manual cleaning
json_parse_failed = True
logger.warning(f"Structured output JSON parse failed, attempting cleanup. Response: {response_text[:300]}")
try:
cleaned = response_text.strip()
if cleaned.startswith('```'):
cleaned = cleaned.split('\n', 1)[1] if '\n' in cleaned else cleaned[3:]
if cleaned.endswith('```'):
cleaned = cleaned[:-3]
result = json.loads(cleaned.strip())
except json.JSONDecodeError as e:
logger.error(f"Failed to parse Gemini response as JSON after cleanup: {e}\nResponse: {response_text[:500]}")
_log_json_parse_failure(company_id, audit_type, response_text[:500])
return {'error': 'Nie udało się przetworzyć odpowiedzi AI', 'raw_response': response_text}
if json_parse_failed:
_log_json_parse_failure(company_id, audit_type, 'Recovered after cleanup')
summary = result.get('summary', '')
actions = result.get('actions', [])
# Save to cache (upsert)
cache = db.query(AuditAICache).filter_by(
company_id=company_id,
audit_type=audit_type
).first()
if cache:
# Preserve previous analysis for comparison
cache.previous_summary = cache.analysis_summary
cache.previous_actions_json = cache.actions_json
cache.previous_generated_at = cache.generated_at
cache.analysis_summary = summary
cache.actions_json = actions
cache.audit_data_hash = data_hash
cache.generated_at = datetime.now()
cache.expires_at = datetime.now() + timedelta(days=CACHE_EXPIRY_DAYS)
else:
cache = AuditAICache(
company_id=company_id,
audit_type=audit_type,
analysis_summary=summary,
actions_json=actions,
audit_data_hash=data_hash,
generated_at=datetime.now(),
expires_at=datetime.now() + timedelta(days=CACHE_EXPIRY_DAYS),
)
db.add(cache)
# Save individual actions to audit_actions table
for action_data in actions:
action = AuditAction(
company_id=company_id,
audit_type=audit_type,
action_type=action_data.get('action_type', 'unknown'),
title=action_data.get('title', 'Akcja'),
description=action_data.get('description', ''),
priority=action_data.get('priority', 'medium'),
impact_score=action_data.get('impact_score'),
effort_score=action_data.get('effort_score'),
platform=action_data.get('platform', 'website'),
ai_model=gemini.model_name,
status='suggested',
created_by=user_id,
)
db.add(action)
db.commit()
result = {
'summary': summary,
'actions': actions,
'cached': False,
'generated_at': datetime.now().isoformat(),
}
if cache and (cache.previous_summary or cache.previous_actions_json):
result['previous'] = {
'summary': cache.previous_summary,
'actions': cache.previous_actions_json or [],
'generated_at': cache.previous_generated_at.isoformat() if cache.previous_generated_at else None,
}
return result
except Exception as e:
db.rollback()
logger.error(f"Error generating AI analysis: {e}", exc_info=True)
return {'error': f'Błąd generowania analizy: {str(e)}'}
finally:
db.close()
def generate_content(company_id: int, action_type: str, context: dict = None, user_id: int = None) -> dict:
"""
Generate specific content for an audit action.
Args:
company_id: Company ID
action_type: Content type (e.g. 'generate_schema_org', 'generate_gbp_post')
context: Additional context (e.g. platform, review_text)
user_id: Current user ID for cost tracking
Returns:
dict with 'content' key
"""
db = SessionLocal()
try:
company = db.query(Company).filter_by(id=company_id, status='active').first()
if not company:
return {'error': 'Firma nie znaleziona'}
# Build context for prompt
prompt_context = {
'company_name': company.name,
'category': (company.category.name if company.category else None) or 'Usługi',
'city': company.address_city or 'Polska',
'website': company.website or '',
'phone': company.phone or '',
'email': company.email or '',
'address': f"{company.address_street or ''}, {company.address_city or ''}".strip(', '),
'services': ', '.join([s.service.name for s in company.services[:5] if s.service]) if hasattr(company, 'services') and company.services else 'brak danych',
}
# Merge in extra context
if context:
prompt_context.update(context)
# Add audit-specific data
if action_type in ('suggest_heading_fix', 'generate_alt_texts'):
analysis = db.query(CompanyWebsiteAnalysis).filter(
CompanyWebsiteAnalysis.company_id == company.id
).order_by(CompanyWebsiteAnalysis.seo_audited_at.desc()).first()
if analysis:
prompt_context.update({
'h1_text': analysis.h1_text or 'brak',
'h1_count': analysis.h1_count or 0,
'h2_count': analysis.h2_count or 0,
'h3_count': analysis.h3_count or 0,
'images_without_alt': analysis.images_without_alt or 0,
})
if action_type in ('content_calendar', 'content_strategy'):
profiles = db.query(CompanySocialMedia).filter(
CompanySocialMedia.company_id == company.id
).all()
all_platforms = ['facebook', 'instagram', 'linkedin', 'youtube', 'twitter', 'tiktok']
present = [p.platform for p in profiles]
missing = [p for p in all_platforms if p not in present]
prompt_context.update({
'platforms': ', '.join(present) if present else 'brak',
'platforms_present': ', '.join(present) if present else 'brak',
'platforms_missing': ', '.join(missing) if missing else 'brak',
})
# Build prompt from template or generate dynamic fallback
if action_type in CONTENT_PROMPTS:
prompt_template = CONTENT_PROMPTS[action_type]
else:
# Dynamic fallback — look up action title/description from DB
existing_action = db.query(AuditAction).filter_by(
company_id=company_id,
action_type=action_type,
status='suggested'
).order_by(AuditAction.created_at.desc()).first()
action_title = existing_action.title if existing_action else action_type.replace('_', ' ')
action_desc = existing_action.description if existing_action else ''
prompt_template = f"""Jesteś ekspertem od marketingu cyfrowego i SEO dla lokalnych firm w Polsce.
Firma: {{company_name}}
Branża: {{category}}
Miasto: {{city}}
Strona: {{website}}
Usługi: {{services}}
ZADANIE: {action_title}
{('KONTEKST: ' + action_desc) if action_desc else ''}
Przygotuj konkretną, gotową do wdrożenia treść lub instrukcję krok po kroku dla tego zadania.
Pisz po polsku, bezpośrednio do właściciela firmy.
Bądź konkretny — podaj gotowe przykłady kodu, tekstów lub konfiguracji, które właściciel może skopiować i użyć.
Max 500 słów."""
try:
prompt = prompt_template.format(**prompt_context)
except KeyError as e:
prompt = prompt_template # Fall back to raw template if context is incomplete
logger.warning(f"Missing prompt context key: {e}")
# Call Gemini
gemini = _get_gemini_service()
content = gemini.generate_text(
prompt=prompt,
temperature=0.5,
thinking_level='low',
feature='audit_content_generation',
user_id=user_id,
company_id=company_id,
related_entity_type=f'audit_action_{action_type}',
model='3-pro',
)
if not content:
return {'error': 'Gemini nie zwrócił odpowiedzi'}
# Update action record if it exists
action = db.query(AuditAction).filter_by(
company_id=company_id,
action_type=action_type,
status='suggested'
).order_by(AuditAction.created_at.desc()).first()
if action:
action.ai_content = content
action.ai_model = gemini.model_name
db.commit()
return {
'content': content,
'action_type': action_type,
'model': gemini.model_name,
}
except Exception as e:
db.rollback()
logger.error(f"Error generating content: {e}", exc_info=True)
return {'error': f'Błąd generowania treści: {str(e)}'}
finally:
db.close()
def get_actions_for_company(company_id: int, audit_type: str = None) -> list:
"""
Get all audit actions for a company, optionally filtered by audit type.
Returns list of action dicts sorted by priority.
"""
db = SessionLocal()
try:
query = db.query(AuditAction).filter_by(company_id=company_id)
if audit_type:
query = query.filter_by(audit_type=audit_type)
actions = query.filter(
AuditAction.status.in_(['suggested', 'approved'])
).order_by(AuditAction.created_at.desc()).all()
# Deduplicate by action_type (keep newest)
seen = set()
unique_actions = []
for a in actions:
if a.action_type not in seen:
seen.add(a.action_type)
unique_actions.append(a)
# Sort by priority
priority_order = {'critical': 0, 'high': 1, 'medium': 2, 'low': 3}
unique_actions.sort(key=lambda a: priority_order.get(a.priority, 2))
return [{
'id': a.id,
'action_type': a.action_type,
'title': a.title,
'description': a.description,
'priority': a.priority,
'impact_score': a.impact_score,
'effort_score': a.effort_score,
'ai_content': a.ai_content,
'status': a.status,
'platform': a.platform,
'created_at': a.created_at.isoformat() if a.created_at else None,
} for a in unique_actions]
finally:
db.close()
def update_action_status(action_id: int, new_status: str) -> dict:
"""Update the status of an audit action."""
valid_statuses = ['suggested', 'approved', 'implemented', 'dismissed']
if new_status not in valid_statuses:
return {'error': f'Niepoprawny status. Dozwolone: {", ".join(valid_statuses)}'}
db = SessionLocal()
try:
action = db.query(AuditAction).filter_by(id=action_id).first()
if not action:
return {'error': 'Akcja nie znaleziona'}
action.status = new_status
if new_status == 'implemented':
action.implemented_at = datetime.now()
db.commit()
return {'success': True, 'id': action.id, 'status': new_status}
except Exception as e:
db.rollback()
logger.error(f"Error updating action status: {e}")
return {'error': str(e)}
finally:
db.close()