Implements comprehensive toxicity analysis following the Bluesky collector architecture: - Analyzer module with async batch processing using GPT-4o-mini - Database schema for toxicity scores and analysis run tracking - 12 toxicity categories (toxic, threat, hate_speech, racism, antisemitism, islamophobia, sexism, homophobia, insult, dehumanization, extremism, ableism) - Web interface routes for analysis dashboard and flagged content review - Manual review API endpoint for human validation - Analysis helper functions for database queries - Dutch language support with coded political term recognition Usage: docker exec mastodon-collector-collector-1 python -m app.analyzer See TOXICITY_ANALYSIS.md for full documentation. 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com>
38 lines
1.2 KiB
Python
38 lines
1.2 KiB
Python
"""Configuration for the toxicity analyzer."""
|
|
|
|
from __future__ import annotations
|
|
|
|
import os
|
|
from dataclasses import dataclass
|
|
|
|
|
|
@dataclass
|
|
class AnalyzerConfig:
|
|
"""Configuration for the toxicity analyzer."""
|
|
|
|
database_url: str
|
|
openai_api_key: str
|
|
model: str = "gpt-4o-mini"
|
|
batch_size: int = 10
|
|
concurrency: int = 5
|
|
flag_threshold: float = 0.5
|
|
limit: int = 0 # 0 = no limit
|
|
log_level: str = "INFO"
|
|
|
|
# Pricing (as of 2025, per million tokens)
|
|
input_cost_per_m: float = 0.150 # $0.150 per 1M input tokens
|
|
output_cost_per_m: float = 0.600 # $0.600 per 1M output tokens
|
|
|
|
@classmethod
|
|
def from_env(cls) -> AnalyzerConfig:
|
|
"""Load configuration from environment variables."""
|
|
return cls(
|
|
database_url=os.environ["DATABASE_URL"],
|
|
openai_api_key=os.environ["OPENAI_API_KEY"],
|
|
model=os.getenv("ANALYZER_MODEL", "gpt-4o-mini"),
|
|
batch_size=int(os.getenv("ANALYZER_BATCH_SIZE", "10")),
|
|
concurrency=int(os.getenv("ANALYZER_CONCURRENCY", "5")),
|
|
flag_threshold=float(os.getenv("ANALYZER_FLAG_THRESHOLD", "0.5")),
|
|
limit=int(os.getenv("ANALYZER_LIMIT", "0")),
|
|
log_level=os.getenv("ANALYZER_LOG_LEVEL", "INFO"),
|
|
)
|