- Update all documentation to use "LLM API" instead of "OpenAI GPT-4o-mini" - Rename OPENAI_API_KEY to LLM_API_KEY in configuration - Update code comments to reflect generic LLM usage - Keep OpenAI-compatible client library (supports any LLM provider) - Add LOCAL_OPERATIONS.md and accounts.txt to .gitignore
38 lines
1.2 KiB
Python
38 lines
1.2 KiB
Python
"""Configuration for the toxicity analyzer."""
|
|
|
|
from __future__ import annotations
|
|
|
|
import os
|
|
from dataclasses import dataclass
|
|
|
|
|
|
@dataclass
|
|
class AnalyzerConfig:
|
|
"""Configuration for the toxicity analyzer."""
|
|
|
|
database_url: str
|
|
llm_api_key: str
|
|
model: str = "gpt-4o-mini"
|
|
batch_size: int = 10
|
|
concurrency: int = 5
|
|
flag_threshold: float = 0.5
|
|
limit: int = 0 # 0 = no limit
|
|
log_level: str = "INFO"
|
|
|
|
# Pricing (as of 2025, per million tokens)
|
|
input_cost_per_m: float = 0.150 # $0.150 per 1M input tokens
|
|
output_cost_per_m: float = 0.600 # $0.600 per 1M output tokens
|
|
|
|
@classmethod
|
|
def from_env(cls) -> AnalyzerConfig:
|
|
"""Load configuration from environment variables."""
|
|
return cls(
|
|
database_url=os.environ["DATABASE_URL"],
|
|
llm_api_key=os.environ["LLM_API_KEY"],
|
|
model=os.getenv("ANALYZER_MODEL", "gpt-4o-mini"),
|
|
batch_size=int(os.getenv("ANALYZER_BATCH_SIZE", "10")),
|
|
concurrency=int(os.getenv("ANALYZER_CONCURRENCY", "5")),
|
|
flag_threshold=float(os.getenv("ANALYZER_FLAG_THRESHOLD", "0.5")),
|
|
limit=int(os.getenv("ANALYZER_LIMIT", "0")),
|
|
log_level=os.getenv("ANALYZER_LOG_LEVEL", "INFO"),
|
|
)
|