37 cron jobs automatski prikupljaju podatke 24/7. Svaki izvor ima vlastiti collector sa schedule-om, retry logikom i dedup filterom.
⚙️ COLLECTORS
Python scripts39
Cron jobs37
Hydra CollectorACTIVE
Unified PipelineACTIVE
Brain BuilderACTIVE
Legal EmbedderACTIVE
Collectors koriste scraping (BeautifulSoup, Playwright), REST API pozive, RSS feed parsanje i WebSocket streaming. Svaki ima retry (3x), rate limiting i TruthGate validaciju.
Schemascivic (88), persona (35), support (18), projects (14), dabi (13), commander (8), platform (6), legal (3)
Svi sirovi podaci ulaze u PostgreSQL 18. Svaki schema ima jasnu domenu. Civic schema drži javne podatke (nabava, proračun, entiteti), persona schema drži DABI osobnost i blog.
Sentinel v11.00 orkestrira sve procesiranje. DataGate (4 workera) radi dedup, HealthChecker (60s interval) nadzire, AnomalyDetector (6h) traži nepravilnosti u nabavi i proračunu.
🧠 AI / LLM
Ollama (local)qwen2.5:7b, llama3.2, nomic-embed
Groqllama-3.3-70b (FREE)
DeepSeekdeepseek-chat ($0.14/1K)
GPURTX 4000 Ada 20GB VRAM
RAG PipelineQdrant + nomic-embed
WaterfallGroq → DeepSeek → Ollama → GPT → Claude
LLM waterfall: besplatni provideri prvi (Groq, Ollama), plaćeni kao fallback. DABI orchestrator automatski routea upite na pravi model prema tipu zadatka.
Svo procesirano znanje dostupno je kroz multiple kanale: RAG search (Qdrant), full-text search (Meili), graph queries (Neo4j), i DABI AI chat koji kombinira sve.