Files
echo-core/memory/kb/tools/infrastructure.md
Marius Mutu 1fbd624195 chore(kb): add memory/kb to git tracking
memory/* was fully ignored; now only memory/kb/ is tracked
so notes, coaching sessions, insights, and project docs are
versioned while embeddings and sqlite databases stay untracked.

Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
2026-04-25 21:35:41 +00:00

12 KiB

Infrastructură (Proxmox + Docker)

Ultima actualizare: 2026-04-24. Scan complet al tuturor nodurilor.

Acces rapid LXC

ID Nume Nod IP SSH direct Via Proxmox
100 portainer pvemini 10.0.20.170 ssh echo@10.0.20.170 ssh echo@10.0.20.201 "sudo pct exec 100 -- bash"
101 minecraft pveelite 10.0.20.162 ssh echo@10.0.20.162 ssh echo@10.0.20.202 "sudo pct exec 101 -- bash"
103 dokploy pvemini 10.0.20.167 ssh echo@10.0.20.167 ssh echo@10.0.20.201 "sudo pct exec 103 -- bash"
104 flowise pvemini 10.0.20.161 (publickey only) ssh echo@10.0.20.201 "sudo pct exec 104 -- bash"
106 gitea pvemini 10.0.20.165 ssh echo@10.0.20.201 "sudo pct exec 106 -- sh" ⚠️ Alpine (sh, nu bash)
108 central-oracle pvemini 10.0.20.121 ssh echo@10.0.20.121 ssh echo@10.0.20.201 "sudo pct exec 108 -- bash"
110 moltbot pveelite 10.0.20.173 ssh echo@10.0.20.173 ssh echo@10.0.20.202 "sudo pct exec 110 -- bash"
171 claude-agent pvemini 10.0.20.171 ssh user@10.0.20.171 ssh echo@10.0.20.201 "sudo pct exec 171 -- bash"

LXC 100 — portainer (pvemini)

  • IP: 10.0.20.170 | OS: Debian/systemd | Tailscale: Da
  • Resurse: 4GB RAM (414MB used) | 20GB disk (4GB used, 20%)
  • Portainer UI: https://10.0.20.170:9443
  • Proiecte docker-compose: /opt/docker/

Containere Docker:

Container Port extern Status Descriere
portainer 9443 healthy Management Docker
hbbs 21115-21116, 21118 RustDesk relay (STUN)
hbbr 21117, 21119 RustDesk relay (TURN)
pulse 7655 healthy Monitoring Proxmox
wol-manager Wake-on-LAN
bt-web-automation 5000, 8081→8080 BT automation
roa-efactura 5003→5000 ⚠️ unhealthy E-Factura ANAF
pdf-qr-app 5002→5000 healthy QR facturi
docker-flask_app-1 5001→5000 ROA Flask

Depanare:

# Logs container
ssh echo@10.0.20.201 "sudo pct exec 100 -- docker logs <container> --tail 50"
# Restart container
ssh echo@10.0.20.201 "sudo pct exec 100 -- docker restart <container>"
# Status toate
ssh echo@10.0.20.201 "sudo pct exec 100 -- docker ps -a"

LXC 101 — minecraft (pveelite)

  • IP: 10.0.20.162 | OS: Debian/systemd | Tailscale: Nu
  • Resurse: 8GB RAM (3.8GB used) | 100GB disk (49GB used, 49%)

Servicii:

Serviciu Port Descriere
crafty 8443 Crafty4 web panel (Python)
minecraft 25565 Server Minecraft (Java)
playit Tunel public pentru Minecraft

Depanare:

ssh echo@10.0.20.202 "sudo pct exec 101 -- systemctl status crafty"
ssh echo@10.0.20.202 "sudo pct exec 101 -- journalctl -u crafty -n 50"

LXC 103 — dokploy (pvemini)

  • IP: 10.0.20.167 | OS: Debian/systemd | Tailscale: Da
  • Resurse: 4GB RAM (1.1GB used) | 50GB disk (5.8GB used, 12%)
  • Dokploy UI: http://10.0.20.167:3000

Containere Docker (managed by Dokploy + Traefik):

Container Port Status Descriere
dokploy-traefik 80, 443 Reverse proxy
dokploy 3000 healthy Deployment platform
dokploy-postgres 5432 (intern) DB Dokploy
dokploy-redis 6379 (intern) Cache Dokploy
utile-icongenerator 80 (intern) Icon generator
qr-qrgenerator 80 (intern) QR generator
qr-pdfqrapp PDF+QR app
constanta-space-booking-backend 8000 (intern) Space booking API

Depanare:

ssh echo@10.0.20.201 "sudo pct exec 103 -- docker ps -a"
ssh echo@10.0.20.201 "sudo pct exec 103 -- docker logs <container> --tail 50"

LXC 104 — flowise (pvemini)

  • IP: 10.0.20.161 | OS: Debian/systemd | Tailscale: Da
  • Resurse: 8GB RAM (418MB used) | 100GB disk (23GB used, 23%)
  • SSH direct: (nu merge cu user echo, doar via pct exec)

Servicii:

Serviciu Port Status Descriere
ollama 127.0.0.1:11434 LLM local (CPU-only, avx2)
flowise 3000 Flow builder AI
ngrok Tunel public

Ollama — modele disponibile:

  • all-minilm:latest — embeddings rapid ← folosit de echo-core memory_search
  • nomic-embed-text:latest — embeddings calitate
  • llama3.2:3b-instruct-q8_0 — LLM conversație
  • llama3.2:3b, llama3.2:1b — LLM general
  • smollm:135m — LLM mic rapid

Note importante:

  • Modele stocate în /usr/share/ollama/.ollama/models/ (user ollama)
  • Serviciul ollama rulează ca user ollama, cu HOME=/usr/share/ollama
  • CPU-only — fără GPU; fără CUDA/ROCm

Depanare Ollama:

# Status
ssh echo@10.0.20.201 "sudo pct exec 104 -- systemctl status ollama"
# Logs (problemă frecventă: $HOME undefined sau permisiuni)
ssh echo@10.0.20.201 "sudo pct exec 104 -- journalctl -u ollama -n 30"
# Fix permisiuni (dacă ollama nu pornește)
ssh echo@10.0.20.201 "sudo pct exec 104 -- chown -R ollama:ollama /usr/share/ollama/.ollama/"
# Test API
ssh echo@10.0.20.201 "sudo pct exec 104 -- curl -s http://localhost:11434/api/tags"
# Pull model
ssh echo@10.0.20.201 "sudo pct exec 104 -- ollama pull all-minilm"

LXC 106 — gitea (pvemini)

  • IP: 10.0.20.165 | OS: Alpine Linux + OpenRC ⚠️ (nu systemd, nu bash!)
  • Resurse: disk 250GB (1.1GB used, 0%)
  • Gitea web: http://10.0.20.165:3000 (sau gitea.romfast.ro)
  • Gitea SSH: port 222

Particularități Alpine:

  • Shell: sh (nu bash) — pct exec 106 -- sh
  • Init: OpenRC (nu systemd) — rc-status, nu systemctl
  • Gitea rulează prin Docker + s6, nu nativ

Servicii OpenRC:

Serviciu Status Descriere
networking Rețea
tailscale VPN
crond Cron
tailscale-gitea CRASHED Script Tailscale custom — de investigat

Depanare:

# Acces (folosește sh, nu bash!)
ssh echo@10.0.20.201 "sudo pct exec 106 -- sh -c 'rc-status'"
ssh echo@10.0.20.201 "sudo pct exec 106 -- sh -c 'docker ps'"
# Logs tailscale-gitea
ssh echo@10.0.20.201 "sudo pct exec 106 -- sh -c 'cat /var/log/tailscale-gitea.log 2>/dev/null || rc-service tailscale-gitea status'"

LXC 108 — central-oracle (pvemini)

  • IP: 10.0.20.121 | OS: Debian/systemd | Tailscale: Nu
  • Resurse: 8GB RAM (4.2GB used) | 50GB disk (15GB used, 29%)

Containere Docker:

Container Port Status Descriere
oracle-xe 1521, 5500 (EM Express) healthy Oracle XE principal
oracle18-xe 1522→1521, 5502→5500 Oracle 18 XE
portainer 9000, 9443, 8000 Management local

Depanare Oracle:

# Status
ssh echo@10.0.20.201 "sudo pct exec 108 -- docker ps -a"
# Logs Oracle
ssh echo@10.0.20.201 "sudo pct exec 108 -- docker logs oracle-xe --tail 50"
# Intră în container Oracle
ssh echo@10.0.20.201 "sudo pct exec 108 -- docker exec -it oracle-xe bash"

LXC 110 — moltbot (pveelite)

  • IP: 10.0.20.173 | OS: Debian/systemd | Tailscale: Da
  • Resurse: 4GB RAM (564MB used) | 30GB disk (15GB used, 48%)
  • Acesta este LXC-ul pe care rulează echo-core

Servicii:

Serviciu Port Descriere
code-server@moltbot 8080 VS Code în browser
ttyd 7681 Web terminal
echo-core dashboard 8088 Echo Task Board
whatsapp-bridge 8098 Baileys bridge (Node.js)
fail2ban Protecție SSH

LXC 171 — claude-agent (pvemini)

  • IP: 10.0.20.171 | Tailscale: 100.95.55.51 | OS: Ubuntu/systemd
  • Resurse: 16GB RAM (982MB used) | 32GB disk (23GB used, 72% — de monitorizat)
  • User principal: claude | Workspace: /workspace/

Servicii:

Serviciu Port Descriere
code-server@claude 8080 VS Code (user: claude)
ttyd 7681 Web terminal (/workspace/start-agent.sh, auth: claude:claude2025)

Claude Code:

  • Instalat și configurat, Git → gitea.romfast.ro
  • Mod programatic: claude -p "task" din directorul proiectului

Proiecte în /workspace/ → detalii complete în kb/tools/claude-agent-projects.md

Proiect Stack Scop
roa2web FastAPI + Vue.js + Oracle ERP web modern ROA
roaauto Vue 3 + wa-sqlite + FastAPI PWA service auto (offline-first)
vfp_roaauto Visual FoxPro (legacy) ROA AUTO versiunea VFP
romfastsql Docs + SQL + Python Infrastructură + Oracle migrare
gomag-vending FastAPI + Oracle PL/SQL Import comenzi GoMag → ROA
space-booking FastAPI + SQLite + Vue Rezervări birouri multi-tenant
service-auto Vue 3 + Vite + Tailwind 4 PWA service auto (versiune nouă)
atm Python 3.11+ Automated Trading Monitor (M2D)
paula-escape HTML Escape room joc

Depanare:

ssh echo@10.0.20.201 "sudo pct exec 171 -- systemctl status code-server@claude ttyd"
ssh echo@10.0.20.201 "sudo pct exec 171 -- df -h /"

VM 201 — roacentral (pvemini)

  • IP: 10.0.20.122 | OS: Windows | QEMU Guest Agent: Da
  • Windows VM cu guest agent activ

Mașini Windows externe (producție)

Mașină IP Port Rol
Oracle producție 10.0.20.36 1521 Oracle 10g Windows, baza de date principală ROA
Oracle DR 10.0.20.37 1521 Disaster recovery Oracle

Proxmox Noduri

User: echo | Acces SSH: ssh echo@<IP> | Sudo: qm, pct, pvesh

pvemini (10.0.20.201) — host principal

  • Resurse: 64GB RAM, 1.4TB disk
  • LXC-uri: 100, 103, 104, 105(stopped), 106, 108, 171
  • VM-uri: 201(running), 300(stopped), 302(stopped)
  • Backup zilnic 02:00: VM 100, 104, 106, 108, 171, 201 → storage "backup"

Scripturi /opt/scripts/:

  • ha-monitor.sh — zilnic 00:00, status cluster HA
  • monitor-ssl-certificates.sh — verifică SSL-uri
  • ups-shutdown-cluster.sh — shutdown orchestrat la UPS critic
  • ups-monthly-test.sh — 1 ale lunii, test baterie UPS
  • ups-maintenance-shutdown.sh — shutdown mentenanță UPS
  • vm107-monitor.sh — monitorizare VM 107

pveelite (10.0.20.202)

  • Resurse: 16GB RAM, 557GB disk
  • LXC-uri: 101(running), 105(stopped), 110(running), 301(stopped)
  • VM-uri: 109(stopped — oracle DR)
  • Backup zilnic 22:00: LXC 101, 110 → backup-pvemini-nfs

Scripturi /opt/scripts/:

  • oracle-backup-monitor-proxmox.sh — zilnic 21:00, verifică backup Oracle
  • weekly-dr-test-proxmox.sh — sâmbătă 06:00, test restore Oracle DR (VM 109)

pve1 (10.0.20.200)

  • Resurse: 32GB RAM, 1.3TB disk
  • Status: Gol (fără VM/LXC activ)

Servicii LLM/AI locale

Serviciu LXC IP:Port Note
Ollama 104 flowise 10.0.20.161:11434 CPU-only; modele: all-minilm, nomic-embed-text, llama3.2
Flowise 104 flowise 10.0.20.161:3000 Flow builder AI

Alertă automată când

  • Container/VM down neașteptat
  • Disk >85% utilizare (LXC 171 deja la 72% — monitorizez)
  • Serviciu unhealthy >1h
  • Erori repetate în logs

Acțiunez singur (fără să întreb)

  • Monitorizare și citire status
  • Diagnozare: logs, configurații, health checks
  • Fix-uri safe: permisiuni, restart servicii

Întreb întâi

  • Start/Stop VM sau LXC
  • Modificări configurare (network, storage, resurse)
  • Orice operație distructivă