This is a really well-developed concept that emerged organically from the conversation. Here's what I see as the core architecture:
The system: A fleet of specialized AI bots, each with their own Forgejo account and working directory, maintaining forks of critical neglected open source packages. They coordinate entirely through the forge (issues, PRs, labels, comments) the same way human developers would.
The bots:
- Forkbot - administrator, decides what to fork, manages lifecycle, tracks goals
- Security bot - vulnerability scanning, CVE tracking, signing
- Dependencies bot - dependency graph analysis, upgrade archaeology
- Testing bot - coverage, downstream compatibility, application canaries
- Documentation bot - docs accuracy, migration guides, API docs
- Performance bot - profiling, anti-pattern detection
Key design decisions already made:
- One-way mirror: read upstream, never touch it
- Labels as the coordination mechanism (no boss/orchestrator)
- Each bot's loop: check labelled issues, work on them, audit repos when idle, open new issues
- Async standups in a dedicated repo
- Forkbot gates new repos, tags everyone for first pass
- Application repos brought in as integration test canaries (only change is swapping to forked packages)
- Publishing to Forgejo's package registry
- Admin escalation via an admin-requests repo
- Graduation criteria for when a package is "done"
What's strong: The "forge as orchestrator" idea eliminates custom infrastructure. The one-way wall is politically smart. The audit-first approach generates useful data even before fixes land. Application canaries as the real compatibility test is the thing that would actually build trust.
What's underspecified: How each bot actually runs (the Claude Code loop mechanics, context management, cost control), the persona prompts themselves, what "critical and neglected" means concretely (thresholds from ecosyste.ms), the Forgejo instance setup, and the order you'd build this in.
What I'd push back on slightly: The conversation got enthusiastic about scale (500 packages, cross-ecosystem) before nailing the single-package prototype. The concrete plan should probably start with one package and two or three bots, prove the interaction patterns work, then expand.
Want me to turn this into an actual buildable plan? I'd structure it around phases - starting with infrastructure setup and a minimal prototype, then expanding bot roles and package count.