Skip to content

Production readiness review — 7 gaps between architecture and deployment #1

@aeoess

Description

@aeoess

Reviewed the full codebase. The architecture is strong — 815+ tests, 6 formal invariants, DFA-based lifecycle, multi-bridge design. This is a real protocol, not a README. Here are the gaps between where it is now and production-ready, in priority order:

1. Not distributed. Packages aren't on npm. You can't npm install @sint/core. Everything is monorepo-local. This is the single biggest gap — without published packages, nobody can integrate. The pnpm workspace structure is ready for publishing, it just hasn't been done. npm publish --access public on @sint/core, @sint/gate-capability-tokens, @sint/gate-policy-gateway, and @sint/bridge-mcp would immediately make the protocol usable.

2. No hosted gateway. No equivalent of a public enforcement endpoint. It's local Docker or Railway deploy-it-yourself. A hosted instance (even a free tier) lets the ecosystem verify tokens and query trust profiles without running their own infrastructure. This is what turns a library into a platform.

3. Phase 5 (Avatar Layer) not started. Trust interface, dynamic consent, CSML-driven tier escalation are planned but not built. The existing 4 phases are solid — Phase 5 is the user-facing trust layer that would make the protocol accessible to non-experts. Not blocking for developer adoption, but blocking for enterprise/operator adoption.

4. Real hardware testing unclear. The ROSClaw paper is genuine research, but it's not clear whether bridge-ros2 has been tested against actual ROS2 nodes with physical actuators. The bridge code exists, the tests exist, but "tested in simulation" vs "tested on hardware" is a material distinction for safety-critical systems. Even a single documented hardware test (a Turtlebot, a simulated Gazebo environment with physics) would strengthen the credibility significantly.

5. Zero ecosystem adoption. 1 star, 0 forks. The GitHub engagement happening right now (A2A, MCP threads) is the first ecosystem visibility. The protocol deserves more attention than it's getting — the formal specification is stronger than most protocols at 10x the star count. Consider: a blog post or demo video, submissions to robotics/AI safety conferences, a "getting started" tutorial that goes from zero to enforced robot command in under 5 minutes.

6. No Python SDK. TypeScript only. The ROS2 ecosystem is heavily Python (rclpy). A Python SDK — even if it only covers capability-tokens and policy-gateway — would unlock the primary audience for physical AI governance. The interop verification we did today proves the crypto layer (Ed25519, did:key) produces identical output in both languages — the port is straightforward.

7. Engine packages (system1/system2/hal) are speculative. Perception pipeline, action predictor, behavior trees, ONNX executor — these are ambitious but appear untested against real workloads. They risk making the repo look like vaporware to someone scanning quickly. Consider: either test them against a concrete scenario (even synthetic) or move them to a separate sint-engines repo so the core protocol stays focused and credible.

None of these are architectural problems. The design is sound. These are distribution and deployment gaps — the difference between "works on my machine" and "works in production." The formal invariants, the DFA model, the physical constraint enforcement — those are the hard parts, and they're done.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions