Just point your LLM requests to Proxene for instant cost savings, PII protection, and compliance auditing—all debuggable with YAML/VS Code.
# Install
pip install proxene
# Start proxy
proxene start --port 8081
# Point your LLM calls to Proxene
# Old: https://api.openai.com/v1/chat/completions
# New: http://localhost:8081/v1/chat/completions- Policy-as-Code - Define cost limits, PII detection, and routing rules in YAML
- Cost Guard - Real-time token counting with per-request/minute/daily caps
- PII Shield - Auto-detect and redact sensitive data (SSN, emails, credit cards)
- Local-First - Debug policies with CLI replay, no cloud dependency
- VS Code Integration - Live request monitoring and policy debugging
# policies/default.yaml
cost_limits:
max_per_request: 0.03
daily_cap: 100.00
pii_detection:
enabled: true
action: redact # or: block, warn
routing:
- if: complexity < 3
use: claude-3-haiku
- else:
use: gpt-4oYour App → Proxene Proxy → LLM Provider
↓
[Policy Engine]
[Cost Guard]
[PII Detector]
[OTEL Logging]
Built-in OpenTelemetry support for monitoring:
- Request costs and token usage
- PII detection hits
- Policy violations
- Model routing decisions
We welcome contributions! Please see our Contributing Guide.
MIT License - see LICENSE file
Built with ❤️ for developers who need simple, powerful AI governance