All resilience patterns. One decorator. Zero dependencies.
Inspired by Java's Resilience4j. Stop juggling tenacity for retries, pybreaker for circuit breakers, and custom code for everything else. pyresilience gives you retry, circuit breaker, timeout, fallback, bulkhead, rate limiter, and cache — all through a single @resilient() decorator that works with both sync and async functions.
pip install pyresilienceAlso works with uv, poetry, and pdm.
import requests
from pyresilience import resilient, RetryConfig, TimeoutConfig, CircuitBreakerConfig
@resilient(
retry=RetryConfig(max_attempts=3, delay=1.0),
timeout=TimeoutConfig(seconds=10),
circuit_breaker=CircuitBreakerConfig(failure_threshold=5),
)
def call_api(endpoint: str) -> dict:
return requests.get(endpoint).json()Retries with exponential backoff. Times out at 10s. Opens the circuit after 5 failures. That's it.
- One library instead of many — No need to wire together
tenacity+pybreaker+ custom timeout/fallback/rate limiting code. One config, one decorator. - Patterns that work together — Circuit breaker state is shared across retries. Rate limiting respects bulkhead limits. Cache short-circuits the entire pipeline. Everything is coordinated.
- Zero dependencies — Pure Python stdlib. Nothing to conflict with your stack.
- Sync and async — Same API for both. Auto-detects your function type.
- Production observability — Built-in event listeners for logging, metrics, and alerting. OpenTelemetry and Prometheus listeners included. Know when circuits open, retries fire, or rate limits hit.
- Thread-safe and async-safe — All stateful components use locks. Async-safe latency tracking via
contextvars. Cache stampede prevention via per-key locking. - Framework integrations — Drop-in support for FastAPI, Django, and Flask.
from pyresilience import resilient, RetryConfig, TimeoutConfig, CircuitBreakerConfig
from pyresilience import FallbackConfig, BulkheadConfig, RateLimiterConfig, CacheConfig
@resilient(
retry=RetryConfig(max_attempts=3, delay=1.0, backoff_factor=2.0),
timeout=TimeoutConfig(seconds=10),
circuit_breaker=CircuitBreakerConfig(failure_threshold=5, recovery_timeout=30),
fallback=FallbackConfig(handler=lambda e: {"status": "degraded"}, fallback_on=[Exception]),
bulkhead=BulkheadConfig(max_concurrent=10),
rate_limiter=RateLimiterConfig(max_calls=100, period=60.0),
cache=CacheConfig(ttl=300.0, max_size=1000),
)
def call_service(endpoint: str) -> dict:
return requests.get(endpoint).json()| Pattern | Config | What it does |
|---|---|---|
| Retry | RetryConfig |
Exponential backoff with jitter |
| Timeout | TimeoutConfig |
Per-call time limits |
| Circuit Breaker | CircuitBreakerConfig |
Stop calling failing services |
| Fallback | FallbackConfig |
Graceful degradation |
| Bulkhead | BulkheadConfig |
Concurrency limiting |
| Rate Limiter | RateLimiterConfig |
Token bucket rate limiting |
| Cache | CacheConfig |
LRU result caching with TTL |
The same decorator works with async functions — no changes needed:
import aiohttp
from pyresilience import resilient, RetryConfig, CircuitBreakerConfig
@resilient(
retry=RetryConfig(max_attempts=3, delay=0.5),
circuit_breaker=CircuitBreakerConfig(failure_threshold=5),
)
async def call_api(url: str) -> dict:
async with aiohttp.ClientSession() as session:
async with session.get(url) as resp:
return await resp.json()Skip the configuration for common use cases:
from pyresilience import resilient
from pyresilience import http_policy, db_policy, queue_policy, strict_policy
@resilient(**http_policy()) # 10s timeout, 3 retries, circuit breaker
def call_api(): ...
@resilient(**db_policy()) # 30s timeout, 2 retries, 10 concurrent max
def query_db(): ...
@resilient(**queue_policy()) # 15s timeout, 5 retries, high failure threshold
async def publish_message(): ...
@resilient(**strict_policy()) # 5s timeout, 1 retry, fail fast
def latency_critical(): ...from pyresilience import resilient, RetryConfig, JsonEventLogger, MetricsCollector
logger = JsonEventLogger()
metrics = MetricsCollector()
@resilient(retry=RetryConfig(max_attempts=3), listeners=[logger, metrics])
def my_func():
...
# After calls:
print(metrics.summary())
# {"my_func": {"events": {"retry": 2, "success": 1}, "success_rate": 1.0, "avg_latency_ms": 15.2}}from pyresilience import resilience_context
# Set trace/request ID for the current context — propagates through all resilience events
resilience_context.set({"trace_id": "abc-123", "request_id": "req-456"})from pyresilience.contrib.otel import OpenTelemetryListener
from pyresilience.contrib.prometheus import PrometheusListener
@resilient(retry=RetryConfig(max_attempts=3), listeners=[OpenTelemetryListener()])
def call_api(): ...
@resilient(retry=RetryConfig(max_attempts=3), listeners=[PrometheusListener()])
def call_db(): ...Prevent retry storms across your service with a shared token bucket:
from pyresilience import resilient, RetryConfig, RetryBudgetConfig, RetryBudget
budget = RetryBudget(RetryBudgetConfig(max_retries=100, refill_rate=10))
@resilient(retry=RetryConfig(max_attempts=3, retry_budget=budget))
def call_api(): ...Apply timeout per attempt instead of a total deadline:
from pyresilience import resilient, TimeoutConfig
@resilient(timeout=TimeoutConfig(seconds=5, per_attempt=True)) # 5s per attempt
def call_api(): ...
@resilient(timeout=TimeoutConfig(seconds=30, per_attempt=False)) # 30s total deadline
def call_db(): ...Inspect circuit breaker states across your registry:
from pyresilience import ResilienceRegistry, health_check
registry = ResilienceRegistry()
# ... register and use services ...
status = health_check(registry)
# {"payment-api": "CLOSED", "inventory-api": "OPEN"}Drain in-flight calls before stopping:
from pyresilience import shutdown
shutdown(wait=True, timeout=30) # Wait up to 30s for in-flight calls to completeBenchmarked against tenacity, backoff, stamina, and pybreaker on macOS (Apple Silicon). Full benchmark code in benchmarks/.
| Library | Mean | vs pyresilience |
|---|---|---|
| bare (no decorator) | 0.07μs | — |
| pyresilience | 0.64μs | 1.0x |
| pybreaker | 0.64μs | 1.0x |
| backoff | 1.29μs | 2.0x slower |
| stamina | 5.33μs | 8.3x slower |
| tenacity | 6.64μs | 10.4x slower |
pyresilience is 10.4x faster than tenacity on the happy path.
| Pattern | Mean Latency |
|---|---|
| Retry (happy path) | 0.64μs |
| Circuit Breaker | 1.03μs |
| Fallback (triggered) | 0.69μs |
| Bulkhead | 0.74μs |
| Rate Limiter | 0.89μs |
| Cache (hit) | 0.68μs |
| All 7 patterns (cache hit) | 0.67μs |
| Library | ops/sec |
|---|---|
| pyresilience | 223,934 |
| tenacity | 58,109 |
pyresilience achieves 3.9x higher throughput under concurrent load.
| Library | Mean |
|---|---|
| pyresilience | 0.82μs |
| tenacity | 11.83μs |
pyresilience is 14.4x faster than tenacity for async functions.
| Library | Memory |
|---|---|
| pyresilience | 1,224 KB |
| tenacity | 2,150 KB |
pyresilience uses 43% less memory.
| pyresilience | tenacity | pybreaker | backoff | stamina | |
|---|---|---|---|---|---|
| Retry | Yes | Yes | - | Yes | Yes |
| Circuit Breaker | Yes | - | Yes | - | - |
| Timeout | Yes | - | - | - | - |
| Fallback | Yes | - | - | - | - |
| Bulkhead | Yes | - | - | - | - |
| Rate Limiter | Yes | - | - | - | - |
| Cache | Yes | - | - | - | - |
| Retry Budget | Yes | - | - | - | - |
| Context Propagation | Yes | - | - | - | - |
| Health Check | Yes | - | - | - | - |
| Prometheus | Yes | - | - | - | - |
| OpenTelemetry | Yes | - | - | - | - |
| Unified API | Yes | - | - | - | - |
| Zero Dependencies | Yes | Yes | - | - | - |
| Async | Yes | Yes | - | Yes | Yes |
Comparison reflects built-in capabilities and unified API model, not every possible custom composition.
Full guides, API reference, and examples at pyresilience.readthedocs.io.
MIT