Async Jobs Patterns for background task processing with Celery, ARQ, and Redis. Covers task queues, canvas workflows, scheduling, retry strategies, rate limiting, and production monitoring. Each category has individual rule files in references/ loaded on-demand. Quick Reference Category Rules Impact When to Use Configuration celery-config HIGH Celery app setup, broker, serialization, worker tuning Task Routing task-routing HIGH Priority queues, multi-queue workers, dynamic routing Canvas Workflows canvas-workflows HIGH Chain, group, chord, nested workflows Retry Strategies retry-strategies HIGH Exponential backoff, idempotency, dead letter queues Scheduling scheduled-tasks MEDIUM Celery Beat, crontab, database-backed schedules Monitoring monitoring-health MEDIUM Flower, custom events, health checks, metrics Result Backends result-backends MEDIUM Redis results, custom states, progress tracking ARQ Patterns arq-patterns MEDIUM Async Redis Queue for FastAPI, lightweight jobs Temporal Workflows temporal-workflows HIGH Durable workflow definitions, sagas, signals, queries Temporal Activities temporal-activities HIGH Activity patterns, workers, heartbeats, testing Total: 10 rules across 9 categories Quick Start
Celery task with retry
from celery import shared_task @shared_task ( bind = True , max_retries = 3 , autoretry_for = ( ConnectionError , TimeoutError ) , retry_backoff = True , ) def process_order ( self , order_id : str ) -
dict : result = do_processing ( order_id ) return { "order_id" : order_id , "status" : "completed" }
ARQ task with FastAPI
from arq import create_pool from arq . connections import RedisSettings async def generate_report ( ctx : dict , report_id : str ) -
dict : data = await ctx [ "db" ] . fetch_report_data ( report_id ) pdf = await render_pdf ( data ) return { "report_id" : report_id , "size" : len ( pdf ) } @router . post ( "/api/v1/reports" ) async def create_report ( data : ReportRequest , arq : ArqRedis = Depends ( get_arq_pool ) ) : job = await arq . enqueue_job ( "generate_report" , data . report_id ) return { "job_id" : job . job_id } Configuration Production Celery app configuration with secure defaults and worker tuning. Key Patterns JSON serialization with task_serializer="json" for safety Late acknowledgment with task_acks_late=True to prevent task loss on crash Time limits with both task_time_limit (hard) and task_soft_time_limit (soft) Fair distribution with worker_prefetch_multiplier=1 Reject on lost with task_reject_on_worker_lost=True Key Decisions Decision Recommendation Serializer JSON (never pickle) Ack mode Late ack ( task_acks_late=True ) Prefetch 1 for fair, 4-8 for throughput Time limit soft < hard (e.g., 540/600) Timezone UTC always Task Routing Priority queue configuration with multi-queue workers and dynamic routing. Key Patterns Named queues for critical/high/default/low/bulk separation Redis priority with queue_order_strategy: "priority" and 0-9 levels Task router classes for dynamic routing based on task attributes Per-queue workers with tuned concurrency and prefetch settings Content-based routing for dynamic workflow dispatch Key Decisions Decision Recommendation Queue count 3-5 (critical/high/default/low/bulk) Priority levels 0-9 with Redis x-max-priority Worker assignment Dedicated workers per queue Prefetch 1 for critical, 4-8 for bulk Routing Router class for 5+ routing rules Canvas Workflows Celery canvas primitives for sequential, parallel, and fan-in/fan-out workflows. Key Patterns Chain for sequential ETL pipelines with result passing Group for parallel execution of independent tasks Chord for fan-out/fan-in with aggregation callback Immutable signatures ( si() ) for steps that ignore input Nested workflows combining groups inside chains Link error callbacks for workflow-level error handling Key Decisions Decision Recommendation Sequential Chain with s() Parallel Group for independent tasks Fan-in Chord (all must succeed for callback) Ignore input Use si() immutable signature Error in chain Reject stops chain, retry continues Partial failures Return error dict in chord tasks Retry Strategies Retry patterns with exponential backoff, idempotency, and dead letter queues. Key Patterns Exponential backoff with retry_backoff=True and retry_backoff_max Jitter with retry_jitter=True to prevent thundering herd Idempotency keys in Redis to prevent duplicate processing Dead letter queues for failed tasks requiring manual review Task locking to prevent concurrent execution of singleton tasks Base task classes with shared retry configuration Key Decisions Decision Recommendation Retry delay Exponential backoff with jitter Max retries 3-5 for transient, 0 for permanent Idempotency Redis key with TTL Failed tasks DLQ for manual review Singleton Redis lock with TTL Scheduling Celery Beat periodic task configuration with crontab, database-backed schedules, and overlap prevention. Key Patterns Crontab for time-based schedules (daily, weekly, monthly) Interval for fixed-frequency tasks (every N seconds) Database scheduler with django-celery-beat for dynamic schedules Schedule locks to prevent overlapping long-running scheduled tasks Adaptive polling with self-rescheduling tasks Key Decisions Decision Recommendation Schedule type Crontab for time-based, interval for frequency Dynamic Database scheduler ( django-celery-beat ) Overlap Redis lock with timeout Beat process Separate process (not embedded) Timezone UTC always Monitoring Production monitoring with Flower, custom signals, health checks, and Prometheus metrics. Key Patterns Flower dashboard for real-time task monitoring Celery signals ( task_prerun , task_postrun , task_failure ) for metrics Health check endpoint verifying broker connection and active workers Queue depth monitoring for autoscaling decisions Beat monitoring for scheduled task dispatch tracking Key Decisions Decision Recommendation Dashboard Flower with persistent storage Metrics Prometheus via celery signals Health Broker + worker + queue depth Alerting Signal on task_failure Autoscale Queue depth > threshold Result Backends Task result storage, custom states, and progress tracking patterns. Key Patterns Redis backend for task status and small results Custom task states (VALIDATING, PROCESSING, UPLOADING) for progress update_state() for real-time progress reporting S3/database for large result storage (never Redis) AsyncResult for querying task state and progress Key Decisions Decision Recommendation Status storage Redis result backend Large results S3 or database (never Redis) Progress Custom states with update_state() Result query AsyncResult with state checks ARQ Patterns Lightweight async Redis Queue for FastAPI and simple background tasks. Key Patterns Native async/await with arq for FastAPI integration Worker lifecycle with startup / shutdown hooks for resource management Job enqueue from FastAPI routes with enqueue_job() Job status tracking with Job.status() and Job.result() Delayed tasks with _delay=timedelta() for deferred execution Key Decisions Decision Recommendation Simple async ARQ (native async) Complex workflows Celery (chains, chords) In-process quick FastAPI BackgroundTasks LLM workflows LangGraph (not Celery) Tool Selection Tool Best For Complexity ARQ FastAPI, simple async jobs Low Celery Complex workflows, enterprise High RQ Simple Redis queues Low Dramatiq Reliable messaging Medium FastAPI BackgroundTasks In-process quick tasks Minimal Anti-Patterns (FORBIDDEN) Load details: Read("${CLAUDE_PLUGIN_ROOT}/skills/async-jobs/references/anti-patterns.md") for full list. Key rules: never run long tasks in request handlers, never block on results inside tasks, never store large results in Redis, always use idempotency for retried tasks. Temporal Workflows Durable execution engine for reliable distributed applications with Temporal.io. Key Patterns Workflow definitions with @workflow.defn and deterministic code Saga pattern with compensation for multi-step transactions Signals and queries for external interaction with running workflows Timers with workflow.wait_condition() for human-in-the-loop Parallel activities via asyncio.gather inside workflows Key Decisions Decision Recommendation Workflow ID Business-meaningful, idempotent Determinism Use workflow.random() , workflow.now() I/O Always via activities, never directly Temporal Activities Activity and worker patterns for Temporal.io I/O operations. Key Patterns Activity definitions with @activity.defn for all I/O Heartbeating for long-running activities (> 60s) Error classification with ApplicationError(non_retryable=True) for business errors Worker configuration with dedicated task queues Testing with WorkflowEnvironment.start_local() Key Decisions Decision Recommendation Activity timeout start_to_close for most cases Error handling Non-retryable for business errors Testing WorkflowEnvironment for integration tests