Turbopack: lazy aggregation optimize via persistent pending flag (#93454)
### What?
Make the aggregation `optimize_queue` in the Turbopack persistent
backend bounded and lazy. Cap the in-memory queue size, persist an
`optimization_pending` flag per task, drain the queue per `process()`
call with a per-queue lifetime budget, and recover dropped optimizations
opportunistically via the flag instead of via an unbounded
scheduler-side queue.
### Why?
The previous implementation pushed an `OptimizeJob` for every
`push_optimize_task` call into a single in-memory queue with no upper
bound. On large workloads (or pathological aggregation churn), this
queue could grow very large and cause the thread that scheduled the
optimization to do unbounded work, regressing latency for the operations
that triggered the schedule.
The goals of this change:
- Bound the worst-case work any single `process()` call does for
optimizations (per-queue budget).
- Bound the in-memory queue size so memory use is predictable.
- Avoid losing optimizations: anything we drop must be eventually
recovered.
- Keep the common fast path cheap — no extra `Meta`-category guard
acquisitions when the optimization flows through normally.
### How?
Persist a new `optimization_pending` flag on `TaskStorage`
(`storage_schema.rs`) and use it to drive lazy recovery in
`aggregation_update.rs`:
- `push_optimize_task` only enqueues an in-memory `OptimizeJob` if the
queue is under `MAX_OPTIMIZE_QUEUE_SIZE` (10000) and the per-queue
lifetime budget `MAX_OPTIMIZATIONS_PER_QUEUE` (1000) hasn't been
exhausted. If we can't enqueue, we set `optimization_pending = true` on
the task so a future operation that visits this task will re-discover
and re-enqueue the optimization.
- Every `AggregationUpdateJob` handler calls
`check_optimization_pending` on the primary task(s) it touches, which
re-enqueues the optimize job if the flag is set (and the queue/budget
allow).
- `process()` drains the `optimize_queue` one job at a time (preserving
the original "root first" ordering), counting against the per-queue
budget. Once the budget is exhausted, further `OptimizeJob`s in the
queue are dropped and the flag is left set on those tasks (so they
recover later).
- `optimize_task` clears `optimization_pending` at entry so the recovery
loop eventually settles.
- The flag is **only** written on the drop path — the common case
(enqueue → process normally) does not touch `optimization_pending`, so
no `Meta`-category guard contention is added on the hot path.
- `OptimizeJob` carries a best-effort `flag_already_set` snapshot so
that when a job is dropped at process time and the snapshot says the
flag was already set, we skip the redundant write entirely. Most jobs
originating from `check_optimization_pending` (the recovery path) and
`optimize_task`'s self-re-enqueue carry this hint.
- `try_enqueue_optimize_job` is `#[must_use]` so the contract \"if this
returns false, set the flag\" is enforced at the type level.
`lock_and_mark_optimization_pending` is shared between
`push_optimize_task_by_id` and the budget-exhausted drop branch.
- `optimizations_executed` is intentionally persisted with the queue so
that suspending and resuming the queue cannot reset the per-queue
budget.
<!-- NEXT_JS_LLM_PR -->
---------
Co-authored-by: v-work-app[bot] <262237222+v-work-app[bot]@users.noreply.github.com>
Co-authored-by: Claude <noreply@anthropic.com>
Co-authored-by: Tobias Koppers <sokra@users.noreply.github.com>