Smoke poured from the server rack long before the alerts fired. The deployment had triggered an FFmpeg large-scale role explosion, and performance dropped to zero in under forty seconds. This was not a bug in FFmpeg itself. It was an infrastructure failure: runaway process creation, ballooning memory use, and thread saturation hidden behind a routine media job.
An FFmpeg large-scale role explosion happens when workload parameters, cluster scheduling, and I/O management collide in the wrong way. High-volume transcoding can fork too many worker processes across nodes, spiking CPU load until the orchestrator fails. In distributed environments, the issue compounds: overlapping roles spawn more codecs, more threads, more disk writes. Resource isolation breaks. Queue latency grows. Jobs start to die in unpredictable order.
The trigger is often a mismatch between codec parallelization flags and the execution environment. Using -threads 0 across many containers on shared hardware can multiply active threads far beyond expectations. Combined with aggressive segmenting (-f segment, -map, or -filter_complex), FFmpeg cascades into hundreds or thousands of active roles. Each role competes for both CPU and I/O, thrashing disks and choking the cluster network. The result is system-wide collapse.