The Hybrid Model
The debate between cloud and on-premises infrastructure is a false dichotomy for most VFX studios. In practice, the answer is almost always "both." The question is not whether to use cloud -- it is which workloads belong in the cloud, which need to stay on-prem, and how to connect the two seamlessly.
The fundamental tension is simple. Interactive VFX work -- compositing in Nuke, animating in Maya, grading in Resolve -- demands sub-10ms latency between the artist and the storage system. That kind of responsiveness is trivial with a local NVMe array sitting three meters from the artist's workstation. It is physically impossible when the data lives in a data center 50 miles away, no matter how fast your internet connection is. The speed of light imposes hard limits that no amount of bandwidth can overcome.
Meanwhile, rendering is the opposite workload profile. A render farm does not care about latency. It cares about throughput -- how many frames per hour it can push through. Rendering is inherently parallelizable, meaning you can throw 1,000 machines at a job for two hours instead of running 10 machines for 200 hours. This burst-and-release pattern is exactly what cloud computing is designed for. You pay for compute only when you need it, and you can scale to capacities that would be financially impossible to own outright.
Why Most Studios Land on Hybrid
Studios that go all-cloud discover that their monthly compute bills for interactive workstations are surprisingly high, and artists complain about latency-induced lag during detailed compositing work. Studios that stay fully on-prem discover that their render farm sits idle 60% of the time and cannot keep up during crunch periods when every shot needs to turn around overnight. The hybrid model solves both problems: on-prem hardware for the workloads that demand it, cloud resources for everything else.
The 80/20 Split
A common pattern among studios with successful hybrid deployments is an 80/20 split: 80% of daily compute runs on-prem (artist workstations, editorial, dailies, baseline rendering), while 20% of compute -- the burst capacity needed during crunch -- runs in the cloud. This ratio keeps the on-prem hardware utilized at a healthy 70-85% while avoiding the capital expense of building for peak demand.
When Cloud Makes Sense
Cloud infrastructure is not universally better or worse than on-prem. It excels in specific scenarios where its strengths -- elastic scaling, geographic distribution, and zero capital expenditure -- create genuine advantages over owned hardware.
Burst Rendering During Crunch
This is the single most compelling use case for cloud in VFX. During the final weeks of a project, render volumes can spike 5-10x above baseline. Building an on-prem farm to handle peak demand means that 80% of your hardware sits idle most of the year. Cloud rendering lets you spin up hundreds or thousands of cores for a few days or weeks, then shut them down when the crunch is over. You pay only for the hours you use.
The major cloud render management platforms include:
- AWS Deadline Cloud. Amazon's managed render farm service. Deep integration with EC2 spot instances, S3 storage, and AWS Thinkbox tools. Supports all major DCC applications. The managed service handles queue management, worker provisioning, and auto-scaling.
- Google Cloud Batch. Google's batch computing service, suitable for render workloads. Strong integration with Google Cloud Storage and their high-memory instance types. Less VFX-specific tooling than AWS Deadline, but competitive pricing.
- Azure Batch. Microsoft's batch computing platform. Good integration with Azure Blob Storage and their NV-series GPU instances. Popular with studios already in the Microsoft ecosystem.
Remote Artist Access
Cloud-hosted virtual workstations enable studios to hire artists anywhere in the world without shipping hardware. An artist in London can work on the same project as a team in Los Angeles, with both accessing the same centralized storage. Virtual workstation platforms like Teradici/HP Anyware and Parsec stream pixels to the artist's local display, keeping all project data in the cloud and never on the artist's personal machine -- a critical requirement for TPN-compliant studios.
Disaster Recovery
Cloud storage provides an excellent offsite backup target. Replicating critical project data to S3 Glacier or Azure Archive ensures that a fire, flood, or ransomware attack at your physical location does not destroy your only copy of a project. Cloud-based DR also enables faster recovery than tape-based offsite backups, since data can be accessed immediately (or within hours for archive tiers) without waiting for physical media to be shipped.
Client Review Sessions
Cloud-hosted review platforms like Frame.io, Moxion, and ShotGrid Review allow clients to view and annotate work from anywhere, at any time. Rather than scheduling in-person screenings or dealing with the security risks of sending files via download links, studios can publish review media to a secure cloud platform where clients can view calibrated streams without ever downloading the source media.
When On-Prem Wins
On-premises infrastructure still holds decisive advantages for several core VFX workloads. These are not edge cases -- they represent the majority of an artist's daily workflow.
Real-Time Interactive Work
Compositing, color grading, animation, and texture painting all require the artist to see results instantly as they manipulate parameters. When a compositor adjusts a color correction in Nuke and the viewport updates 200ms later instead of immediately, the feedback loop breaks down and productivity plummets. On-prem storage -- particularly NVMe arrays connected via 100GbE -- delivers consistent sub-5ms latency that cloud storage simply cannot match from a remote data center.
The math is unforgiving. Light travels roughly 200km per millisecond in fiber. A round trip to an AWS region 40km away adds a minimum of 0.4ms of latency just from physics, before any processing overhead. In practice, cloud storage latency for random reads is typically 5-20ms, compared to 0.1-0.5ms for local NVMe. For sequential reads of large EXR sequences, the gap narrows, but for the small random I/O patterns that interactive DCC tools generate, on-prem storage is 10-50x faster.
Large-Scale Data Sets
When your active project data exceeds 100TB, the economics of cloud storage shift dramatically. At S3 Standard pricing of $23/TB/month, 200TB of active project data costs $4,600/month -- $55,200/year -- just for storage, before any compute or egress. A 200TB on-prem NAS costs roughly $30,000-50,000 as a one-time purchase, with annual maintenance of perhaps $5,000. The on-prem system pays for itself in under a year.
Consistent High-Bandwidth I/O
VFX workloads are unusually I/O intensive. A single Nuke composite might read 50-100 multi-layer EXR files per frame, each 50-200MB. A render farm with 100 nodes can easily saturate a 100Gbps network link. On-prem networks with 25GbE or 100GbE to each workstation and 100-400GbE backbones provide consistent, dedicated bandwidth that is not shared with other tenants or subject to internet congestion.
Security-Sensitive Content Under TPN
Studios working on major theatrical releases under NDA must comply with the Trusted Partner Network (TPN) security assessment. While TPN does not prohibit cloud infrastructure, many studios and their clients prefer the control and auditability of on-prem systems for the most sensitive content. Physical security controls -- badge access, camera surveillance, visitor logs, locked server rooms -- are easier to demonstrate and audit when the hardware is on your premises. Cloud deployments can achieve equivalent security, but the configuration is more complex and the audit trail requires more documentation.
The Latency Test
Before committing to cloud infrastructure for interactive work, run a simple test. Open your heaviest Nuke composite on a cloud-hosted workstation and compare the viewport responsiveness to the same setup running locally. If the difference is noticeable -- and for compositing work with heavy node trees, it almost always is -- that workload belongs on-prem. Save the cloud for rendering, review, and collaboration.
Cost Comparison
Cost comparisons between cloud and on-prem are notoriously difficult because they depend on utilization rates, project patterns, and how you account for capital expenditure versus operational expenditure. The following table provides a realistic monthly cost breakdown for a 50-seat VFX studio under three scenarios: fully on-prem, fully cloud, and hybrid.
| Cost Category | On-Prem Only | Cloud Only | Hybrid |
|---|---|---|---|
| Artist Workstations (50) | $8,300/mo* | $25,000/mo | $8,300/mo* |
| Render Farm | $4,200/mo* | $12,000-30,000/mo | $4,200/mo* + $3,000-8,000/mo cloud burst |
| Storage (300TB) | $1,500/mo* | $7,500/mo | $1,500/mo* + $500/mo cloud backup |
| Networking | $800/mo | $2,000-5,000/mo | $1,200/mo |
| Egress/Transfer | $0 | $3,000-8,000/mo | $500-2,000/mo |
| IT Staff (2 FTE) | $20,000/mo | $20,000/mo | $20,000/mo |
| Facility (power, cooling, space) | $3,000/mo | $0 | $2,500/mo |
| Monthly Total | $37,800/mo | $69,500-95,500/mo | $38,700-46,700/mo |
* On-prem hardware costs amortized over 4-year lifecycle. Assumes $200K workstations, $100K render farm, $60K storage infrastructure.
Several patterns emerge from this comparison. First, fully on-prem and hybrid are surprisingly close in total cost, because the hybrid model's cloud spending is offset by avoided overprovisioning of on-prem render capacity. Second, fully cloud is dramatically more expensive, driven primarily by the cost of running 50 cloud workstations 10 hours a day versus owning workstations outright. Third, egress costs are a significant line item in the cloud-only model -- they add up quickly when artists and render nodes are constantly reading and writing large files.
Egress: The Hidden Killer of Cloud Budgets
Cloud providers charge $80-90 per TB for data leaving their network (egress). In a VFX pipeline, data moves constantly -- renders go from cloud compute to cloud storage, review media goes to client platforms, dailies go to editorial. A busy studio can easily transfer 20-50TB per month in egress alone, adding $1,600-4,500/month that does not appear in simple storage or compute estimates. Always model egress costs explicitly when budgeting cloud infrastructure. Consider using providers with free or reduced egress (Cloudflare R2, Backblaze B2) for data that needs to be accessed frequently.
Latency & Data Transfer
The single biggest technical challenge in hybrid VFX infrastructure is moving data between on-prem and cloud fast enough that it does not become a bottleneck. VFX data sets are large -- a single feature film project can be 50-200TB -- and traditional internet upload speeds are far too slow for bulk transfers.
Accelerated Transfer Tools
Standard TCP-based transfers (SCP, rsync, FTP) are limited by TCP's congestion control algorithm, which does not fully utilize high-bandwidth, high-latency links. Accelerated transfer protocols use UDP-based transport that can saturate the available bandwidth regardless of latency.
- IBM Aspera. The industry standard for high-speed file transfer in media. Aspera's FASP protocol can transfer data at line speed over any distance. A 10Gbps connection will deliver close to 10Gbps of actual throughput, compared to 1-3Gbps with TCP-based tools over the same link. Aspera licenses are expensive ($10,000-50,000+/year depending on throughput tier), but for studios moving tens of terabytes regularly, the time savings justify the cost.
- Signiant. A direct competitor to Aspera, widely used in broadcast and post-production. Signiant's acceleration technology delivers similar throughput to Aspera, with a cloud-native management console that integrates with S3 and Azure Blob. Pricing is typically per-TB transferred, making it more predictable than Aspera's tiered licensing.
- AWS DataSync / Storage Gateway. Amazon's first-party data transfer service. Not as fast as Aspera or Signiant for single large transfers, but deeply integrated with AWS services and included at no additional cost beyond standard data transfer rates. Suitable for automated, scheduled synchronization rather than ad-hoc bulk transfers.
Dedicated Cloud Interconnects
For studios with consistent high-volume data transfer needs, a dedicated network connection to the cloud provider eliminates internet variability and provides guaranteed bandwidth.
- AWS Direct Connect. Dedicated 1Gbps, 10Gbps, or 100Gbps connections between your facility and the nearest AWS region. Reduces per-GB data transfer costs by roughly 50% compared to internet-based transfer. Typical provisioning time is 2-4 weeks. Monthly cost ranges from $200/month for a 1Gbps hosted connection to $5,000+/month for a 10Gbps dedicated connection, plus data transfer at $0.02/GB.
- Google Cloud Interconnect. Google's equivalent to Direct Connect. Available in Dedicated (10Gbps or 100Gbps) and Partner (50Mbps to 50Gbps through a partner) configurations. Pricing is similar to AWS. Google offers a discount on egress for traffic that traverses the interconnect.
- Azure ExpressRoute. Microsoft's dedicated connection service. Available in 50Mbps to 100Gbps configurations. Includes an unlimited data plan option that caps egress costs at a fixed monthly fee -- useful for studios with unpredictable transfer volumes.
Realistic Bandwidth Expectations
Here is how long common VFX data transfers take at different bandwidth levels:
| Data Volume | 1 Gbps | 10 Gbps | 100 Gbps |
|---|---|---|---|
| 1 TB (single shot) | 2.2 hours | 13 minutes | 80 seconds |
| 10 TB (episode/reel) | 22 hours | 2.2 hours | 13 minutes |
| 50 TB (full project sync) | 4.6 days | 11 hours | 67 minutes |
| 200 TB (studio migration) | 18.5 days | 1.85 days | 4.4 hours |
These numbers assume sustained throughput at the listed speeds, which is achievable with accelerated transfer tools but not with standard TCP. For transfers over 50TB, consider AWS Snowball or Google Transfer Appliance -- physical devices shipped to your facility that you load with data and ship back. While it sounds absurd, for very large data sets, a truck full of hard drives genuinely beats the internet.
Building a Hybrid Pipeline
A hybrid pipeline is not just "some stuff on-prem, some stuff in the cloud." It requires deliberate architecture to ensure that data flows efficiently, tools work consistently across both environments, and artists do not need to think about where their work is running.
Production Tracking Integration
Your production tracking system -- ShotGrid, ftrack, or similar -- is the central nervous system of the pipeline. It needs to know which shots are being worked on locally, which are queued for cloud rendering, and where the latest versions live. Most modern tracking tools have APIs that support custom integrations for routing work between on-prem and cloud resources based on priority, deadline, and resource availability.
Asset Management and Synchronization
The most critical decision in hybrid pipeline design is your synchronization strategy. There are three main approaches:
- Full sync. Mirror the entire project between on-prem and cloud storage. Simple to reason about but expensive in bandwidth and storage costs. Only practical for smaller projects (under 10TB) or studios with dedicated high-bandwidth interconnects.
- Selective sync. Synchronize only the files needed for the current cloud workload (render inputs, for example) and leave everything else on-prem. Requires pipeline tooling to identify and stage the correct files, but dramatically reduces transfer volumes. This is the most common approach.
- On-demand streaming. Do not pre-sync anything. Instead, cloud workers pull files from on-prem storage as needed through a caching proxy. This eliminates pre-staging but requires a low-latency, high-bandwidth connection between on-prem and cloud. Works well with dedicated interconnects; poorly over standard internet.
Cloud Rendering Queues
Set up your render management system (Deadline, Tractor, or Royal Render) to treat cloud instances as an extension of your on-prem farm. Artists should submit renders through the same interface regardless of where the job will execute. The render manager should automatically route jobs to cloud workers when on-prem capacity is exhausted, or when a supervisor flags a job as high-priority and needing maximum parallelism.
Key configuration decisions for cloud rendering:
- Spot vs On-Demand instances. Spot instances offer 60-90% savings but can be interrupted with 2 minutes notice. For renders, this is usually acceptable -- the render manager simply re-queues interrupted frames. Set up your farm to use spot instances by default and fall back to on-demand only for urgent, deadline-critical jobs.
- Instance right-sizing. Do not default to the largest available instance. Profile your typical render jobs to determine actual CPU, RAM, and GPU requirements, then select instances that match. Oversized instances waste money; undersized instances cause swapping and slowdowns.
- Pre-baked AMIs/images. Build custom machine images with all required DCC software, plugins, and license servers pre-configured. Booting a fresh instance and installing software on the fly wastes time and introduces configuration inconsistencies.
On-Prem Interactive Workstations
Keep artist workstations on-prem with direct access to local high-speed storage. The workstation-to-storage path should be 25GbE at minimum, with 100GbE preferred for heavy compositing and grading workflows. Use NVMe caching tiers on the storage server to accelerate reads of active shot data, with HDD tiers for less frequently accessed project files.
Start With Rendering, Expand From There
If you are building your first hybrid pipeline, start with cloud rendering only. It is the simplest integration point -- you are adding capacity to an existing render farm, not rearchitecting the entire pipeline. Get comfortable with cloud rendering workflows, data staging, and cost management before expanding to cloud workstations, cloud storage tiers, or more complex hybrid topologies.
Recommendations
There is no universal right answer -- the optimal infrastructure mix depends on your studio's size, project patterns, budget structure, and growth trajectory. Here are our tiered recommendations.
Small Studio (5-15 Artists)
At this scale, keep it simple. Cloud infrastructure adds operational complexity that a small team may not have the IT staff to manage. Focus on strong on-prem fundamentals.
- Workstations: All on-prem. Tower workstations with local NVMe for active work.
- Storage: On-prem NAS (Synology, QNAP, or TrueNAS) with 25GbE connectivity.
- Rendering: On-prem render nodes for baseline capacity. Use a cloud rendering service like Conductor or Google Cloud Batch for occasional burst needs, paying per-frame rather than managing cloud infrastructure yourself.
- Cloud: Backblaze B2 or Wasabi for offsite backup only. No cloud workstations or cloud-native pipeline.
- Budget: $5,000-8,000/month all-in.
Mid-Size Studio (15-50 Artists)
This is where hybrid begins to pay off. You have enough scale to justify dedicated IT staff and the pipeline complexity of a dual-environment setup.
- Workstations: On-prem for all full-time artists. Cloud virtual workstations (HP Anyware or Parsec) for remote contractors and overflow during crunch.
- Storage: On-prem primary storage (100-500TB). Cloud storage (S3 or GCS) for render outputs, review media, and offsite backup.
- Rendering: On-prem farm sized for 60-70% of peak demand. AWS Deadline Cloud or similar for burst capacity. Use spot instances for cost savings.
- Networking: 10Gbps internet minimum. Evaluate AWS Direct Connect or GCP Interconnect if monthly cloud transfer exceeds 20TB.
- Budget: $25,000-45,000/month.
Large Studio (50+ Artists)
At this scale, hybrid is not optional -- it is a competitive necessity. The capex savings of right-sized on-prem infrastructure combined with elastic cloud burst capacity can save hundreds of thousands of dollars per year compared to either pure approach.
- Workstations: On-prem for co-located artists. Cloud virtual workstations for all remote artists, with dedicated GPU instances and Teradici/HP Anyware.
- Storage: Enterprise on-prem SAN (Isilon, Pure, NetApp) with tiered storage policies. Cloud storage with automated lifecycle management (hot to cold tier migration).
- Rendering: On-prem farm for baseline (sized to 50-60% of peak). Cloud burst integrated into render management with automatic scaling policies. Reserved instances for predictable baseline cloud usage; spot for burst.
- Networking: Dedicated cloud interconnect (10-100Gbps Direct Connect or equivalent). Aspera or Signiant for accelerated transfers.
- Security: Full TPN compliance across both environments. VPC isolation in cloud, segmented VLANs on-prem, unified identity management (SSO/SAML).
- Budget: $60,000-150,000/month, heavily dependent on render volumes and remote artist count.
Plan for Where You Will Be, Not Where You Are
Infrastructure decisions have long lead times and switching costs. When designing your hybrid architecture, plan for your studio's size and workload 18-24 months from now, not today. It is much cheaper to build flexibility into the architecture upfront than to rearchitect mid-production because you outgrew your initial design. If you are a 20-person studio expecting to be 40 people in a year, design for 40.
Need Help Designing Your Hybrid Infrastructure?
We have helped studios of all sizes architect and deploy hybrid pipelines. Contact us for a free infrastructure assessment.