Lower Storage TCO & Improve Efficiency for M&E Orgs

As studios & post-production houses develop new content, it is critical to lower storage TCO & improve efficiency for remote video workflows.

mapping-storage-workflows-nvme-nas-san-archive-cloud-archive
Mapping Video Data Storage to WorkflowsUsing NVMe, Ultra Fast SAN/NAS, NAS/Filer, Local Archive and Cloud Archive

In today’s economic climate, it’s necessary to do more with less—perhaps more now than ever before. With studios and post-production houses just starting up again, there’s been a large loss of revenue…to an industry that purchases based on when projects take place. Thus, investing in new infrastructure is difficult…even when the need is there.

Tier 1 Storage Capacity for Mission-Critical Applications & Data

With tier 1 storage (also referred to as production storage), available capacity has to be there when it is needed for a project. Customers look to tier 1 storage to provide high performance that meets the demanding requirements of applications. It also must provide protection against hardware failures, data durability, and flexibility to support multiple diverse applications simultaneously. But with these requirements, comes cost. For example, Quantum’s latest f1000, entry level NVME SAN provides 78.6 TB of capacity for just over $81,000 MSRP. That is over $1,000 per TB of storage. In this day and age, it’s easy to see how quickly that can be consumed when an uncompressed, hour-long 4k video file often exceeds 80GB.

What About Data and Content That is Aging?

Even though tier 1 storage is expensive and intended/designed for appropriate high performance, it tends to also house aging data/content that needs to be instantly accessible. Most tier 1 storage is designed for a 3-year lifecycle, and then often becomes cost-prohibitive to maintain. Even though most organizations backup files to tape storage or cloud storage, a lot of this data still remains on tier 1 storage. This may be due to a client or user wanting it online and available if needed, or due to a project that is still in process with already rendered content. Other times, data growth has spun out of control and has become difficult to manage. The latter is a reality (and maybe a nightmare) for many organizations today. Implementing a 2nd tier of less expensive storage while maintaining the resilience and instant availability of the content is the best way to solve this challenge.

How Often Do You Access Data on Primary Storage?

By only keeping recently accessed content on Tier 1 and shifting less frequently accessed content to a 2nd tier of storage, it allows an organization to keep costs down. For example, Caringo Swarm’s Single Server Appliance offers 168TB of capacity in a 1u form factor with a MSRP just over $77,000 or $458/per TB. That is more than 50% less than the Quantum model mentioned above. With the tiered model, you would have approximately ⅓ of the capacity in tier 1 (78.6 TB) and about ⅔ in tier 2 (168TB). This would save the organization about $85,000 (3 x f1000 vs 1xf1000 + 1 SSA).

Should I Move Some of My Data Off of Tier 1 Storage?

While lowering TCO on tier 1 storage and using it more efficiently is important, it’s equally if not more important to be able to access that data quickly when needed. And, you need to ensure that it’s on reliable and scalable storage designed for the long term. Caringo Swarm Object Storage is a great fit for that. It uses standard servers and is scalable to hundreds of PBs. Swarm also has built-in data protection, security, multi-tenancy, metadata search and accessibility via HTTP, S3, NFS and SMB.

How Can I Move My Data Off Tier 1 SAN and NAS Storage to Object Storage?

By integrating with tier 1 storage platforms such as NetApp, EditShare, Pixit, StorNext, etc., content and data can easily be transferred from tier 1 production storage to your Swarm cluster. With most MAMs supporting S3, you can also use a MAM to transfer data between storage tiers. With Swarm’s built-in versioning, you can easily select the object with the corresponding metadata that needs to be retrieved. Swarm also has built-in content management, so that objects can be viewed or streamed directly from our storage and then shared with another user. Lastly, Swarm can back up to any S3-compliant 3rd party service or S3 endpoint. So, you can have data/content on prem for a certain period of time if required for the client, and then back it up to a public cloud such as Amazon, Microsoft Azure or Wasabi for long-term storage.

Ready to Learn More About How to Lower Your Storage TCO?

If you are ready to learn more about how to lower your storage TCO while keeping your content and data protected and accessible, email me today!

David Fabrizio
David Fabrizio

About The Author

David Fabrizio joined Caringo in 2016 as the Western US Regional Manager and is currently our Business Development Manager. With 19 years in data storage and backup sales. David enjoys helping customers solve their storage challenges and meet their business goals. David loves spending time with family, listening to music and traveling.


More From Author