Data Caps, Zero-rated, Net Neutrality: The Video Tsunami Doesn’t Take Sides

We Need to Work Together to Conserve Bits in the Zettabyte Era

Over the past year, and again last week, there has been no shortage of articles and discussion around data caps, binge-on, zero rated content, and of course network neutrality.

We know the story. Consumer demand for Internet and over-the-top video content is insatiable. This is creating an unstoppable tsunami of video.

Vendors like Cisco have published the Visual Network Index to help the industry forecast how big that wave is, so we can work together to find sustainable ways to deliver it.

The Cisco VNI is projecting that internet video traffic will more than double to 2.3 Zettabytes by 2020. (Endnote 1.) To put it another way, that’s 1.3 Billion DVDs of video crossing the internet daily in 2020, versus the 543 Million DVDs of video that crossed the internet today.

That’s still tough to visualize, so here’s a back-of-the-envelope thought experiment

Let’s take the single largest TV event in history, Super Bowl 49.

114 million viewers on average, every minute, watched Super Bowl 49 in 2015. The broadcast is about 3 hours and 35 minutes.  We might say that 24.5 Billion cumulative viewer-minutes of video were watched.

Assume that a DVD holds 180 minutes of video. (Note, this is an inexact guess assuming a conservative video quality.) If one person watched 543 Million DVDs of video, she would have to spend 97.8 billion cumulative minutes watching all of it. That’s four Super Bowl 49s every day.

And in 2020, it’s going to be close to 10 Super Bowl 49s of cumulative viewer-minutes of video trafficking across the network. In one day.

That is a lot of traffic and it is going to be hard work to transport those bits in a reliable, high-quality fashion that is also economically sustainable.

And that’s true no matter whether you are a network operator or an over-the-top content distributor. Here’s why.

All Costs are Variable in the Long-run

Recently, Comcast and Netflix have agreed to partner, which bodes well for both companies’ business models, and for the consumer at large. However, last week there were several news headlines about data caps and zero-rated content. These will undoubtedly continue.

Now, it’s obvious that OTT companies like Netflix & M-GO need to do everything they can to reduce the costs of video delivery. That’s why both companies have pioneered new approaches to video quality optimization.

On the other hand, it might seem that network operators have a fixed cost structure that gives them wiggle room for sub-optimal encodes.

But it’s worth noting this important economic adage: In the long run, all costs are variable. When you’re talking about the kind of growth in video traffic that industry analysts are projecting to 2020, everything is a variable cost.

And when it comes to delivering video sustainably, there’s no room for wasting bits. Both network operators and over-the-top content suppliers will need to do everything they can to lower the number of bits they transport without damaging the picture quality of the video.

In the age of the Zettabyte, we all need to be bit conservationists.

 

Endnote 1: http://www.cisco.com/c/dam/m/en_us/solutions/service-provider/vni-forecast-widget/forecast-widget/index.html

Translating Opinions into Fact When it Comes to Video Quality

This post was originally featured at https://www.linkedin.com/pulse/translating-opinions-fact-when-comes-video-quality-mark-donnigan 

In this post, we attempt to de-mystify the topic of perceptual video quality, which is the foundation of Beamr’s content adaptive encoding and content adaptive optimization solutions. 

National Geographic has a hit TV franchise on its hands. It’s called Brain Games starring Jason Silva, a talent described as “a Timothy Leary of the viral video age” by the Atlantic. Brain Games is accessible, fun and accurate. It’s a dive into brain science that relies on well-produced demonstrations of illusions and puzzles to showcase the power — and limitation — of the human brain. It’s compelling TV that illuminates how we perceive the world.(Intrigued? Watch the first minute of this clip featuring Charlie Rose, Silva, and excerpts from the show: https://youtu.be/8pkQM_BQVSo )

At Beamr, we’re passionate about the topic of perceptual quality. In fact, we are so passionate, that we built an entire company based on it. Our technology leverages science’s knowledge about the human vision system to significantly reduce video delivery costs, reduce buffering & speed-up video starts without any change in the quality perceived by viewers. We’re also inspired by the show’s ability to turn complex things into compelling and accessible, without distorting the truth. No easy feat. But let’s see if we can pull it off with a discussion about video quality measurement which is also a dense topic.

Basics of Perceptual Video Quality

Our brains are amazing, especially in the way we process rich visual information. If a picture’s worth 1,000 words. What’s 60 frames per second in 4k HDR worth?

The answer varies based on what part of the ecosystem or business you come from, but we can all agree that it’s really impactful. And data intensive, too. But our eyeballs aren’t perfect and our brains aren’t either – as Brain Games points out. As such, it’s odd that established metrics for video compression quality in the TV business have been built on the idea that human vision is mechanically perfect.

See, video engineers have historically relied heavily on two key measures to evaluate the quality of a video encode: Peak Signal to Noise Ratio, or PSNR, and Structured Similarity, or SSIM. Both metrics are ‘objective’ metrics. That is, we use tools to directly measure the physics of the video signal and construct mathematical algorithms from that data to create metrics. But is it possible to really quantify a beautiful landscape with a number? Let’s see about that.

PSNR and SSIM look at different physics properties of a video, but the underlying mechanics for both metrics are similar. You compress a source video where the properties of the “original” and derivative are then analyzed using specific inputs, and metrics calculated for both. The more similar the two metrics are, the more we can say that the properties of each video are similar, and the closer we can define our manipulation of the video, i.e. our encode, as having a high or acceptable quality.

Objective Quality vs. Subjective Quality


However, it turns out that these objectively calculated metrics do not correlate well to the human visual experience. In other words, in many cases, humans cannot perceive variations that objective metrics can highlight while at the same time, objective metrics can miss artifacts a human easily perceives.

The concept that human visual processing might be less than perfect is intuitive. It’s also widely understood in the encoding community. This fact opens a path to saving money, reducing buffering and speeding-up time-to-first-frame. After all, why would you knowingly send bits that can’t be seen?

But given the complexity of the human brain, can we reliably measure opinions about picture quality to know what bits can be removed and which cannot? This is the holy grail for anyone working in the area of video encoding.

Measuring Perceptual Quality

Actually, a rigorous, scientific and peer-reviewed discipline has developed over the years to accurately measure human opinions about the picture quality on a TV. The math and science behind these methods are memorialized in an important ITU standard on the topic originally published in 2008 and updated in 2012. ITU BT.500 (International Telecommunications Union is the largest standards committee in global telecom.) I’ll provide a quick rundown.

First, a set of clips is selected for testing. A good test has a variety of clips with diverse characteristics: talking heads, sports, news, animation, UGC – the goal is to get a wide range of videos in front of human subjects.

Then, a subject pool of sufficient size is created and screened for 20/20 vision. They are placed in a light-controlled environment with a screen or two, depending on the set-up and testing method.

Instructions for one method is below, as a tangible example.

In this experiment, you will see short video sequences on the screen that is in front of you. Each sequence will be presented twice in rapid succession: within each pair, only the second sequence is processed. At the end of each paired presentation, you should evaluate the impairment of the second sequence with respect to the first one.

You will express your judgment by using the following scale:

5 Imperceptible

4 Perceptible but not annoying

3 Slightly annoying

2 Annoying

1 Very annoying

Observe carefully the entire pair of video sequences before making your judgment.

As you can imagine, testing like this is an expensive proposition indeed. It requires specialized facilities, trained researchers, vast amounts of time, and a budget to recruit subjects.

Thankfully, the rewards were worth the effort for teams like Beamr that have been doing this for years.

It turns out, if you run these types of subjective tests, you’ll find that there are numerous ways to remove 20 – 50% of the bits from a video signal without losing the ‘eyeball’ video quality – even when the objective metrics like PSNR and SSIM produce failing grades.

But most of the methods that have been tried are still stuck in academic institutions or research labs. This is because the complexities of upgrading or integrating the solution into the playback and distribution chain make them unusable. Have you ever had to update 20 million set-top boxes? Well if you have, you know exactly what I’m talking about.

We know the broadcast and large scale OTT industry, which is why when we developed our approach to measuring perceptual quality and applied it to reducing bitrates, we were insistent on staying 100% inside the standard of AVC H.264 and HEVC H.265.

By pioneering the use of perceptual video quality metrics, Beamr is enabling media and entertainment companies of all stripes to reduce the bits they send by up to 50%. This reduces re-buffering events by up to 50%, improves video start time by 20% or more, and reduces storage and delivery costs.

Fortunately, you now understand the basics of perceptual video quality. You also see why most of the video engineering community believes content adaptive sits at the heart of next-generation encoding technologies.

Unfortunately, when we stated above that there were “all kinds of ways” to reduce bits up to 50% without sacrificing ‘eyeball video quality’, we skipped over some very important details. Such as, how we can utilize subjective testing techniques on an entire catalog of videos at scale, and cost efficiently.

Next time: Part 2 and the Opinionated Robot

Looking for better tools to assess subjective video quality?

You definitely want to check out Beamr’s VCT which is the best software player available on the market to judge HEVC, AVC, and YUV sequences in modes that are highly useful for a video engineer or compressionist.

VCT is available for Mac and PC. And best of all, we offer a FREE evaluation to qualified users.

Learn more about VCT: http://beamr.com/h264-hevc-video-comparison-player/

 

The TV of Tomorrow Needs Standards Today: Why the streaming video industry must work together to solve video delivery quality issues

Nearly 50 percent of Americans have an entertainment subscription service like Netflix, Amazon Prime, or Hulu, accessed via a connected television or devices like Amazon Fire TV, Roku, or Apple TV, according to recent research from Nielsen. Furthermore, a quarter of those in the coveted 18-to-34 demographic have either cut their cable or satellite services or never signed up for a pay-TV package, according to ComScore.

It’s Not Just Millennials Cutting the Cord – Content Providers Are Too

For decades cable and satellite services provided the exclusive gateway to mass audiences for premium and niche content channels.  Today, with the ease and availability to go consumer-direct via the Internet and over-the-top streaming (OTT), new networks are joining video platforms and licensing content to transactional, and subscription video-on-demand services, at an unprecedented rate.  The future of streamed TV whenever and wherever the viewer desires, is becoming a reality.  Or is already the reality for an ever-growing percentage of US households.

Yet to reach the consumer where they are means today’s content publisher must support a wide array of devices and players to enable video viewing ‘anytime and anywhere’ across computers, televisions, and mobile devices.  But device capabilities can vary significantly, and any modification means the content publisher must build different applications to support each device, and to ensure the best possible user experience.

Solving these issues will require collaboration among many players, who each have a vested interest in building the digital (streaming) OTT industry, in a quest to meet and exceed the “broadcast quality” standard that viewers have come to expect.

As streaming or OTT moves from a novelty to dominate distribution method, viewers are demanding better quality.  Leading streaming experience measurement company, Conviva, consistently reports in their user experience consumer survey results, that re-buffering events and video quality are the most cited frustrations for consumers watching online video.  With the adoption of new technologies such as 4K, virtual reality and OTT delivery of broadcast events, the demands on bandwidth will notably increase. Which explains why M-GO, a leading video on demand premium movie service partnered with Samsung and recently acquired by Fandango, reported that when they reduced bitrates using perceptual content adaptive technology, they experienced improvements in their streaming user experience and consumer satisfaction.

The key role that video quality plays in impacting user engagement and UX – and consequently service provider revenues, has incited recent efforts to improve video quality. This includes progress on adaptive bitrate selection, better transport-layer algorithms, and CDN optimization. Think about it – a single IP video packet contains approximately 1,400 bytes of information, and each IP packet contains multiple MPEG encapsulated video packets. The loss of even one IP packet can lead to video impairments lasting a half second or more.

The Need for Standardization Before Reaching the End User

While efforts are valuable and demonstrate potential improvements, one of the key missing pieces is an understanding of the structure that handcuffs video quality. That starts at the client-side before reaching the client. Standardization of online video quality, particularly the quality of experience (QoE), is more important than ever. But traditional methods of measuring quality do not translate well to OTT video.

Pay TV operators such as cable companies, have a specific advantage when it comes to the quality they can deliver, and that is, they control every aspect of the delivery process including the network and playback device, known as the STB or set-top-box. In contrast, the OTT delivery structure is fragmented, dangling by multiple vendors – from delivery, storage, transcoding – all who are responsible for parts of the overall system. Viewers care little about the complex network or routes involved to get their content to a device.  They simply expect the same high-quality viewer experience that they are accustomed to with traditional pay TV or broadcast systems.

This fragmentation, coupled with numerous formats that must be supported across devices, the need for standardization and the related challenges are apparent. While we rely on monitoring and analysis, there is enough variation in the measurement methodologies and definitions across the industry to impede our ability to not only maintain – but improve video quality. More than one video engineer would likely admit privately, that they spend their day just making sure the video is working, and only after this task is accomplished, do they consider what can be done to improve the quality of the video that they are delivering.

Strides are being made to develop and evangelize best practices for high-quality delivery of video over the Internet, thanks in part to the Streaming Video Alliance (SVA). The recommendations, requirements, and guidelines being assembled by the SVA are helping to define new industry architectures and contribute to building best practices across the streaming video ecosystem to accelerate adoption worldwide.

Standards Pave the Way for Valuable Metrics

Without agreed-upon industry standards for both quality of service (QoS) and quality of experience (QoE), there can be no objective benchmark or performance measurement in the video delivery ecosystem.

The SVA’s guidelines define a common language that describes the effectiveness of network delivery and outlines key network delivery metrics: ideal video startup time, acceptable re-buffering ratio, average video encoding bitrates, and video start failure metrics.

The alliance’s main focus is on the bits you’ll never see – like optimization and delivery techniques.  As a technology enabler that addresses these bits, improving all the above, we are excited to join with content providers, CDNs and service providers to tackle the most pressing issues of streaming video delivery.

Content Is Going Everywhere

To feed the beast, the industry must band together to provide constant easy access to high-quality video content.  The name of the game is getting content to your consumer with the best quality and highest user experience possible, and the only way to do that is to increase file efficiency by optimizing file sizes and conserving bandwidth, to cut through the Internet clutter.

Today, consumers have widespread access to streaming video services with content choices coming online in ever greater quantities and at vastly improved quality. What is critically lacking is a broad-spectrum understanding of the nature of video quality problems as they occur.  Also, the cost of bandwidth in the age of data caps continues to be an open question. To help navigate through the clutter and help answer critical questions, visit our resource center for useful information.

 

NAB 2015, Here We Come!

The Beamr team will be at the NAB Show in Las Vegas on April 13-16, and we would love to meet you!

Beamr-Video-Nab-Show-2015

Check out the top 5 reasons to stop by Beamr’s main booth #SU10425, South Upper Hall:

1) Experience first-hand how Beamr Video optimization improves user experience for streaming video while reducing delivery costs.
2) Learn how Beamr can cut video bitrates by up to 50%, while retaining their full resolution and quality.
3) Find out how Beamr fits seamlessly into your existing workflow.
4) T-shirt giveaways. . . have you ever seen an optimized T-shirt?
5) Think you’ve got ‘the eye’? We’ve got two videos. Tell us which is the original video and which is the optimized video.

We’ll also be presenting at the Amazon AWS booth, #SL9016 in the South Lower Hall. This is where we’ll show our cloud solution, which provides fully scalable video optimization capabilities accessible via REST APIs or cloud storage integration. Don’t miss our session on Media Optimization in the Cloud at the Amazon AWS Theater on Tuesday, April 14, at 1:30 PM.

We can’t wait to see you at NAB!

 

Jan Ozer: “Beamr Delivers The Benefit of HEVC/VP9 Without The Implementation Hassles”

Jan Ozer, one of the world’s top experts on video compression and streaming and Contributing Editor of Streaming Media Magazine, has done a thorough evaluation of Beamr Video using test clips that he encoded.  The evaluation included analyzing the quality of the resulting clips using objective measures such as PSNR, SSIM and VQM, and conducting extensive viewing of the optimized clips, assessing their subjective quality compared to the original clips.  Following this evaluation, Jan Ozer concludes that “for 99.99% of the expert and average viewers, the Beamr-processed clips would be perceptually identical to their source”.

As for bitrate reduction, Jan has found that Beamr Video was able to reduce the bitrate of his test clips by 50% on average in High Quality mode, and 35% on average in Best Quality mode.  Providing such significant bitrate reduction, while producing video clips that are perceptually identical to the source, brings Jan to the following conclusion: “If you’re spending big dollars on bandwidth, Beamr represents an opportunity to reduce it significantly, essentially delivering the benefit of HEVC/VP9 without the implementation hassles”.

Jan Ozer has worked in digital video since 1990, and is the author of over 20 books on video compression, editing and production, which have been adopted as textbooks by multiple colleges and universities.  As Contributing Editor at Streaming Media Magazine, Jan has reviewed most cloud, enterprise and desktop encoding tools, and has worked with many online video platforms (OVPs) and live streaming services.

You can read Jan’s article summarizing his findings about Beamr video here and download his full report here.

Beamr-Video-Reduced-Bitrate