We Need a Revolution of 4K!

Don’t panic or stop reading, we used the word ‘revolution’ in the title and though admittedly it’s provocative being less than a week from the US Presidential elections, we are talking about entertainment and TV, not politics. Cue the massive sigh of relief here…

Our story starts with a recent article published in PC Magazine titled “Meet Two Companies That Want to Revolutionize 4K Video”, where the author Troy Dreier examines the state of 4K and some of the issues surrounding the rate of 4K adoption, specifically a chicken-and-egg problem. As Dreier points out, 4K UHD TVs are being bought in considerable numbers “over 8 million 4K TVs to date, 1.4 million in the US.”

But what about content?

Although, 4K is already far more widely deployed in Asia Pacific and Western Europe, in the US cable and satellite customers are seeing limited content choices, with almost no options in broadcast, leaving consumers turning to online distribution services to satisfy their needs.

But with this comes another problem facing streaming providers, the commodity of the internet: bits.

Though the internet is getting much faster and infrastructure is improving, overall average speeds are still just 15.3 Mbps per household, making it difficult to deliver 4K UHD video sustainably. Or at least with the quality promise that the TV vendors are making. This ultimately, puts the pressure on network operators and over-the-top content suppliers to do everything they can to lower the number of bits they transport without damaging the picture quality of the video.

To this point, Dreier suggests that video optimization solutions are needed to “condense 4K video.” Dreier goes on to point out two solutions that are solving this problem, and one of them he highlights is Beamr’s content adaptive optimization solution, Beamr Video.

At the heart of our video encoding and processing technology solutions is the Beamr content adaptive quality measure that is backed up by more than 20 granted patents with another 30 still pending.  

The Beamr Video optimization technology is based on a proprietary, low complexity, reliable, perceptual quality measure. Or put simply, we have the most advanced commercially available content adaptive quality measure available. The existence of this measure enables controlling a video encoder, to obtain an output clip with maximal compression of the video input, while still maintaining the input video resolution, format and visual quality. This is performed by controlling the compression level frame by frame, in such a way that the maximum number of bits are squeezed out of the file, while still resulting in a perceptually identical visual output.

An important characteristic of our quality measures is that it operates as a full-reference to the source which insures that artifacts are never introduced as a result of the bitrate reduction process. Many “alternative” solutions struggle with inconsistent quality as they operate in an open loop, which means at times quality may be degraded while at other times they leave “bits on the table.”

With so much at stake for next generation entertainment formats, it is critical that every new encoding and video processing technology be evaluated for quality and useability. This is why we are proud of the customers we have which include major Hollywood studios, premium OTT content distributors, MSOs and large video platforms.

Beamr Video in the real world with 720p VBR input, reduced 21%:

beamr_video_live

For more information on the why and how behind content adaptive solutions, download the free Beamr Content Adaptive Tech Guide.

Translating Opinions into Fact When it Comes to Video Quality

This post was originally featured at https://www.linkedin.com/pulse/translating-opinions-fact-when-comes-video-quality-mark-donnigan 

In this post, we attempt to de-mystify the topic of perceptual video quality, which is the foundation of Beamr’s content adaptive encoding and content adaptive optimization solutions. 

National Geographic has a hit TV franchise on its hands. It’s called Brain Games starring Jason Silva, a talent described as “a Timothy Leary of the viral video age” by the Atlantic. Brain Games is accessible, fun and accurate. It’s a dive into brain science that relies on well-produced demonstrations of illusions and puzzles to showcase the power — and limitation — of the human brain. It’s compelling TV that illuminates how we perceive the world.(Intrigued? Watch the first minute of this clip featuring Charlie Rose, Silva, and excerpts from the show: https://youtu.be/8pkQM_BQVSo )

At Beamr, we’re passionate about the topic of perceptual quality. In fact, we are so passionate, that we built an entire company based on it. Our technology leverages science’s knowledge about the human vision system to significantly reduce video delivery costs, reduce buffering & speed-up video starts without any change in the quality perceived by viewers. We’re also inspired by the show’s ability to turn complex things into compelling and accessible, without distorting the truth. No easy feat. But let’s see if we can pull it off with a discussion about video quality measurement which is also a dense topic.

Basics of Perceptual Video Quality

Our brains are amazing, especially in the way we process rich visual information. If a picture’s worth 1,000 words. What’s 60 frames per second in 4k HDR worth?

The answer varies based on what part of the ecosystem or business you come from, but we can all agree that it’s really impactful. And data intensive, too. But our eyeballs aren’t perfect and our brains aren’t either – as Brain Games points out. As such, it’s odd that established metrics for video compression quality in the TV business have been built on the idea that human vision is mechanically perfect.

See, video engineers have historically relied heavily on two key measures to evaluate the quality of a video encode: Peak Signal to Noise Ratio, or PSNR, and Structured Similarity, or SSIM. Both metrics are ‘objective’ metrics. That is, we use tools to directly measure the physics of the video signal and construct mathematical algorithms from that data to create metrics. But is it possible to really quantify a beautiful landscape with a number? Let’s see about that.

PSNR and SSIM look at different physics properties of a video, but the underlying mechanics for both metrics are similar. You compress a source video where the properties of the “original” and derivative are then analyzed using specific inputs, and metrics calculated for both. The more similar the two metrics are, the more we can say that the properties of each video are similar, and the closer we can define our manipulation of the video, i.e. our encode, as having a high or acceptable quality.

Objective Quality vs. Subjective Quality


However, it turns out that these objectively calculated metrics do not correlate well to the human visual experience. In other words, in many cases, humans cannot perceive variations that objective metrics can highlight while at the same time, objective metrics can miss artifacts a human easily perceives.

The concept that human visual processing might be less than perfect is intuitive. It’s also widely understood in the encoding community. This fact opens a path to saving money, reducing buffering and speeding-up time-to-first-frame. After all, why would you knowingly send bits that can’t be seen?

But given the complexity of the human brain, can we reliably measure opinions about picture quality to know what bits can be removed and which cannot? This is the holy grail for anyone working in the area of video encoding.

Measuring Perceptual Quality

Actually, a rigorous, scientific and peer-reviewed discipline has developed over the years to accurately measure human opinions about the picture quality on a TV. The math and science behind these methods are memorialized in an important ITU standard on the topic originally published in 2008 and updated in 2012. ITU BT.500 (International Telecommunications Union is the largest standards committee in global telecom.) I’ll provide a quick rundown.

First, a set of clips is selected for testing. A good test has a variety of clips with diverse characteristics: talking heads, sports, news, animation, UGC – the goal is to get a wide range of videos in front of human subjects.

Then, a subject pool of sufficient size is created and screened for 20/20 vision. They are placed in a light-controlled environment with a screen or two, depending on the set-up and testing method.

Instructions for one method is below, as a tangible example.

In this experiment, you will see short video sequences on the screen that is in front of you. Each sequence will be presented twice in rapid succession: within each pair, only the second sequence is processed. At the end of each paired presentation, you should evaluate the impairment of the second sequence with respect to the first one.

You will express your judgment by using the following scale:

5 Imperceptible

4 Perceptible but not annoying

3 Slightly annoying

2 Annoying

1 Very annoying

Observe carefully the entire pair of video sequences before making your judgment.

As you can imagine, testing like this is an expensive proposition indeed. It requires specialized facilities, trained researchers, vast amounts of time, and a budget to recruit subjects.

Thankfully, the rewards were worth the effort for teams like Beamr that have been doing this for years.

It turns out, if you run these types of subjective tests, you’ll find that there are numerous ways to remove 20 – 50% of the bits from a video signal without losing the ‘eyeball’ video quality – even when the objective metrics like PSNR and SSIM produce failing grades.

But most of the methods that have been tried are still stuck in academic institutions or research labs. This is because the complexities of upgrading or integrating the solution into the playback and distribution chain make them unusable. Have you ever had to update 20 million set-top boxes? Well if you have, you know exactly what I’m talking about.

We know the broadcast and large scale OTT industry, which is why when we developed our approach to measuring perceptual quality and applied it to reducing bitrates, we were insistent on staying 100% inside the standard of AVC H.264 and HEVC H.265.

By pioneering the use of perceptual video quality metrics, Beamr is enabling media and entertainment companies of all stripes to reduce the bits they send by up to 50%. This reduces re-buffering events by up to 50%, improves video start time by 20% or more, and reduces storage and delivery costs.

Fortunately, you now understand the basics of perceptual video quality. You also see why most of the video engineering community believes content adaptive sits at the heart of next-generation encoding technologies.

Unfortunately, when we stated above that there were “all kinds of ways” to reduce bits up to 50% without sacrificing ‘eyeball video quality’, we skipped over some very important details. Such as, how we can utilize subjective testing techniques on an entire catalog of videos at scale, and cost efficiently.

Next time: Part 2 and the Opinionated Robot

Looking for better tools to assess subjective video quality?

You definitely want to check out Beamr’s VCT which is the best software player available on the market to judge HEVC, AVC, and YUV sequences in modes that are highly useful for a video engineer or compressionist.

VCT is available for Mac and PC. And best of all, we offer a FREE evaluation to qualified users.

Learn more about VCT: http://beamr.com/h264-hevc-video-comparison-player/