Super-Resolution: Why is it good and how can you incorporate it? – Display Daily

Welcome to Part 2 ofBitmovins Video Tech Deep Dive series: Super-Resolution with Machine learning.Before you get started, I highly recommend that you readSuper Resolution: Whats the buzz and why does it matter?. But if you would rather prefer to directly jump into it, here is a quick summary:

The focus of this series of blog posts will be on machine learning-based super-resolution.

In this post, we will examine:

Super-resolution, Machine learning (ML), and Video Upscaling are a match made in heaven. The three factors coming together is the reason behind the current popularity inMachine-learning based super-resolutionapplications. In this section, we will see why.

The concept of super-resolution has existed since the 1980s. The basic idea behind super-resolution was (and continues to be) tointelligentlycombinenon-redundant informationfrom multiple related low-resolution images to generate a single high-resolution image.

Some classic early applications were finding license plate information from several low-resolution images.

Several low-resolution snapshots of a moving car provides non-redundant but related information. Super-Resolution uses this related non-redundancy to create higher-resolution images, which can be useful in finding information such as license plate information or driver identification [Source].

But the recent wave of interest in super-resolution has been primarily driven by ML.

So, why ML and what changed now?

ML, in essence, is about learning theintelligencefor awell-defined problem. With the right architecture and enough data, ML can be significantly moreintelligentthan a human-defined solution (at least in that narrow domain). We saw this demonstrated stunningly in the case ofAlphaZero(for chess) andAlphaGo(for the board gameGo).

Super-resolution is awell-defined problem, and one could reasonably argue that ML would be a natural fit to solve this problem. With that motivation, early theoretical solutions were already proposed in the literature.

But, the exorbitant computational power and fundamental unresolved complexities kept the practical applications of ML-based super-resolution at bay.

However, in the last few years, there were two major developments:

These developments have led to a resurgence and come back for ML-based super-resolution methods.

It should be mentioned that ML-based super-resolution is a versatile hammer that can be used to drive manynails. It has wide applications, ranging frommedical imaging, remote sensing, astronomical observations, among others. But as mentioned inPart 1of this series, we will focus on howthe ML super-resolutionhammer can nail the problem ofvideo upscaling.

The last missing puzzle piece in this arc of the story isVideo upscaling.

When you think about it, video upscaling is almost a perfect nail for the ML-based super-resolution hammer.

Video provides the core features needed for the ML-based Super-Resolution. Namely:

The convergence of these three factors is why we are witnessing ahuge uptick in theresearchin this area, and also thefirst practical applicationsin the field of ML Super-Resolution powered Video upscaling.

I provided a historical timeline and the factors that lead to ML Super-Resolution powered Video upscaling. But, it might still not be clear on why it is superior to other traditional methods (bilinear,bicubic,Lanczos, among others). In this section, I will provide a simplified explanation to provide an intuitive understanding.

The superior performance simply boils down to the fact that the algorithm understands the nature of the content it is upsampling. And how it tunes itself to upsample that content in the best way possible. This is in contrast to the traditional methods where there is no tuning. In traditional methods, the same formula is applied without any consideration of the nature of the content.

One could say that:

ML-based super-resolution is to upsampling, whatPer-Titleis to encoding.

InPer-Title, we use different encoding recipes for the different pieces of content. In a similar way, ML-based super-resolution uses different upsampling recipes for different pieces of content.

The recipes can adapt itself on both at the:

Hopefully, by now, you are already excited about the possibilities of this idea. In this section, I would like to provide some suggestions on how you can incorporate this idea into your own video workflows and the potential benefits you might expect from it.

Broadly speaking, a video processing workflow typically has three steps involved:

Typically, there is a heavy emphasis on the encoding block for visual quality optimizations (Per-Title,3-Pass,Codec-Configuration, among others).

But, the other two (often overlooked) blocks are as important when it comes to visual quality optimization. In this instance, upsampling is a preprocessing step. And by choosing the right upsampling methods, such as super-resolution, one can improve the visual quality of the entire workflow. Sometimes, significantly more than that could be provided from the other blocks.

In the Part-3 of this series, we will delve more deeply into this. We will quantify how much quality improvements one could expect from tuning the pre-processing block with super-resolution. And use some real-life examples.

(This specific section is primarily meant for advanced readers who understand whatPer-Title,VMAF,convex-hullmeans. Please feel free to skip this section).

Like explained earlier, there are broadly three blocks in a video workflow. Roughly speaking, they work independently. But if we are smart about the design, we can extract synergies and use that to improve the overall video pipelines, that otherwise would not have existed.

One illustrative example is how Per-Title can work in conjunction with the Super-Resolution. This idea is depicted in the following figure.

VMAFvs Bitrate Convex hulls of video content. Green => 360p, Red => 720p, Blue => 1080p. BC : Bicubic, SR : SuperResolution.

In the above figure, for the illustrated bitrate: When using the traditional method the choice is clear. We will pick the 720p rendition. But, when using Super-Resolution, the choice is not very clear. We could either pick

The choice is determined by the complexity (vs) quality tradeoff that we are willing to make.

The takeaway message is two blocks synergistically working together to give more options and flexibility for the Per-Title algorithm to work with. Overall, a higher number of options translate to better overall results.

This is just one illustrative example, but within your own video workflows, you could identify regions where super-resolution can work synergically and improve the overall performance.

If your entire video catalog is a specific kind of content (anime for example), and you want to do a targeted upsample of these contents, then without doubtML Super-Resolution is the way to go!

In fact, that is what many companies alreadydo.This specific trend will only accelerate in the future, especially considering the popularity of consumer 4K TVs.

Visual quality enhancements,Synergies, andTargeted upsamplingare some ideas on how you can incorporate Super-Resolution into your video workflows.

Super-Resolution applied for targeted content such as Anime [Source]

We continued the story fromPart 1. We learned that :

In the follow-up, Part 3 of this series, we will look at how to do practical deployments, tools to use, and some real-life results.

This article was originally published as a blog post on the Bitmovin website byAdithyan Ilangovanand is re-published here with kind permission.

Originally posted here:
Super-Resolution: Why is it good and how can you incorporate it? - Display Daily

Related Posts

Comments are closed.