Pixelated images, ‘potato’ cameras, and pictures deliberately blurred to hide people’s faces, could become a thing of the past thanks to Google’s latest algorithm development. The company’s neural network can now take an image of just eight by eight pixels and use clever guesswork, to work its way back to a relatively decent quality photo.
Traditionally, blurred or pixellated images required a lot of manual attention to improve. Photoshop has some anti-camera-shake filters you can use, or you can manually go down to the pixel level and correct the pictures by hand. But that’s massively time-consuming and no good if you don’t know what was captured in the image, looks like. With Google’s system, you don’t need to (thanks Guardian).
The image on the left was the original input, while those on the right show various outputs by the algorithm
Google describes the technique as having the algorithm “hallucinate,” the missing details of the image. In essence, the system is taking an educated guess about what single pixels could mean and enhances them with additional ones. Google’s system was able to take eight by eight pixel images and turn them into 32 x 32 pixel images.
Adding a little blur to the edges, makes for a pretty good looking image, considering its source.
The images on the left, were what Google’s algorithm created, while those on the right were the original source images.
In its research paper entitled: Pixel Recursive Super Resolution, Google showed the ability to take images with no discernible details and create something that resembles a much higher quality version of the input image. This technique could theoretically be used to unblur family faces from photos, improve old photographs from the past automatically, or even help identify criminals from blurry, old security footage.
However, there are some concerns about using it for serious purposes. While the system can make educated guesses, it isn’t always entirely accurate. In some cases, the people’s faces that Google’s system ‘hallucinates’ into existence, are a little different than the people who were used for the test.
Discuss on our Facebook page, HERE.
KitGuru Says: There’s some very interesting potential here. I wonder too, if this sort of algorithmic method will make it possible to upscale a lot of old images in the future, to something resembling high-resolution.