Google Lens: Silent Revolution Nobody Foresaw
This week at Google’s I/O conference its CEO Sundar Pichai introduced a feature that will overhaul our online experience forever. Meet Google Lens – your door into the augmented reality. This is the thing IT crowd has been talking for years, but when the Google Glass didn’t take off, many viewers swiped the augmented reality left. It would never happen in our lifetime, they said. And they couldn’t be more wrong.
It took Google engineers two years to hone the technology to integrate with Google Photo, Google Assistant and… your life. Point your camera at a flower and Google will recognize it and give its name to you. Point the camera onto a concert’s bill and Google assistant will put an entry into your “to-visit list”. Your camera will help you to get info about shops, bars, cafes and restaurants, about parks and gyms, landmarks and routes. And no, we aren’t talking about QR codes here.
Google Lens is capable of recognizing objects in a human way. You wouldn’t think a flower from the first example features a QR code on its petals, would you? If you look at a picture of a rose you see the flower, the color of it, the background, the setting. You can say if it’s late summer or early fall and if there’s a bee or a dew drop on a petal. With the help of Google Lens algorithm your smartphone can do it, too! More than that! If the picture you’ve taken is dark or noisy Google Lens can lighten it. Or if a twig squeezed into the frame taking up the front Google Lens can erase it. Not in a Photoshop way, leaving the white space behind but simply remove the obstacle as if it never had been there in the first place!
Devoted Web surfers may not see anything outstanding in it, as they may have used ‘Search by picture’ function for a long time. But this is the other story. When asked to find an image Google and any other search engine, as the search by picture is a common feature now, compare simply the arrangement of color spots and gamut. That’s why, looking for a red rose against a blue wall you can get all kinds of red and blue pictures, the one with a fire hydrant included. It’s annoying, I know. But we used to shrug it away, saying: ‘Oh, it’s just a computer, it can’t do more’. Now it can!
It can tell a rose from a fire hydrant and a concert bill from a street art. Why? Because now Google is capable of processing all this big data. The company built the first ever AI-based data center. This is the silent revolution nobody foresaw and never heard of. You see, Google featured warehouse data centers for years, since the Google Assistant voice control launch. It took hundreds and hundreds of servers to process your request ‘Ok, Google, pizza delivery phone’ barked somewhere between 6 and 11 p.m.
So Google faced a problem: either to build some more giant data centers and bear the expenses and electricity bills or invent something completely new. And here when the TPU arrived. Google developed the state-of-art Tensor Processing Unit. Tensor calculations are so complicated that not every mathematician can grasp them. They are used to described complicated physical processes on Earth and in open space and they have come useful for processing big data as well.
Installing TPU in the servers, Google can save the space and money. But this is just the beginning. The TPU is used in neural network environment. What is neural network? This is a network architecture that imitates the human brain and its neurons, hence the name – neural. It’s a fresh technology and it has a great advantage over the old regular networks.
You see, brain neurons cooperate with each other to process any info we percept through our eyes, ears, skin. There are myriads of them, more than stars in the universe. And they all send signals to each other and communicate in a seemingly chaotic way. We, humans, do not list variants and do not compare them before doing anything. For instance, it takes as a fraction of a second to remember that this very rose is like the one you saw last year in your aunt Augusta’s garden, but not quite the same. And she would be delighted to see a new breed of it. Computer engineers of all the world are trying to imitate this intricate web of associations and memories in their products.
And Google pioneers the industry. So, thanks to TPU powered data centers Google Lens can recognize the flower and search the floral wiki and bring back the answer. Or it can warn you, ‘NO! This is not a cutie green creature, it’s the poisonous ivy, RUN!’
So far, I’m sorry to say, Google Lens is not on Google Store yet. But we in iGotOffer promise to download it first thing it turns up and test it for you. Stay tuned!