Welcome to BlogsInsider's beginner’s guide to AI. This (currently) four-part feature should provide you with a very basic understanding of what AI is, what it can do, and how it works. The guide contains articles on (in an order published) neural networks, computer vision, natural language processing, and algorithms. It’s not necessary to read them all, but doing so may better help your understanding of the topics covered.
Artificial intelligence has become a focal point for the global tech community thanks to the rise of deep learning. The radical advance of computer vision and natural language processing, two of AI’s most important and useful functions, are directly related to the creation of Artificial Neural Networks.
One of the more complex and misunderstood topics making headlines lately is artificial intelligence. People like Elon Musk warn that robots could one day destroy us all, while other experts claim that we’re on the brink of an AI winter and the technology is going nowhere. Making heads or tails of it all is difficult, but the best place to start is with deep learning. Here’s what you need to know.
For the purpose of this article, we’ll refer to artificial neural networks as, simply, neural networks. But, it’s important to know that deep learning techniques for computers are based on the brains of humans and other animals.
What is a Neural Network?
Scientists believe that a living creature’s brain processes information through the use of a biological neural network. The human brain has as many as 100 trillion synapses – gaps between neurons – which form specific patterns when activated. When a person thinks about a specific thing, remembers something, or experiences something with one of their senses, it’s thought that specific neural patterns “light up” inside the brain.
In the field of deep learning, a neural network is represented by a series of layers that work much like a living brain’s synapses. We know that researchers teach computers how to understand what a cat is – or at least what a picture of a cat is – by feeding it as many images of cats as they can. The neural network takes those images and tries to find out everything that makes them similar so that it can find cats in other pictures.
Think of it like this: when you were learning to read you might have had to sound out the letters so that you could hear them out loud and lead your young brain to a conclusion. But, once you’ve read the word cat enough times you don’t have to slow down and sound it out. At this point, you access a part of your brain more associated with memory than problem-solving, and thus a different set of synapses fire because you’ve trained your biological neural network to recognize the word “cat.”
Scientists use neural networks to teach computers how to do things for themselves. Here are a few examples of what neural networks do:
- This neural network takes dark images and makes them clear
- This one analyzes MRIs and displays what you’re thinking
- Here’s one that plays Super Mario Bros
- And one that creates pick-up lines
- Finally, this one is self-replicating
There are many different kinds of deep learning and several types of neural network, but we’ll be focusing on generative adversarial networks (GANs), convolutional the neural networks (CNN), and recurrent neural networks (RNNs)
As you can see neural networks tackle a wide variety of problems. In order to understand how they work – and how computers learn – let’s take a closer look at three basic kinds of neural network.
Generative Adversarial Network:
First up, the GAN. Ian Goodfellow, one of Google’s AI gurus, invented the GAN in 2014. To put it in laymen’s terms, a GAN is a neural network comprised of two arguing sides — a generator and an adversary — that fight among themselves until the generator wins. If you wanted to create an AI that imitates an art style, like Picasso’s for example, you could feed a GAN a bunch of his paintings.
One side of the network would try to create new images that fooled the other side into thinking they were painted by Picasso. Basically, the AI would learn everything it could about Picasso’s work by examining the individual pixels of each image. One side of it would start creating an image while the other determined if it was a Picasso. Once the AI fooled itself, the results could then be viewed by a human who could determine if the algorithm needed to be tweaked to provide better results, or if it successfully imitated the desired style.
Convolutional Neural Network:
CNN's, not to be confused with the news outlet, are convolutional neural networks. These networks, at least in theory, have been around since the 1940s, but thanks to advanced hardware and efficient algorithms they’re just now becoming useful. Where a GAN tries to create something that fools an adversary, a CNN has several layers through which data is filtered into categories. These are primarily used in image recognition and text language processing.
CNN's are among the most common and robust neural networks. Researchers use them for a myriad of things, including outperforming doctors in diagnosing some diseases.
Recurrent Neural Network:
Finally, we have the RNN or recurrent neural network. RNNs are primarily used for AI that requires nuance and context to understand its input. An example of such a neural network is a natural language processing AI that interprets human speech. One need look no further than Google’s Assistant and Amazon’s Alexa to see an example of an RNN in action.
To understand how an RNN works, let’s imagine an AI that generates original musical compositions based on human input. If you play a note the AI tries to ‘hallucinate’ what the next note ‘should’ be. If you play another note, the AI can further anticipate what the song should sound like. Each piece of context provides information for the next step, and an RNN continuously updates itself based on its continuing input – hence the recurrent part of the name.
Go Deeper:
There are at least a dozen other kinds of neural network, and the three covered here are much more nuanced than the scope of this article. But, if you’ve made it this far, you should have a conversational understanding of what neural networks are, and what they do. If you’d like to know more here’s a few suggestions to take your machine learning education to the next level:
- This free AI course from the University of Helsinki and Reaktor
- Some free tutorials for using Google’s open-source AI platform TensorFlow
- And this excellent article on The Grandfather by MIT’s Martin Giles.
And for all the biggest news on what’s happening in the world of neural networks, check out our artificial intelligence section.