What is machine learning? A machine learning algorithm (an algorithm is nothing but a set of rules or steps to achieve some outcome), is being trained how to play the classic Atari game 'breakout.' Ten minutes into the game, it was clumsy and missed the ball. Give it some more time, and it plays better than a human! The differentiation here is that instead of programing a traditional 'if-then-else' construct which explicitly directs machines to take rules-based decisions; we instead create algorithms that allow them to learn how to perform a particular task optimally, getting better and better with each iteration.
Now the first thing to note is that machine learning is not new. Check out this definition by Arthur Samuel:
“Field of study that gives computers the ability to learn without being explicitly programmed”. - Arthur Samuel, 1959
Arthur Samuel is perhaps the father of machine learning. He certainly coined the term. The Samuel checkers-playing program is probably the world's first self-learning program, and an early demonstration of a broader concept: artificial intelligence.
A very small history lesson
Reading Samuel's 1959 quote, you may have guessed that machine learning and AI in general are oldish. Which period do you think marks the beginning of AI? The '30s, '40s, '50s? Before reading on, take a guess.
The conception of AI is actually ancient. Thousands of years ago, the Greeks discussed the possibility of placing a mind inside a mechanical body. This automaton was exemplified by Talos, a legendary mechanical man created to protect the land from pirates and invaders. As early as the seventeenth century, intellectuals including Gottfried Leibniz, Thomas Hobbes and Rene Descartes imagined that all thought could be represented as mathematical symbols (an idea which is fundamental to neural networks).
Ancient and medieval roots aside, the true 'modern' birth of AI can be attributed to Alan Turing's publication of 'Computing Machinery and Intelligence' in 1950. The paper gave rise to the famous 'Turing test.' This test involved a computer, a human player and a human judge in a game. If the judge is unable to differentiate the human from the computer based on their interactions, the computer wins the game.
Over the next several decades, there were many milestones in the development of machine learning. These ranged from the conceptualization of the first neural network (called Perceptron), an algorithm inspired by how neurons behave in the brain; to IBM Watson's defeat of the world champions at Jeopardy! in 2011. This was a triumph of natural language processing and represented a significant leap forward for 'cognitive' technologies, as Jeopardy! is not a mathematically precise rules-based game like chess where the number of possible moves are limited.
Besides these big milestones, think about how we encounter machine learning every day, without even realising it. From the more obvious virtual assistants like Siri and Cortana (notice how they seem to get smarter after every interaction with you), to chatbots that can easily be mistaken for human service agents, recommender systems on search engines and websites, your ride-sharing app telling you the ETA of your next ride, the autopilot on the next flight you take, and your robo vacuum cleaner. Machine learning, and in a broader sense, AI, is creeping into virtually every aspect of our lives.
So what is the difference between AI, Machine Learning, and the even more mysterious Deep Learning?
You may have noticed that I use the term 'machine learning' instead of AI. Most experts agree that 'true AI' is still very far from reality. In other words, the Terminator is not going to be around blowing things up very soon, unless of course it is sent from the very distant future:-)
As a rule of thumb, machine learning is a subset of AI. And another form of highly specialised machine learning called Deep Learning is a subset of machine learning.
Deep learning deals with learning data representations, rather than task-specific algorithms. Deep neural networks, deep belief networks and recurrent neural networks have shown remarkable results in computer vision, speech recognition, natural language processing, machine translation and bioinformatics. Deep learning, requiring greater computational power than simpler machine learning algorithms, have come to the fore largely because of advances in computational technology such as the creation of GPUs and CPU-GPU architecture.
Sign up for Computerworld eNewsletters.