Deep Learning vs. Machine Learning vs. Pattern Recognition

Introduction:

Deep learning, machine learning, and pattern recognition are highly relevant topics commonly used in the field of robotics with artificial intelligence. Despite the overlapping similarities, these concepts are not identical. In this article, we will be discussing some of the differences of the three concepts and their applications.

Figure 1: An algorithm to detect the character "3" using sub-blocks

Three Popular Terms Correlated with "Learning"

Pattern recognition is the oldest form of learning and has become a relatively obsolete term. On the other hand, deep learning is a new and popular topic in the field of artificial intelligence. Machine learning, unlike the other two terms, is a fundamental form of learning and is one of the hottest areas in many start-ups and research labs. The Google Trends image below that shows the recent increase in interest for deep learning. Additionally, from the image, we can conclude that:

  • Starting from 2010, machine learning is steadily becoming popular again.
  • Pattern recognition used to be the hottest topic at the very beginning of the graph but is steadily declining.
  • Deep learning is a new and fast-rising area, beating the popularity of pattern recognition in 2015.

Figure 2: The Google search index of the three concepts since 2004 (Picture source: Google Trends)

Pattern Recognition: The Beginning of Intelligent Programs

The term pattern recognition became popular between the 1970s and the 1980s. It focuses on how to make computer programs perform intelligent and human-like tasks, such as an object from an image. Initially, people were not very interested in knowing how machines can achieve this intelligence, as long as it works correctly. However, technologies such as filters, boundary detection, and morphological processing have shown to be effective when applied to an image detection algorithm. Researchers in the pattern recognition community showed an increasing interest in this topic, spawning the field of optical character recognition.

It is appropriate to say that pattern recognition was the most innovative and "intelligent" signal processing of the 1970s, the 1980s, and even the early 1990s. Concepts such as decision tree, heuristic method, and quadratic discriminatory analysis were all introduced during this period. Pattern recognition slowly shifted from being a topic in electrical engineering to a topic of interest in computer science. One of the most famous books in pattern recognition, Pattern Classification by Duda and Hart, was released in 1973. Despite being published more than four decades ago, it is still a good introductory textbook for beginners seeking to know more about the pattern recognition field.

Machine Learning: Intelligent Programs that Learn from Samples

In the early 1990s, many realized that there was a more effective way to create pattern recognition algorithms, particularly replacing researchers with probability and statistics. This paradigm led to the creation of machine learning. The goal of machine learning is to give a computer a collection of data and let the computer make its own conclusion with minimal human intervention. Specifically, it implies that the computer (or the machine) collects statistics from data, and then generating a probabilistic model to determine the most possible outcome. When designed correctly, a machine learning algorithm always performs better than a person would because it is immune to cognitive bias and fatigue.

Figure 3: Typical machine learning process (Picture source: Dr. Natalia Konstantinova's blog).



In the middle of the 21st century, machine learning has emerged as an important research topic in computer science. Scientists have begun applying the concept broadly, creating new businesses using this technology. Machine learning has been used in robotics, genetic analysis, and in predictions for the financial market. Moreover, machine learning's combination with the graph theory created a new topic of research – the graph model.

Machine learning has become a basic skill for many people, but it has also caused a lot of confusion especially to people new to this field. We have seen a wide variety of methods and schools of thoughts in the machine-learning field, all having its own benefits.

Deep Learning: A Framework to Unite the World

Deep learning is currently a hot topic of research, specifically Convolutional Neural Network (or ConvNet), which has been used in large-scale graphic recognition.

Figure 4: ConvNet framework (Picture source: Torch's textbook)



In deep learning, there is minimal human intervention and bias because the parameters in the modes are learned from statistics. However, deep learning is only possible with an ample amount of statistics (big data) and strong arithmetic capabilities (graphic processor or GPU) to optimize the mode. Because convolution computation has been widely applied in computer vision, it is the natural choice for the mode of deep learning.

To understand deep learning, you should have a basic knowledge of linear algebra and programming. If you are not familiar with these topics, we strongly recommend Andrej Karpathy's blog "Hacker's Guide to Neural Networks."

Despite the benefits of deep learning, there are still many unsolved questions in its application. There are no existing theories concerning the validity of deep learning, nor textbooks on specific guidelines of deep learning. There have also been valid concerns for the possibility of artificial intelligence taking over jobs through deep learning. However, successful implementation of deep learning and artificial intelligence still requires plenty of human intervention. A high-quality product requires great vision, expertise of the field, market development, and most of all, the creativity of human beings.

Additional Relevant Technical Terms

  • Big Data: is an important concept that covers many aspects, such as the storage of massive data, and the mining of hidden information in data. For an enterprise operation, big data can offer valuable insights in decision-making. It was only several years ago that machine learning was integrated with big data.
  • Artificial Intelligence: is the oldest as well as the most encompassing technical term. Artificial intelligence is sometimes used to describe all topics related to learning, and its popularity has fluctuated in the past 50 years. In simple terms, artificial intelligence is the potential of a computer program or a device to think, learn, and interact with a human user. It is widely applied in fields such as healthcare, robotics, and finance.

Conclusion

The three popular terms relevant to artificial intelligence – deep learning, machine learning and pattern recognition – are highly correlated but are also unique are used in different applications. Pattern recognition was the first concept to be introduced in image processing, and eventually evolved to machine learning. To maximize the scope of application for machine learning, researchers actively searched for a method to automate machine learning, creating the field of deep learning. Deep learning is a relatively nascent field; there remains a lot of research to explore the full potential of deep learning.

时间: 2024-09-08 16:08:42

Deep Learning vs. Machine Learning vs. Pattern Recognition的相关文章

Learning Machine Learning, Part 3: Application

This post features a basic introduction to machine learning (ML). You don't need any prior knowledge about ML to get the best out of this article. Before getting started, let's address this question: "Is ML so important that I really need to read thi

Awesome Machine Learning

  Awesome Machine Learning  A curated list of awesome machine learning frameworks, libraries and software (by language). Inspired by awesome-php. If you want to contribute to this list (please do), send me a pull request or contact me @josephmisiti A

(转) Graph-powered Machine Learning at Google

    Graph-powered Machine Learning at Google     Thursday, October 06, 2016 Posted by Sujith Ravi, Staff Research Scientist, Google Research Recently, there have been significant advances in Machine Learning that enable computer systems to solve comp

booklist for machine learning

Recommended Books Here is a list of books which I have read and feel it is worth recommending to friends who are interested in computer science. Machine Learning Pattern Recognition and Machine Learning Christopher M. Bishop A new treatment of classi

Learning Machine Learning, Part 2: Algorithms and Techniques

The previous blog post, Introduction to Machine Learning, presented the Machine Learning concept. Now, let's discuss representative methods used in the technology. Regression Algorithms In most Machine Learning courses, regression algorithms are the

(转)Introduction to Gradient Descent Algorithm (along with variants) in Machine Learning

Introduction Optimization is always the ultimate goal whether you are dealing with a real life problem or building a software product. I, as a computer science student, always fiddled with optimizing my code to the extent that I could brag about its

(转)Is attacking machine learning easier than defending it?

  转自:http://www.cleverhans.io/security/privacy/ml/2017/02/15/why-attacking-machine-learning-is-easier-than-defending-it.html   Is attacking machine learning easier than defending it? Feb 15, 2017 by Ian Goodfellow and Nicolas Papernot In our first po

Teaching Your Computer To Play Super Mario Bros. – A Fork of the Google DeepMind Atari Machine Learning Project

  Teaching Your Computer To Play Super Mario Bros. – A Fork of the Google DeepMind Atari Machine Learning Project   Posted by ehrenbrav on August 25, 2016Leave a comment (14)Go to comments     For those who want to get right to the good stuff, the in

《Machine Learning with Spark》书评与作者访谈

机器学习是基于现有数据,通过数据驱动做出决策或者预测的技术.Apache Spark及其机器学习库MLlib为开发可伸缩的机器学习应用,提供了多种有用的算法.InfoQ采访了<Machine Learning with Spark>一书的作者Nick Pentreath,共同探讨了有关数据科学和机器学习的话题. By Srini Penchikala 机器学习以现有数据为基础创建模型,通过数据驱动,对未来做出决策或者预测.机器学习在最近几年获得大量的关注,因为它对企业与商业做出决策非常有帮助.