What is machine learning, anyways?
A brief definition, history and overview about how computers learn.
Definition of machine learning
Arthur Samuel, an American pioneer in the field of computer gaming and artificial intelligence, coined the term “machine learning”. He defined machine learning as a “Field of study that gives computers the ability to learn without being explicitly programmed” (Samuel 1959). This definition emphasizes the ability of machines to improve their performance on a specific task over time, based on experience, without the need for human intervention to explicitly program their behaviors.
A more formal, mathematical definition of machine learning involves concepts from statistics and computer science, focusing on the development of algorithms that can learn from and make predictions or decisions based on data. Here is a more detailed formalization:
Machine learning algorithms build a model based on input data (known as “training data”) to make predictions or decisions without being explicitly programmed to perform the task. These models can be understood in terms of functions, optimization, and probability.
Function: A machine learning model can be seen as a function
Optimization: The process of learning involves finding the function
Probability: In a probabilistic view, machine learning involves estimating the conditional probability
Therefore, a more formal definition could be: Machine learning seeks to learn a function
The history of artificial intelligence and machine learning
The history of machine learning is a fascinating journey that intertwines with the development of computers and artificial intelligence (AI). It begins in the mid-20th century, when the concept of “machine learning” was still nascent, and computational power was a fraction of what we have today. One of the earliest milestones in machine learning was in 1950, when Alan Turing introduced the Turing Test as a criterion of intelligence, proposing that a machine could be considered intelligent if it could mimic human responses under specific conditions. This period also saw Arthur Samuel define machine learning in 1959 as a “Field of study that gives computers the ability to learn without being explicitly programmed,” showcasing his work on a checkers-playing program that improved with experience.
The 1960s and 1970s witnessed the advent of fundamental algorithms that are still in use today. For instance, the development of the perceptron by Frank Rosenblatt in 1957 laid the groundwork for neural networks and deep learning. However, the initial enthusiasm for AI and machine learning faced significant challenges during the “AI winters” of the late 1970s and late 1980s, when expectations did not meet reality, leading to a decrease in funding and interest in the field. Despite these setbacks, the foundation for future developments was laid during this time, with important theoretical work on algorithms and the limits of computation being carried out.
The resurgence of interest in machine learning began in the late 1990s and early 2000s, fueled by the increasing availability of digital data and significant improvements in computing power. This era saw the development of Support Vector Machines, decision trees, and ensemble methods like Random Forests, which provided powerful tools for data analysis and prediction. The real game-changer, however, came with the advent of deep learning in the 2010s, where neural networks, particularly those with many layers (deep neural networks), achieved remarkable success in tasks such as image and speech recognition. The victory of DeepMind’s AlphaGo over world champion Go player Lee Sedol in 2016 marked a watershed moment, demonstrating the potential of deep learning and machine learning not just as academic pursuits but as technologies with significant real-world impact.
Today, machine learning is an integral part of everyday life, powering search engines, recommendation systems, voice assistants, and more, reflecting the dramatic evolution of the field from its humble beginnings to its current status as a cornerstone of modern artificial intelligence.
Domains of ML, how can it be divided?
- Supervised
- Unsupervised
- RL
Reality is more complex – the different domains are often mixed.