Varol Cagdas Tok

Personal notes and articles.

Introduction to Machine Learning

Machine learning has become a central technology. The core ideas behind it are rooted in centuries of inquiry into knowledge, intelligence, and learning. We provide an introduction to the field, placing it in the context of other mechanisms for creating complex systems and tracing its lineage from philosophy to artificial intelligence.

Three Paradigms of System Improvement

To understand what makes learning special, consider it as one of three mechanisms for the design and improvement of complex, autonomous systems:

  1. Evolution: A process of improvement through trial and error, described by Richard Dawkins as the "Blind Watchmaker." In nature, it operates through genetic mutation and natural selection over timescales. In technology, it can be seen in the advancement of the state-of-the-art through competition and innovation. It is a self-optimizing process, but it is slow, wasteful, and not directed by foresight.
    1. Explicit Design: The top-down approach where a knowledgeable creator—an engineer, a programmer, or a designer—builds a system based on explicit knowledge and analytical thinking. This is the foundation of most traditional engineering and software development. Its advantage is that the resulting system is understood and can be analyzed and improved methodically. Its disadvantage is the need for an expert designer.
      1. Learning: This mechanism occupies a middle ground. It is an optimization process that happens within the lifetime of an individual system, allowing it to improve its behavior through interaction with its environment or by analyzing data. Machine learning is the technical embodiment of this principle, aiming to create systems that can improve themselves autonomously.
      2. Philosophical Roots: Rationalism vs. Empiricism

        The intellectual tension that gave rise to machine learning can be traced back to a debate in epistemology, the theory of knowledge: how do we acquire knowledge?

        • Rationalism: Championed by philosophers like Plato and René Descartes, rationalism emphasizes deductive reasoning from general principles. It is a top-down approach that posits we can derive knowledge from a set of foundational axioms using logic. This philosophy is the intellectual ancestor of classical Artificial Intelligence, which focused on creating intelligent systems by programming them with explicit, symbolic rules and using logical inference to reason about the world.
        • Empiricism: Advocated by thinkers like Aristotle and John Locke, empiricism argues that knowledge is derived primarily from sensory experience and observation. It is a bottom-up approach, suggesting the mind begins as a "blank slate" (tabula rasa) and builds knowledge by generalizing from data. This philosophy is the direct intellectual forerunner of machine learning, which focuses on inducing general patterns and models from specific examples.

        A Brief History of Machine Learning

        The journey of machine learning has been marked by several phases, often swinging between the poles of rule-based and data-driven approaches.

        1. The Statistical Origins (Pre-1950s): The mathematical foundations were laid by statisticians. Thomas Bayes developed a theorem for updating beliefs based on evidence, which is the cornerstone of Bayesian statistics. Figures like Karl Pearson and Ronald Fisher established the foundations of frequentist statistics, developing concepts like correlation and hypothesis testing.
          1. Neural Computation I: The Perceptron (1950s-1960s): The first wave of interest in learning machines was inspired by neuroscience. The McCulloch-Pitts neuron (1943) was a mathematical model of a biological neuron. This led to Frank Rosenblatt's Perceptron (1958), a learning algorithm that could learn to classify patterns. However, the publication of the book Perceptrons (1969) by Marvin Minsky and Seymour Papert, which highlighted the model's limitations (e.g., its inability to solve the XOR problem), led to a decline in funding and interest, ushering in the first "AI winter."
            1. The Rise of Classical AI (1960s-1980s): With neural network research on the wane, the focus shifted to symbolic, logic-based AI. This era was dominated by the development of expert systems, which aimed to capture the knowledge of human experts in a set of formal rules. The belief was that intelligence could be achieved by creating vast knowledge bases and powerful inference engines.
              1. Neural Computation II: The Return of Connectionism (1980s): The limitations of purely symbolic systems became apparent, and interest in neural networks was rekindled. The breakthrough was the popularization of the backpropagation algorithm, which provided an efficient way to train Multilayer Perceptrons (MLPs)—networks with one or more hidden layers. These models could overcome the limitations of the original Perceptron and learn complex, non-linear functions, leading to a resurgence of the field.
                1. The Rise of Statistical Machine Learning (1990s-2000s): In the 1990s, the field underwent another shift, moving away from the biologically inspired heuristics of neural networks towards more mathematically rigorous, statistically grounded models. This era was dominated by the development of Support Vector Machines (SVMs), kernel methods, and graphical models. These methods had strong theoretical foundations (e.g., VC-theory) and often provided performance with fewer parameters to tune, making them the state-of-the-art for many years.
                  1. Neural Computation III: The Deep Learning Revolution (2010s-Present): The current era began around 2010, driven by the convergence of three factors: the availability of massive datasets, significant increases in computational power (especially from GPUs), and algorithmic improvements. Deep learning—the use of neural networks with many layers—achieved breakthrough performance on benchmark problems in computer vision, speech recognition, and natural language processing.
                    1. The Age of Generative AI and Foundation Models (2020s-Present): The latest evolution is the rise of massive, pre-trained models, often called foundation models. Models like GPT are trained on large amounts of unlabeled text and image data from the internet in a self-supervised manner. They learn general-purpose representations of the world that can then be fine-tuned for a variety of specific tasks.
                    2. Conclusion

                      Machine learning is the application of the empiricist tradition, offering a paradigm for building intelligent systems that learn from experience. Its trajectory encompasses a range of competing ideas, shifts in research focus, and advancements influenced by theory, data, and computational resources. From its philosophical origins to the foundation models today, the development of machines capable of learning remains an area of research and technological innovation.