A brief history of Artificial Intelligence

“And now, O Muses, tell me who these people were and from what land they came.”
– Herodotus –

Maurício Pinheiro

Welcome to our brief history of Artificial Intelligence! We will explore its history from its early beginnings in the 19th century, highlighting important milestones and significant achievements. Our goal is to provide an understanding of the evolution of AI and its current role in our lives.

19th Century:

Ada Lovelace wrote about the possibility of machines performing tasks beyond numerical calculation. Her collaboration with Charles Babbage on the design of the Analytical Engine made her one of the first individuals to conceive the idea of a programmable computer.

The philosopher and writer Samuel Butler speculated about the possibility of machines becoming intelligent and even surpassing human intelligence. He addressed these questions in his book “Darwin Among the Machines” (1863), exploring the evolution and relationship between humans and machines.

1943: Warren McCulloch and Walter Pitts published “A Logical Calculus of Ideas Immanent in Nervous Activity,” which established the mathematical foundations for the field of artificial neural networks (ANNs). In this work, they used electrical circuits to model a simple artificial neural network inspired by the structure and function of the human brain and introduced algorithms known as threshold logic.

1949: In his book “The Organization of Behavior,” Donald Hebb presents a work that describes a fundamental concept for how humans learn. He proposes that neural pathways are strengthened each time they are used. Hebb argues that when two neurons fire simultaneously, the connection between them is enhanced. This principle, known as Hebb’s rule or “cells that fire together, wire together,” describes how coordinated neural activity contributes to the strengthening of synapses between neurons. This means that when a particular association or pattern of activity occurs repeatedly, the synaptic connections involved in that activity become more effective, facilitating the transmission of signals between those neurons.

The standard interpretation of the Turing Test involves player C, the interrogator, who is tasked with trying to determine which player – A or B – is a computer and which is a human. The interrogator is limited to using the written responses to make the determination. By Juan Alberto Sánchez Margallo. Source: Wikimedia Commons.

1950: Alan Turing publishes the article “Computing Machinery and Intelligence,” in which he introduces the Turing Test, a benchmark for determining whether a machine can exhibit intelligent behavior equivalent to or indistinguishable from that of a human. The test involves a human evaluator engaging in natural language conversations with both a human and a machine, without knowing which is which. If the evaluator cannot distinguish the machine from the human, it is said that the machine has passed the Turing Test.

1956: The term “artificial intelligence” is coined at the Dartmouth Conference, where researchers gather to discuss the possibility of creating machines that can think and learn. This marks the beginning of AI as a formal academic discipline.

Marvin Minsky, Claude Shannon, Ray Solomonoff, and other scientists at the Dartmouth Summer Research Project on Artificial Intelligence (Photo: Margaret Minsky).

1957: Frank Rosenblatt, a neurobiologist, began working on the Perceptron, an artificial analogue of a neuron. He was intrigued by the functioning of a fly’s eye. Much of the processing done by a fly when deciding to flee is performed in its eyes. The Perceptron, resulting from this research, was built in hardware and is the oldest neural network still in use today. A single-layer Perceptron was useful for classifying a set of continuous-valued inputs into one of two classes.

The Rosenblatt’s Perceptron, as described by Mitchell in “Machine Learning,” page 87. October 24, 2012. Source: Wikimedia Commons.

1959: Arthur Samuel develops the first self-learning computer program, a checkers game system that improves its performance through experience. The program is capable of learning from its own mistakes and adjusting its strategy accordingly. This marks an early example of machine learning, a subfield of AI that involves using algorithms to enhance the performance of a system based on data and experience.

1966: John von Neumann explored the concept of self-reproduction and autonomous machines in his work “Theory of Self-Reproducing Automata” (1966).

GNU Emacs 21 mostrando ELIZA. Fonte: Wikimedia Commons (Dez. 2022).

1966: Joseph Weizenbaum creates ELIZA, the first natural language processing (NLP) program capable of engaging in simple conversations with humans. ELIZA uses pattern matching and substitution rules to generate responses to user inputs, giving the impression of understanding and intelligence. This marks an early example of NLP, a subfield of AI that involves the development of systems capable of understanding, interpreting, and generating human language.

1969: The first internet connection is established on October 29, linking the first two nodes of what would become ARPANET. Leonard Kleinrock’s Network Measurement Center, located at the UCLA School of Engineering and Applied Science, is connected to Douglas Engelbart’s NLS system at SRI International (SRI) in Menlo Park, California. This historic milestone symbolizes the beginning of the network that would become the foundation of the modern internet and plays a crucial role in the development of distributed artificial intelligence systems.

1972: The first expert system, called MYCIN, is developed at Stanford University. It is capable of diagnosing and recommending treatment for blood infections using a set of rules and knowledge provided by human experts. Expert systems are designed to mimic the decision-making abilities of a human expert in a specific domain and are an early example of AI applications in the healthcare field.

1980: Kunihiko Fukushima proposes Neoconitron, a deep convolutional neural network (CNN), which is used for handwriting recognition and other pattern recognition problems. The Neocognitron was inspired by a model proposed by Hubel & Wiesel in 1959. They discovered two types of cells in the primary visual cortex called simple cells and complex cells, and also proposed a cascaded model of these two types of cells for use in pattern recognition tasks.

1980: John Hopfield proposes a new model of neural network. Hopfield’s neural network model is a recurrent network capable of storing and retrieving associative memory patterns. It consists of interconnected neurons and uses an energy function to achieve stable states corresponding to stored patterns. This network has the ability to retrieve complete information from incomplete stimuli and is applied in pattern recognition and optimization problems.

1984: Judea Pearl publishes a seminal book titled “Probabilistic Reasoning in Intelligent Systems,” which has had a significant impact on the field of Artificial Intelligence. This work established the theoretical foundations of Bayesian inference and Bayesian networks in AI. The book introduced key concepts, such as the use of probabilities to represent and reason about uncertainties in intelligent systems. Bayesian inference is based on Bayes’ theorem to update probabilities as new evidence is obtained. Bayesian networks, on the other hand, are graphical models that allow representing and inferring probabilistic relationships between different variables in a system. Judea Pearl’s work was instrumental in advancing probabilistic reasoning techniques in AI, providing a solid foundation for modeling and decision-making in scenarios with uncertainties. Since then, Bayesian inference and Bayesian networks have been widely used in various domains of AI, contributing to the development of more efficient and reliable intelligent systems.

1986: David Rumelhart, Geoffrey Hinton, and Ronald Williams published the seminal paper titled “Learning representations by back-propagating errors,” which had a profound impact on the field of neural networks. In this paper, they proposed and popularized the method of backpropagation, also known as the error backpropagation algorithm, for training multilayer neural networks. This training method allowed neural networks to learn representations of complex data, enabling the solving of challenging pattern recognition and information processing problems. Backpropagation involves calculating the gradient of the error with respect to the network’s weights and then updating those weights using optimization techniques. Through the efficient use of this algorithm, Rumelhart, Hinton, and Williams demonstrated that multilayer neural networks could learn and generalize from large datasets more effectively than traditional methods at the time. This work was a fundamental milestone in the AI renaissance and sparked renewed interest in the development of deep neural networks and their applications in various fields, driving the advancement of machine learning and artificial intelligence.

1989: Yann LeCun, Leon Bottou, Yoshua Bengio, and Patrick Haffner developed LeNet-5, a groundbreaking convolutional neural network specifically designed for handwritten digit recognition. Released in 1998, LeNet-5 was a significant milestone in the field of computer vision and laid the foundation for many modern architectures of convolutional neural networks. The LeNet-5 architecture consists of convolutional layers, pooling layers, and fully connected layers, all working together to extract discriminative features from images of handwritten digits. With training on large datasets, LeNet-5 achieved impressive results in digit recognition, setting a new benchmark for accuracy and performance in optical character recognition (OCR) systems. The success of LeNet-5 paved the way for the application of convolutional neural networks in a wide range of computer vision tasks, including image classification, object detection, and semantic segmentation. LeNet-5 is considered a significant milestone in the history of AI and a starting point for the development of more advanced architectures of convolutional neural networks.

1980s and 1990s: The AI winter refers to a period of reduced funding and interest in AI research. The term was coined in 1984 to describe a cycle of enthusiasm followed by disappointment and criticism, resulting in funding cuts and a slowdown in research in the field. During this period, exaggerated expectations, inflated promises, and intense media hype contributed to the decline of interest in AI. However, despite the ups and downs in the reputation of AI, the field continued to develop successful new technologies. Since the decline of AI in the 1990s, there has been a gradual increase in enthusiasm and optimism, culminating in the current boom in the field, particularly in the field of machine learning. The main causes of this “AI winter” include technical limitations related to hardware and the availability of data for training AI systems.

Deep Blue, a computer similar to this one, defeated world chess champion Garry Kasparov in May 1997. It was the first computer to win a match against a world champion. Photo taken at the Computer History Museum. Source: Wikimedia Commons.

1997: IBM’s Deep Blue defeated the world chess champion, Garry Kasparov, marking a significant milestone in the development of artificial intelligence. Deep Blue employed advanced search and evaluation techniques, combined with powerful hardware, to analyze and evaluate millions of positions per second, surpassing its human opponents. This achievement represents a significant advancement in the application of AI in the field of games, demonstrating the potential of AI systems to outperform human experts in strategic and complex tasks. Deep Blue’s victory over Kasparov captured the world’s attention to the emerging capabilities of artificial intelligence and spurred further research and advancements in AI algorithms and techniques applied to games and other challenging domains.

Sample images from the MNIST test dataset. The image contains handwritten numbers from 0 to 9. The size of all the small images is 28×28 pixels. The image can be used to create a number detection program using machine learning or neural networks. By Suvanjanprasai. January 2, 1998. Source: Wikimedia Commons.

1999: A milestone in AI is the creation of the MNIST database. The MNIST database, which stands for “Modified National Institute of Standards and Technology,” is an extensive collection of handwritten digits widely used for training and testing machine learning algorithms, particularly in the field of image recognition and classification. The database consists of thousands of labeled examples of handwritten digits, providing a standardized benchmark for evaluating the performance of AI models in recognizing and interpreting handwritten characters. The MNIST database has made significant contributions to the advancement of AI and has been a fundamental resource for the development and benchmarking of various image processing and pattern recognition algorithms.

Early 2000s: The term “deep learning” starts gaining popularity after an article by Geoffrey Hinton and Ruslan Salakhutdinov demonstrates how a multi-layer neural network could be pretrained one layer at a time.

2009: The NIPS Workshop on Deep Learning for Speech Recognition takes place, and it is discovered that with a sufficiently large dataset, neural networks do not need pretraining, and error rates significantly decrease.

2011: Apple introduces Siri, a virtual assistant that uses natural language processing to answer questions and perform tasks. Siri is capable of understanding and responding to spoken requests, such as setting reminders or providing directions, and is widely seen as a significant step in the development of practical natural language processing (NLP) systems.

2012: Artificial pattern recognition algorithms achieve human-level performance in certain tasks. Additionally, Google’s deep learning algorithm is able to identify cats.

2014: Google acquires the UK-based artificial intelligence startup DeepMind for £400m.

2014: Google DeepMind’s AlphaGo defeats world Go champion Lee Sedol, demonstrating the capabilities of machine learning algorithms in complex and strategic situations. Go is a board game with a large number of possible moves, making it difficult for humans to evaluate and analyze all possibilities. AlphaGo uses a combination of deep learning techniques and Monte Carlo tree search to evaluate and select its moves, leading to its remarkable victory over the human champion. This marks a significant milestone in the development of AI in the gaming field, showcasing the potential of machine learning algorithms to surpass human experts in complex tasks.

2015: Facebook deploys deep learning technology, called DeepFace, to automatically tag and identify Facebook users in photographs. Algorithms perform superior facial recognition tasks using deep networks that take into account 120 million parameters.

2016: AlphaGo Zero, a variant of AlphaGo, defeats the original version without receiving any human-generated data. AlphaGo Zero is able to learn to play Go at a superhuman level by playing against itself using a self-play reinforcement learning algorithm. This marks a significant milestone in the development of autonomous learning AI systems.

2017: OpenAI’s machine learning model, GPT-2, is capable of generating human-like text, raising concerns about the potential malicious use of AI for generating fake news. GPT-2 is trained on a large dataset of human-generated text and can generate coherent and credible paragraphs as well as complete articles on a wide range of topics. This marks an important milestone in the development of AI in the field of natural language processing, demonstrating the potential of machine learning algorithms to generate human-like text.

2017: Mass adoption of deep learning in various corporate applications, along with advancements in research. All technology events related to data science, AI, and big data point to deep learning as the leading technology for creating intelligent systems.

2019: GPT-3 (Generative Pre-trained Transformer 3) by OpenAI becomes the most powerful and large-scale language model to date, with 175 billion parameters. It is capable of performing a wide range of language tasks, including translation, summarization, and question-answering, with impressive accuracy and fluency. GPT-3 is seen as a major breakthrough in the field of natural language processing and machine learning.

ChatGPT. Source: Wikimedia Commons (Dez. 2022).

2020: Research and development in AI continue to advance rapidly, with new applications and technologies emerging in fields such as healthcare, finance, and transportation. AI systems are also being used to address social challenges, such as climate change and the COVID-19 pandemic. However, concerns about the potential impacts of AI on society, such as job displacement and bias in decision-making, remain a topic of debate and research.

December 2022: AI-Talks was born.


#AI #ArtificialIntelligence #EvolutionofAI #HistoryofAI #Technology #Innovation #TechnologicalAdvancements #Intelligence #MachineLearning #DeepLearning #NeuralNetworks #Automation #Future #DigitalTransformation #AIinToday’sWorld #TechnologicalDevelopment #AIforEveryone #AIforLife #TechnologyoftheFuture



Copyright 2024 AI-Talks.org

Leave a Reply

Your email address will not be published. Required fields are marked *