Technologyfeatured

[All About AI] The Origins, Evolution & Future of AI

By October 14, 2024 No Comments
AI has revolutionized people’s lives. For those who want to gain a deeper understanding of AI and use the technology, the SK hynix Newsroom has created the All About AI series. This first episode covers the historical evolution of AI and explains how it became integrated into today’s world.

 

AI-powered robots that walk, talk, and think like humans have long been a staple of sci-fi comics and movies. However, AI and robotics are no longer merely works of fiction—they have become a reality. Now that AI is here and transforming people’s lives, it is prudent to look back and consider AI’s origins, the milestones which have shaped the technology’s evolution, and consider what the future might hold.

From the Turing Test to Machine Learning: AI’s Early Beginnings

An overview of AI’s evolution through the decades from the 1950s to the 2020s

Figure 1. An overview of AI’s evolution through the decades from the 1950s to the 2020s

 

The birth of AI can be traced back to the 1950s. In 1950, British mathematician Alan Turing proposed that machines could “think,” introducing what is now known as the “Turing test” to evaluate this capability. This is widely recognized to be the first study to present the concept of AI. In 1956, the Dartmouth Summer Research Project on Artificial Intelligence formally introduced the term “AI” to the wider world for the first time. Held in the U.S. state of New Hampshire, the conference fueled further debates on whether machines could learn and evolve like humans.

During the same decade, the development of artificial neural network1 models marked a significant milestone in computing history. In 1957, U.S. neuropsychologist Frank Rosenblatt introduced the “perceptron” model2, empirically demonstrating that computers can learn and recognize patterns. This practical application built on the “neural network theory” developed in 1943 by neurophysiologists Warren McCulloch and Walter Pitts, who conceptualized nerve cell interactions into a simple computational model. Despite these early breakthroughs raising high expectations, research in the field soon stagnated due to limitations in computing power, logical framework, and data availability.

1Neural network: A machine learning program, or model, that makes decisions in a manner similar to the human brain. It creates an adaptive system to make decisions and learn from mistakes.
2Perceptron: The simplest form of a neural network. It is a model of a single neuron that can be used for binary classification problems, enabling it to determine whether an input belongs to one class or another.

Then in the 1980s, “expert system” emerged which operated solely based on human-defined rules. These systems could make automated decisions to perform tasks such as diagnosis, categorization, and analysis in practical fields such as medicine, law, and retail. However, during this period, expert systems were limited by their reliance on rules set by humans and struggled to understand the complexities of the real world.

In the 1990s, AI evolved from following human commands to autonomously learning and discovering new rules by adopting machine learning algorithms. This became possible due to the advent of digital technology and the internet, which provided access to vast amounts of online data. At this point, AI was able to unearth new rules even humans could not discover. This period marked the start of renewed momentum for AI research, based on machine learning.

The Rise of Deep Learning: A Key Technology in AI’s Growth

Timeline showing advances in artificial neural networks and deep learning

Figure 2. Timeline showing advances in artificial neural networks and deep learning

 

While the 1990s presented opportunities for AI to grow, the journey and evolution of AI has had its share of setbacks. In 1969, early artificial neural network research hit a roadblock when it was discovered that the perceptron model could not solve nonlinear problems3, leading to a prolonged downturn in the field. However, computer scientist Geoffrey Hinton, often hailed as the “godfather of deep learning,” breathed new life into artificial neural network research with his groundbreaking ideas.

For example, in 1986, Hinton applied the backpropagation4 algorithm to a “multilayer perceptron” model, essentially layers of artificial neural networks, proving it could address the limitations of the initial perceptron model. This seemed to spark a revival in artificial neural networks research, but as the depth of the networks increased, issues began to emerge in the learning process and outcomes.

In 2006, Hinton introduced the “deep belief network (DBN),” which enhanced the performance of a multilayer perceptron, in his paper “A Fast Learning Algorithm for Deep Belief Nets.” By pre-training each layer through unsupervised learning5 and then fine-tuning the entire network, the DBN significantly improved the speed and efficiency of neural network learning—which had previously been deemed an issue. This progress paved the way for future advancements in deep learning.

3The initial perceptron model was a single-layer perceptron that could not solve nonlinear problems such as the XOR problem, which involves two input values; it outputs 0 if the two input values ​​are the same and 1 if they are different.
4Backpropagation: An algorithm used in neural networks to minimize errors by adjusting the weights. It works by calculating the difference between the predicted and actual values and then updating the weights in reverse order, starting from the output layer.
5Unsupervised Learning: A type of machine learning where the model is trained on input data without explicit labels or predefined outcomes. The goal is to discover and understand hidden structures and patterns within the data.

In 2012, deep learning made a historic leap forward when Hinton’s team won the ImageNet Large Scale Visual Recognition Challenge (ILSVRC) with their deep learning-based model, AlexNet. This triumph demonstrated deep learning’s immense power by recording an error rate of just 16.4%, surpassing the 25.8% of the previous year’s winner.

An Overview of ILSVRC’s Image Recognition Error Rate by Year (Kien Nguyen, Arun Ross, Iris Recognition With Off-the-Shelf CNN Features: A Deep Learning Perspective, IEEE Access, Sept. 2017 p.3)

Figure 3. An Overview of ILSVRC’s Image Recognition Error Rate by Year (Kien Nguyen, Arun Ross, Iris Recognition With Off-the-Shelf CNN Features: A Deep Learning Perspective, IEEE Access, Sept. 2017 p.3)

 

Deep learning, a focal point of AI research, has grown rapidly since the 2010s for two primary reasons. First, advances in computer systems, including graphics processing units (GPUs), have driven AI development. Originally designed for graphics processing, GPUs can process repetitive and similar tasks in parallel. This capability enables GPUs to process data faster than central processing units (CPUs). In the 2010s, general-purpose computing on GPUs (GPGPU) emerged, enabling GPUs to be used for broader computational tasks beyond graphics rendering and allowing them to replace CPUs in some instances. The use of GPUs has further increased as they have been utilized for training artificial neural networks, accelerating the development of deep learning. Deep learning, which needs to perform iterative computations during analysis of large datasets to extract features, benefits from the parallel processing capability of GPUs.

Second, the expansion of data resources has fueled progress in deep learning. Training an artificial neural network requires vast amounts of data. In the past, data was primarily sourced from users manually inputting information into computers. However, the explosion of the internet and search engines in the 1990s exponentially increased the range of data available for processing. In the 2000s, the advent of technologies such as smartphones and the Internet of Things (IoT) contributed to the birth of the Big Data era, where real-time information flows from every corner of the globe. Deep learning algorithms use this large quantity of data for training, growing increasingly sophisticated. This data revolution has therefore set the stage for significant advancements in deep learning technology.

Figure 4. Google DeepMind’s AlphaGo – The Movie is a documentary film about the epic battle between AlphaGo and Lee Sedol on March 9, 2016

 

By 2016, the evolution of AI reached a dramatic turning point with the development of AlphaGo, an advanced AI program created by Google DeepMind to play the board game Go. This extraordinary AI program captivated the world when it defeated Go grandmaster Lee Sedol by an impressive 4-1 score. Combining deep learning with reinforcement learning6 and Monte Carlo tree search (MCTS)7 algorithms, AlphaGo learned to mimic human intuition, predict moves, and strategize through tens of thousands of self-played games. AlphaGo’s victory over a legendary human player signaled the beginning of a new AI era.

6Reinforcement Learning: A type of machine learning where an AI agent learns to make decisions by interacting with an environment. The agent receives rewards or penalties based on its actions and aims to maximize cumulative rewards over time by optimizing its strategy.
7Monte Carlo tree search (MCTS): A stochastic algorithm that repeatedly generates a series of random numbers to derive a numerical approximation of a function’s value. It structures the possible actions of the current situation into a search tree and uses random simulations to infer the pros and cons of each, ultimately determining the optimal course of action.

ChatGPT: The Catalyst for the Generative AI Boom

Generative AI explained through key AI subsets

Figure 5. Generative AI explained through key AI subsets

 

At the close of 2022, humanity stood on the brink of a transformative leap with AI technology. OpenAI unveiled ChatGPT, powered by a type of LLM8 known as generative pre-trained transformer (GPT) 3.5, marking the dawn of the generative AI era. Most notably, this leap propelled AI into the creative realm, a domain once considered uniquely human. Now, generative AI can produce high-quality content across diverse formats, moving beyond traditional deep learning, which merely predicts or classifies data. Instead, generative AI, using LLMs or various image generation models such as variational autoencoders (VAEs), generative adversarial networks (GANs), and diffusion models, creates original results tailored to user needs.

8Large language model (LLM): Deep learning algorithms that perform a variety of natural language processing tasks by leveraging extensive data.

To provide a clearer context for the evolution of generative AI, it is essential to examine its origins and key developments. The roots of generative AI trace back to 2014, when American scientist and researcher Ian Goodfellow introduced GANs. In GANs, two neural networks engage in a continuous duel: one generates new data from a dataset, while the other network compares this new data to the original dataset to determine its authenticity. Through this iterative process, GANs produce increasingly refined and sophisticated outputs. Over time, researchers have enhanced and expanded upon this model, leading to its widespread use in applications such as image generation and transformation.

In 2017, the natural language processing (NLP)9 model “transformer” was introduced. This model considers the relationships between data as key variables. By focusing more attention to certain information, transformers can learn complex data patterns and relationships between data, capturing essential details to produce higher quality results. This advancement transformed NLP tasks such as language comprehension, machine translation, and conversational systems, leading to the development of LLMs such as the aforementioned GPT.

9Natural language processing (NLP): A subfield of AI that uses algorithms to analyze and process natural language data. By examining syntactic structures, semantic relationships, and contextual patterns, NLP systems can perform tasks such as language translation.

First released in 2018, GPTs have rapidly advanced in performance by expanding their parameters and training on data every year. By 2022, OpenAI’s chatbot ChatGPT, powered by GPT-3.5, completely changed the paradigm of AI. ChatGPT, with its exceptional ability to understand user context, deliver relevant responses, and handle diverse queries, quickly gained traction. Within a week of its launch, it drew over 1 million users and attracted more than 100 million active users within two months.

The rapid advancements in AI culminated in a major technological leap forward in 2023 with the launch of GPT-4 by OpenAI. This new model is built on a dataset roughly 500 times larger than that of GPT-3.5. GPT-4, now considered a Large Multimodal Model (LMM)10, can simultaneously process diverse formats of input data, including images, audio, and video, expanding far beyond its text-only predecessors. In 2024, OpenAI introduced GPT-4o, an enhanced model offering faster, more efficient processing of text, voice, and images. Capitalizing on the generative AI boom triggered by ChatGPT, companies have rolled out diverse services. For example, Google’s Gemini can simultaneously recognize and understand text, images, and audio; Meta’s SAM accurately identifies and isolates objects in images; and OpenAI’s Sora generates videos from text prompts.

10Large Multimodal Model (LMM): A deep learning algorithm that can handle many types of data, including images, audio, and more, not just text.

The generative AI market is only beginning to unleash its potential. According to a report from the global market research firm International Data Corporation (IDC), the market is set to be worth 40.1 billion USD in 2024—2.7 times larger than the previous year. Looking ahead, the market is expected to continue its growth each year and reach 151.1 billion USD by 2027. As generative AI evolves, its influence will extend beyond software to various formats including hardware and internet services. The world can expect a leap in capabilities and a push towards greater accessibility, making cutting-edge AI technology available to an ever-growing audience.

AI’s Impact on Revolutionizing Today and Redefining Tomorrow

Just as Google search revolutionized the early 2000s and mobile social media reshaped the 2010s, AI is now driving transformative changes across society. The pace of this technological advancement is unprecedented, and the challenges and concerns of humanity are growing along with it.

So what is the “next generative AI”? The most notable technology around today is perhaps on-device AI. Unlike traditional AI that relies on large cloud servers to pull data to edge devices, on-device AI operates directly on electronic devices such as smartphones and PCs through integrated AI chipsets and smaller LLMs (sLLMs). This shift promises to enhance security, conserve resources, and deliver more personalized AI experiences.

Cloud-based AI vs on-device AI structures

Figure 6. Cloud-based AI vs on-device AI structures

 

AI will seamlessly integrate into an increasing number of devices, continuously evolving in form and function. Thus, innovations that once seemed like science fiction are becoming reality. For instance, in 2023, U.S. startup Humane launched the AI Pin, a wearable device with a laser-ink display that projects a menu onto the user’s palm. At CES 2024, Rabbit’s R1 and Brilliant Labs’ Frame showcased their own cutting-edge wearable AI technology. Meanwhile, mixed reality (MR) headsets, like Apple’s Vision Pro and Meta’s Quest, are pushing beyond traditional virtual reality (VR) and metaverse experiences, opening up new markets.

However, as technology races forward, it not only creates new opportunities but also brings about social challenges. The rapid rise of AI has sparked concerns about society’s ability to keep up with these advancements. In particular, AI’s potential misuse and impact on real-world issues has heightened these fears. Sophisticated AI-generated content, such as deepfake videos and manipulated images, creates fake news and disrupts society. Recently, concerns about fake content have intensified in many countries ahead of major elections, including the 2024 U.S. presidential election.

Social anxiety and disruption due to deepfake technology portrayed by DALL-E, a generative AI platform

Figure 7. Social anxiety and disruption due to deepfake technology portrayed by DALL-E, a generative AI platform

 

There are also risks associated with the development and use of AI. As generative AI crawls and merges publicly available web contents to train its AI models, there are concerns about plagiarism. Moreover, copyright disputes can arise from creating content using similar prompts with the same generative AI program. The potential for AI to shift from enhancing productivity to replacing jobs and disrupting the labor market presents a troubling reality for some as well.

AI has created a world beyond human imagination. As this new world unfolds, it is crucial to prepare for the changes ahead. Addressing this new era involves thoughtful planning and social discussion. These action items first require a deep understanding of AI’s potential and implications, which will be provided throughout the All About AI series.