What does Artificial Intelligence mean?
According to data from Forbes, the term "Artificial Intelligence" was coined by John McCarthy in 1956 when he brought together a team of researchers in different disciplinary areas such as language simulation, neural networks, and complexity theory.
As Britannica indicates, the concept of Artificial Intelligence is usually applied "to the project of developing systems endowed with the intellectual processes characteristics of humans," some examples are the ability to reason, constant curiosity, and the discovery and reassessment of meaning, among other features.

What does Intelligence mean?
The definition of intelligence, according to Oxford data, is "the ability to learn, understand and think logically about things, and the ability to do this well."
The dictionary defines it as "the capacity for learning, reasoning, comprehension, and similar forms of mental activity, and the ability to grasp the truth, relate facts, and understand meaning."
At first glance, it might seem like an acquired capacity, learned through human interactions, language, and the transmission of knowledge. But on the other hand, we must also point out the importance of the cognitive characteristics of the human brain. This leaves scientists creating machines in a bind, as they must simulate the thought process.

Where does Artificial Intelligence come from?

1950 - 1970
The early stage of AI was focused on developing neural networks, which are mathematical models inspired by the structure and function of the human brain. Researchers started exploring the idea of creating machines that could think and make decisions like humans, which led to the development of neural networks.

1980 - 2010
Machine learning: In this stage, the focus shifted from neural networks to machine learning, which is a type of AI that enables computers to learn and improve their performance based on data. Machine learning algorithms use statistical techniques to identify patterns in data and make predictions or decisions based on that data.

2020 - 2023
Deep learning. This is the latest stage in the evolution of AI where the focus is on developing deep learning algorithms, which are a type of machine learning that uses neural networks with multiple layers. The goal of deep learning is to create AI systems that have a higher level of awareness and the ability to respond to complex situations without being explicitly programmed to do so.
So far, we have seen some definitions of intelligence, and simply the functionality of the mind under some concepts of reasoning, which leads us to conclude that intelligence is a human trait until we consider language and learning, which science has been able to replicate through codes and software.
The evolution of artificial intelligence (AI) can be divided into three stages according to data from SAS. These stages are as follows:
Mind Creation
A 2020 publication by the National Center for Biotechnology Information for the journal Dialogues Clin Neurosci defines intelligence as "a general mental capacity to reason, solve problems, and learn. Due to its general nature, intelligence integrates cognitive functions such as perception, attention, memory, language, or planning..."
The article highlights widespread individual differences in the ability to reason, problem-solve, and learn, "leading to differences among humans in their overall ability to cope with challenging situations."
The evolution of the study of thought, intelligence, and the functioning of the mind has been supported by neuroscience and neuropsychology.
In the article "Areas of Research in Neurosciences" from New York University Grossman School of Medicine, it is stated that there are approximately 90 billion neurons in the human brain. Each neuron is connected to many others, and its function is like a network to react to information from the outside world.
On the other hand, neuropsychology is another pillar of the study of the mind. In his article "Neuropsychology: Theoretical Basis," Giovanni Berlucchi defines neuropsychology as the study of the relationships between the brain, on the one hand, and the mind and behavior control, on the other.

The pioneers of Artificial Intelligence: Alan Turing and John McCarthy
Alan Turing, a mathematician, logician, theoretical computer scientist, cryptographer, philosopher, and British theoretical biologist, is widely considered one of the fathers of computer science and a pioneer in the field of artificial intelligence (AI). During World War II, Turing designed the bombe machine which helped speed up the decoding of messages.
In 1950, Turing wrote a paper entitled "Computing Machinery and Intelligence," in which he speculated on the concept of "thinking machines" and proposed the idea of the Turing Test, a measure of a machine's ability to exhibit intelligent behavior that is indistinguishable from that of a human. This paper is considered one of the first discussions of AI and remains an important reference in the field today.
American mathematician, scientist, and computer scientist John McCarthy followed Turing's work and is credited with coining the term "artificial intelligence". McCarthy recognized that the key to AI was automation, which enables machines to solve problems and provide answers automatically without human intervention. He developed the Lisp computer language in 1958, which later became the standard programming language for AI.
At the time, the high cost of computing and the lack of mechanisms to store commands made it difficult to pursue AI research. In the early 1950s, the cost of renting a computer was $200.00 per month, and only large technology companies and prestigious universities had access to the machines.
Years later, Allen Newell, Cliff Shaw, and Herbert Simon created a program called The Logic Theorist, funded by the RAND Corporation, to imitate the problem-solving abilities of a human being. The program was presented at the Dartmouth Summer Research project on AI in 1956 by John McCarthy, who brought together the best researchers in the field, but his goal of reaching an agreement on standard methods for AI was not achieved.


The Evolution of Artificial Intelligence in Literature and Hollywood
For decades, literature, TV, cinema, and radio have convinced us of simplistic theories about the growth of the mind and artificial intelligence.
Based on the premise of protecting humanity, androids or other forms of artificial intelligence decide to exterminate the very humanity that created them, or in less extreme cases, take away human freedom.
It may sound fatalistic in science fiction, but if we review history, we can understand the extent to which humanity has sought "peace and freedom."
Below, we present Hollywood cinema's portrayal of technological advancement over time, aided by literature and science fiction.
COMO Group and all its affiliates (“COMO”) hereby ensure that any and all content provided in this forum is for information purposes only and envisages merely to disseminate knowledge and act as a forum for exchange of ideas. Furthermore, COMO guarantees that this project is strictly non-profit and that all content will be shared on a completely free-access basis. Moreover, any third-party content that may eventually be shared by COMO as part of this forum shall be solely open-source content (creative common) and will not, in any way, infringe any third-party trademarks and/or copyrights. Should you have any issue with any content posted by COMO on this forum, please contact us.
Metróplis (1927)
They can be easily accessed through a digital banking app, and users do not need to carry a physical card with them. Additionally, digital cards may offer enhanced security features, such as the ability to lock or disable the card if it is lost or stolen.
The day the earth stood still (1951)
Did you know: By calling the number *99#, mobile transactions can be carried out without an app and internet connection.
2001: Space Odyssey (1968)
AEPS is a digital payment system that allows users to conduct financial transactions using their Aadhaar number (a unique identification number issued by the Government of India to every individual resident of India) and biometric data.
Colossus: The Forbin Project (1970)
UPI is a digital payment system that allows users to make and receive payments instantly and securely using their mobile phones. It is an open-source platform that is owned and operated by the National Payments Corporation of India (NPCI).
Westworld (1973)
Digital wallets have made it easier for users to make payments and manage their finances by eliminating the need to remember card PINs, CVV details, and carry cash. Instead, users can store their bank account and card credentials in their digital wallet and easily add funds and make payments to other merchants with similar applications.
Star Wars (1977)
Commonly known as e-banking, this refers to obtaining certain banking services over the internet. These services may include fund transfers, opening and closing accounts, and other core banking functions.
Tron (1979)
Commonly known as e-banking, this refers to obtaining certain banking services over the internet. These services may include fund transfers, opening and closing accounts, and other core banking functions.
Blade Runner (1982)
Commonly known as e-banking, this refers to obtaining certain banking services over the internet. These services may include fund transfers, opening and closing accounts, and other core banking functions.
Back to the Future (1985)
Commonly known as e-banking, this refers to obtaining certain banking services over the internet. These services may include fund transfers, opening and closing accounts, and other core banking functions.
Daryl (1985)
Commonly known as e-banking, this refers to obtaining certain banking services over the internet. These services may include fund transfers, opening and closing accounts, and other core banking functions.
Terminator (1984)
Commonly known as e-banking, this refers to obtaining certain banking services over the internet. These services may include fund transfers, opening and closing accounts, and other core banking functions.
Matrix (1999)
Commonly known as e-banking, this refers to obtaining certain banking services over the internet. These services may include fund transfers, opening and closing accounts, and other core banking functions.
A.I. (2001)
They can be easily accessed through a digital banking app, and users do not need to carry a physical card with them. Additionally, digital cards may offer enhanced security features, such as the ability to lock or disable the card if it is lost or stolen.
I. Robot (2004)
Commonly known as e-banking, this refers to obtaining certain banking services over the internet. These services may include fund transfers, opening and closing accounts, and other core banking functions.
Stealth (2005)
Commonly known as e-banking, this refers to obtaining certain banking services over the internet. These services may include fund transfers, opening and closing accounts, and other core banking functions.
Iron Man (2008)
Commonly known as e-banking, this refers to obtaining certain banking services over the internet. These services may include fund transfers, opening and closing accounts, and other core banking functions.
Prometeus (2012)
Commonly known as e-banking, this refers to obtaining certain banking services over the internet. These services may include fund transfers, opening and closing accounts, and other core banking functions.
Her (2013)
Commonly known as e-banking, this refers to obtaining certain banking services over the internet. These services may include fund transfers, opening and closing accounts, and other core banking functions.
Elysium (2013)
Commonly known as e-banking, this refers to obtaining certain banking services over the internet. These services may include fund transfers, opening and closing accounts, and other core banking functions.
Interestellar (2014)
Commonly known as e-banking, this refers to obtaining certain banking services over the internet. These services may include fund transfers, opening and closing accounts, and other core banking functions.
Automata (2014)
Commonly known as e-banking, this refers to obtaining certain banking services over the internet. These services may include fund transfers, opening and closing accounts, and other core banking functions.
Ex Machina (2014)
Commonly known as e-banking, this refers to obtaining certain banking services over the internet. These services may include fund transfers, opening and closing accounts, and other core banking functions.
Exploring the world of AI in film: A journey through time


1957 - 1974
Programmers managed to store more information on computers and with more speed.

1961
Creation of the industrial robot called Unimate by George Devol. It was initially used by General Motors for assembly and handling die-cast parts and welding car bodies.

1964
The first chatbot called Eliza was created by Joseph Wiezenbaum. This psychotherapeutic robot generated pre-designed responses to users to make them feel as though they were in a conversation and understood their requests.

1969
Shakey, the first general-purpose mobile robot, was introduced. It had a television monitor and bars as motion sensors.

1995
Richard Wallace introduced Alice, a chatbot inspired by Eliza. He made improvements, including the implementation of natural language processing, which later served as the basis for the movie "Her."

1997
IBM created Deep Blue, a chess-playing computer. A year later, Kismet was introduced as one of the first robots with the ability to interact socially and emotionally with humans. It had expressions, eyes, head movements, and emotions projecting sadness, anger, tranquility, surprise, or disgust.

2002
Roomba was introduced as a domestic cleaning appliance with filters, long-term mobility, and the ability to avoid obstacles such as furniture and stairs. Amazon released Alexa, a voice-controlled virtual assistant. The same year, Google introduced its own virtual assistant programming.

2005
Stanford University achieved a "futuristic" goal by developing the first car with an automatic police car, which won a robot vehicle competition.

2008
Apple showed its commitment to leading artificial intelligence by introducing the iPhone and Siri. These options provided users with access to a vast amount of information on the Internet and the ability to make calls, send messages, or synchronize information from one device to another.

2011
Siri was introduced as a virtual assistant with a conventional interface, prior knowledge of the user, and a hierarchy of services and requests.

2012
Google created a computer that could learn to identify cats, faces, and human bodies through YouTube.

2014
A computer bot named Eugene Goostamn managed to fool 30 out of 150 judges during a Turing test by making them believe they were interacting with a 13-year-old Ukrainian boy. The same year, Amazon introduced Alexa, a virtual assistant available on various platforms such as smartwatches, car monitors, and TVs.

2015
The AlphaGo machine won against a professional Go player for the first time. The following year, Hanson Robotics introduced Sophia, a better-characterized humanoid robot than Kismet, which was introduced 12 years prior. Sophia could imitate facial expressions, language, and interactions with other human beings.

2017
Amper was introduced as the first artificial intelligence to compose music and serve as an assistant for musicians.

2019
Arhav Kapur from MIT created a device called Alter Ego that could detect what others were saying by detecting neuromuscular signals sent by the brain to the vocal cords.

2020
Marked by the COVID-19 pandemic, but this did not stop the rapid progress and advances in artificial intelligence and machine learning, particularly in the fields of health, economics, and language analysis. According to The Newstack, the development of natural language procedural models, a subfield of artificial intelligence, was a focus in 2020 and was used in language translation and chatbot applications.

2021
It was considered one of the toughest mathematical puzzles to crack: Could AI create its own algorithms to speed up matrix multiplication, one of machine learning’s most fundamental tasks? In a paper published in Nature on October 5, research lab DeepMind unveiled AlphaTensor, the “first artificial intelligence system for discovering novel, efficient and provably correct algorithms.” The Google-owned lab said the research “sheds light” on a 50-year-old open question in mathematics about finding the fastest way to multiply two matrices.

2022
AI will be the “new fuel for the modern economy” and argued that companies need to open up to AI to be ready for the future.
As we move forward in 2023, experts believe AI will continue to be a key part of digital transformation strategies in various sectors.
“Chatbots specifically will become more sophisticated and could soon be the future of fintech customer service,”

2023
AI will be the “new fuel for the modern economy” and argued that companies need to open up to AI to be ready for the future.
As we move forward in 2023, experts believe AI will continue to be a key part of digital transformation strategies in various sectors.
“Chatbots specifically will become more sophisticated and could soon be the future of fintech customer service,”
Advances through the years

Alan Turing, a mathematician, logician, theoretical computer scientist, cryptographer, philosopher, and British theoretical biologist, is widely considered one of the fathers of computer science and a pioneer in the field of artificial intelligence (AI). During World War II, Turing designed the bombe machine which helped speed up the decoding of messages.
In 1950, Turing wrote a paper entitled "Computing Machinery and Intelligence," in which he speculated on the concept of "thinking machines" and proposed the idea of the Turing Test, a measure of a machine's ability to exhibit intelligent behavior that is indistinguishable from that of a human. This paper is considered one of the first discussions of AI and remains an important reference in the field today.
American mathematician, scientist, and computer scientist John McCarthy followed Turing's work and is credited with coining the term "artificial intelligence". McCarthy recognized that the key to AI was automation, which enables machines to solve problems and provide answers automatically without human intervention. He developed the Lisp computer language in 1958, which later became the standard programming language for AI.
At the time, the high cost of computing and the lack of mechanisms to store commands made it difficult to pursue AI research. In the early 1950s, the cost of renting a computer was $200.00 per month, and only large technology companies and prestigious universities had access to the machines.
Years later, Allen Newell, Cliff Shaw, and Herbert Simon created a program called The Logic Theorist, funded by the RAND Corporation, to imitate the problem-solving abilities of a human being. The program was presented at the Dartmouth Summer Research project on AI in 1956 by John McCarthy, who brought together the best researchers in the field, but his goal of reaching an agreement on standard methods for AI was not achieved.

Literature and technology: Science Fiction
