Compartilhar:

Artigo - PDF

 

Scientific Society Journal  ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​ ​​​​ 

ISSN: 2595-8402

Journal DOI: 10.61411/rsc31879

REVISTA SOCIEDADE CIENTÍFICA, VOLUME 7, NÚMERO 1, ANO 2024
.

 

 

ARTIGO ORIGINAL

Notes on artificial intelligence: concepts, applications and techniques

Fabíola Alves Alcântara1 ; Eugênio da Silva2 ; Rodrigo Siqueira Batista3

 

Como Citar:

ALCÂNTARA, Fabíola Alves; DA SILVA, Eugênio; BATISTA, Rodrigo Siqueira. Notes on artificial artificial intelligence: concepts, applications and techniques. Revista Sociedade Científica, vol.7, n. 1, p.2970-3008, 2024.

https://doi.org/10.61411/rsc202457217

 

DOI: 10.61411/rsc202457217

 

Área do conhecimento: Interdisciplinar.

.

Palavras-chaves: Inteligência Artificial; Aprendizagem de Máquina; Aprendizagem Supervisionada; Problemas de Classificação.

 

Publicado: 03 de junho de 2024.

Resumo

A Inteligência Artificial (IA) tem estado cada vez mais presente no mundo, com aplicabilidade em diversas áreas do conhecimento. Soluções baseadas em IA, implementadas com diferentes técnicas, estão presentes em diferentes sistemas. O objetivo deste artigo é (1) apresentar um breve histórico e os principais conceitos relacionados à Inteligência Artificial – com foco em técnicas de aprendizado de máquina (AM) –, (2) discutir suas nuances, técnicas e (3) apresentar algumas aplicações em diversas áreas. A intenção é abordar o assunto de forma acessível ao público não especializado, de forma a promover a compreensão de seus principais conceitos, mas sem a necessidade de recorrer a detalhes muito técnicos. Para isso, é realizada uma revisão narrativa da literatura sobre IA abordando brevemente sua história, conceitos, interseções com outras áreas e aplicações. O segmento de aprendizado de máquina recebe atenção especial, com destaque para o aprendizado supervisionado e sua aplicação em problemas de classificação. O artigo também destaca as capacidades apresentadas pelas mais modernas técnicas de aprendizado de máquina que, em alguns casos, apresentam resultados melhores do que os obtidos por humanos. Dados esses resultados promissores, o futuro da IA aponta para a criação de sistemas com capacidades que se assemelharão cada vez mais às do intelecto humano. Porém, a criação de máquinas verdadeiramente pensantes, e não de simuladores pensantes, ainda deve permanecer por muito tempo como um objetivo a ser alcançado.

 

 

..

Abstract

Artificial Intelligence (AI) has been increasingly present in the contemporary world, with applicability in various fields of knowledge. AI-based solutions, implemented with different techniques, are present in different systems. The purpose of this article is to (1) present a brief history and the main concepts related to Artificial Intelligence - with a focus on machine learning (ML) techniques -, (2) discuss its nuances, techniques and (3) present some application examples in several areas. The intention is to approach the subject in an accessible way to the non-specialized public, in order to promote the understanding of its main concepts, but without the need to resort to very technical details. For this, a narrative review of the literature on AI is carried out, briefly addressing its history, concepts, intersections with other areas and applications. The machine learning segment receives special attention, with emphasis on supervised learning and its application in classification problems. The article also highlights the capabilities presented by the most modern machine learning techniques which, in some cases, present better results than those obtained by humans. Given these promising results, the future of AI points to the creation of systems with capabilities that will increasingly resemble those of the human intellect. However, the creation of truly thinking machines, and not of thinking simulators, should still remain for a long time as an objective to be achieved.

Keywords: Artificial Intelligence; Machine Learning; Supervised Learning; Rating Problems

..

1. Introduction

Artificial Intelligence (AI) is currently one of the main topics of debate in the world. Important cinematographic works (for example, Her, Eve, Ex machina and Westworld) and various documentaries (Coded Bias, AlphaGo and The dilemma of networks) on streaming platforms have contributed to the discussion on the limits and applications of the subject (MARQUES ANDREON et al., 2022). However, the phenomenon called ChatGPT has gained the global spotlight in recent months, literally causing a transformation in the technological landscape, which has consequently made the subject increasingly "familiar" to the non-specialized public.

In 2020, McKinsey Analytics published The State of AI in 2020, an online survey answered by 2,395 individuals with the aim of analyzing the conditions for implementing AI in institutions (MCKINSEY ANALYTICS, 2020). According to Stanford University's The AI Index 2022 Annual Report, more than 93.5 billion dollars were invested in AI in 2021. This is more than double the total invested in 2020 (DANIEL et al., 2022). The results showed that companies are increasingly investing in AI, but it was during the COVID-19 pandemic that there was a significant increase in the adoption of these technologies, especially in the health and pharmaceutical sectors (MOTTA et al., 2020).

AI has applicability in various fields of knowledge, whether in astronomy, physics and mathematics, or in medicine, genetics and cell biology, as well as bioethics, law, economics and the arts (BITTAR, ALVES & MELO, 2018; CAETANO, MANZOLLI & ZUBEN, 2005; CHANG, 2018; GUNKEL, 2017; WUERGES & BORBA, 2010). In this sense, Russell & Norvig (2021, p. 19) describe AI as "one of the most interesting and fast-growing fields".

Today, AI-based solutions, implemented with different techniques, are present in the most diverse types of systems. These include systems for translation, logistics organization, investments, image and voice recognition, games, robot control and many other artifacts (CASTRO & FERRARI, 2016; JUNIOR et al., 2017; NASCIMENTO, SILVA & SIQUEIRA-BATISTA, 2018). Also according to Russel & Norvig (2021, p. 27), AI is not "magic or science fiction, but science, engineering and mathematics" and, as a result, its incorporation among so many media requires a further understanding of its potential and limitations.

Based on these considerations, the aim of this article is to (1) present a brief history and the main concepts related to AI - with a focus on machine learning (ML) techniques -, (2) discuss its nuances and (3) present some examples of application in various areas. The special interest in ML is due to the large number of studies carried out using techniques of this type. It should be noted that the intention of the text is to present the concepts related to the topic - as well as reflections on its applicability - in a way that is accessible to a non-specialist audience. To this end, this review includes sections on the concept of AI from different points of view; areas of knowledge that have contributed to the development of AI; machine learning, its paradigms and applications; and some final considerations on the subject.

..

2. Methodology

The article presents a narrative literature review - which is useful "to describe and discuss the state of the art of a given subject, from a theoretical and contextual point of view, which basically consists of an analysis of the literature published in books, articles in printed and/or electronic journals and the author's personal interpretation and critical analysis" (ROTHER, 2007, p. 5) - involving a survey of scientific knowledge about AI. To compose the text, publications were selected from books and scientific journals that deal with the subject of "Artificial Intelligence", with a special interest in those that present the concepts and paradigms of machine learning, as well as their applicability in the most diverse areas of knowledge.

..

3.History

Although in its infancy, the idea of artificial intelligence comes from times past, when concerns were already focused on developing machines or methods that could reproduce human abilities. The philosopher Aristotle (384-322 BC), for example, set out to find a way of codifying "correct thought" (ARISTÓTELES, 2021). As a result of these efforts, formal logic emerged which, after a series of improvements, evolved into symbolic logic - which uses symbols for its concepts/conclusions (GEORGE BOOLE, 1854) - and from there into mathematical logic - which underpins mathematical principles and demonstrations (WHITEHEAD & RUSSEL, 1910) - becoming an essential tool in the reasoning processes used, above all, in the exact sciences.

More concrete initiatives around AI only emerged during the Second World War (1939-1945), motivated by demands to create technologies that would boost the war industry. During this period, studies began to be carried out (in concentration camps) in order to study the brains of human beings (TEIXEIRA, 2019). It is worth noting that the studies carried out at this time had inhumane, cruel, highly reprehensible and often deadly methods.

The Hixon Symposium in 1948 in the United States was a major scientific milestone. On that occasion, the researchers present were able to visualize existing and approximate relationships between the human brain and computers. More than a decade before this event, in 1936, the young English mathematician Alan Turing had written the article "On Computable Numbers With an Application to the Entscheidungsproblem", in which he established the principles of computing through a theoretical, abstract artifact that can be reproduced with paper and pencil, commonly referred to as the "Turing Machine" (TURING, 1936).

"(...) Turing's breakthrough consists in the fact that he demonstrated, through the invention of his machine, that any and all tasks that can be represented in the form of an effective procedure can be mechanized, i.e. can be carried out by a computer... and that any and all types of computer can ultimately be reduced to a Turing machine... because they can be imitated by his machine... making it a true universal principle" (TEIXEIRA, 2019, p. 14).

The Turing Machine is of great importance to computing because it formalizes the concept of algorithm, the essence of what lies behind computing. However, it wasn't until 1950, with the article "Computing Machinery and Intelligence", that Turing made a definitive contribution to the development of AI, mainly by presenting an operational definition for the term "intelligence". On that occasion, Turing proposed a kind of imitation game that became popular under the name "Turing test" (TURING, 1950).

Figure 1- Illustration of the Turing Test. Source: Prepared by the authors

.

As Figure 1 illustrates, there are three participants in this game: two people and a computer. The interrogator (one of the two people) remains in a separate room asking subjective questions to try to find out who of the other two is a person. If, after a certain time, the interrogator is unable to identify this difference, it is proposed that the computer has assumed a behavior considered intelligent (TURING, 1950).

Another important milestone was an academic seminar that addressed questions about intelligence, automata and neural networks, which took place in Dartmouth, USA, in 1956 (MCCARTHY, MINSKY, ROCHESTER & SHANNON, 1955). This event discussed issues related to problem solving, the computer's ability to "think" and the feasibility of a machine capable of playing chess (MCCARTHY, MINSKY, ROCHESTER & SHANNON, 1955; NEWELL, SHAW & SIMON, 1958). It's worth noting that this was the first use of the term AI (RUSSEL & NORVIG, 2021 p. 17). From then on, various works and programs emerged that boosted the growth and development of the field, such as GPS (General Problem Solver), a problem "solver" created by Newell & Simon (1961); Lisp, the first programming language for AI and the Advice Taker program, both developed by McCarthy (1959); as well as the creation of perceptrons by Rosenblatt (1962) and the first neural networks (WINOGRAD & COWAN, 1963). In the 1970s, the first knowledge-based systems appeared, most notably DEBDRAL (FEIGENBAUM, LEDERBERG & BUCHANAN, 1968) and MYCIN (SHORTLIFFE, 1976). Later, in the 1980s, more flexible neural networks with greater learning capacity emerged (RUMELHART & HINTON, 1986), which could be applied to solving more complex problems.

Currently, deep learning techniques are the latest in machine learning and are present, for example, in Convolutional Neural Networks (CNN), Deep Belief Networks (DBN), Deep Boltzmann Machines (DBM), as well as Recurrent Neural Networks (RNN) (PONTI & COSTA, 2017). More recently, generative AI, also based on the concept of deep learning, has stood out for being the basis of the intelligence behind ChatGPT (Generative Pre-trained Transformer) (OPENAI, 2022) and Midjourney (MIDJOURNEY, 2022) - the former aimed at generating texts and the latter at generating images.

.

4.What is artificial intelligence?

The characterization of the concept of AI ultimately depends on an understanding of the terms "Intelligence" and "Artificial", since these two words make up the term "Artificial Intelligence". Mora (2001, p. 1873) understands Intelligence "as the capacity of certain organisms to learn and to apply learning". In relation to the term Artificial, it can be designated as "everything that is made by human beings, in other words, an artifact" (ZUBEN, 2011, p. 4). Roughly speaking, combining the definitions of the words separately, it can be said that AI corresponds to the ability of artifacts (computers) to perform tasks that require intelligence to be carried out, i.e. that have historically depended on the human intellect to be performed.

According to Rich (1988), AI is a field of study whose goal is to create systems capable of performing tasks that humans are better at. On the other hand, Bigonha (2018) states that AI examines and produces systems prepared to perform actions similar to those of people, such as learning and making decisions, only with a great deal of agility. Goldschmidt (2010, p. 7) defines computational intelligence as "a multidisciplinary science that seeks to develop and apply computational techniques that simulate human behavior in specific activities".

Russel & Norvig (2021) provide different definitions for AI, organized into four categories - "acting like humans", "acting rationally", "thinking like a human" and "thinking rationally" - in which they present a different - sometimes complementary - view of the subject.

.

4.1 Acting like human beings

Here, the ability of a computer to impersonate a human being, or in other words, to overcome the Turing Test, stands out as a model of intelligence. In this "Imitation Game", the physical need for a human being to issue answers is "discarded" and the computer, in an "intelligent" way, manages to replace him satisfactorily. Two definitions, considering this aspect, are presented below:

a."The art of creating machines that perform functions that require intelligence when performed by people" (KRZWEIL, 1990, p. 14)

b."The study of how computers can do tasks that are now better performed by people" (RICH & KNIGHT, 1991, p. 3).

.

4.2 Acting rationally

From this point of view, the intelligence model is associated with the rational agent, where rationality is the means to guarantee the most appropriate results according to some predefined utility metric. It's important to note that human and rational behavior are not mutually exclusive. This differentiation is just to highlight that humans don't behave in a strictly rational way all the time. In other words, human decisions are not always made with the aim of achieving the maximum measure of satisfaction. The following definitions take this point of view into account:

a.  "Computational Intelligence is the study of the design of intelligent agents" (POOLE et al., 1998, p. 1).

b."is related to an intelligent performance of artifacts" (NILSSON, 1998, p. 1).

 

4.3 Thinking like a human

In this context, knowledge from cognitive science is used to idealize a model of intelligence that consists of "reproducing" the human mind, i.e. the way human beings reason. The following definitions are in line with this perspective:

  • ​​ "The new and interesting effort to make computers think (...) machines with minds, in the full and literal sense" (HAUGELAND, 1985, p. 2)

  •  ​​​​ "[Automation of] activities that we associate with human thought, activities such as decision-making, problem-solving, learning..." (BELLMAN, 1978, p. 3).

.

4.4 Thinking rationally

Finally, we highlight a model of intelligence that is based on the Aristotelian proposal of "codification of thought". This codification presupposes the use of logical resources to build mechanizable and irrefutable reasoning processes, allowing a computational device to conduct "correct thinking". The following definitions take this aspect into account:

  • "The study of mental faculties through the use of computer models" (CHARNIAK & MCDERMOTT, 1985, p. 6)

  • "The study of the computations that make it possible to perceive, reason and act" (WINSTON, 1992, p. 5).

.

4.5 A brief summary

It is possible to state that the tasks of interest to AI are those that, at least in theory, cannot be mechanized, i.e. they cannot be converted into an effective procedure (algorithm) and, consequently, cannot be carried out by a computer. These are tasks in which an extra "ingredient", called intelligence, needs to be used to perform them. Tasks of this type are those that essentially involve decision-making and are associated with a wide variety of areas of knowledge, such as health, law, finance, engineering, business, tourism and marketing, as will be discussed in the next section.

.

5.Intersections between artificial intelligence and other areas

AI is an area of knowledge whose foundations have been provided by other more traditional and historically consolidated areas (SIQUEIRA-BATISTA & SILVA, 2019; RUSSEL & NORVIG, 2021). On the other hand, AI has already acquired a sufficient degree of maturity to allow it to be applied to the development of various areas, including those on which it is based. The following are the main areas of knowledge that contribute concepts and techniques to building the foundations of AI, as well as how AI has contributed to the development of these areas.

.

5.1Philosophy

The construction of an AI project, according to Porto (2006), must take into account the metaphysical foundations of AI, paying attention to the use of the term "intelligent programs", meaning that the "intelligence" of this program is only related to the complexity of the program and not precisely to "intelligence" in general. Indeed, in agreement with Gava (2017), the view of AI researchers that computers can and/or could reproduce the functional characteristics of the human mind has had a major impact on disciplines focused on the philosophy of mind:

"The philosophy of mind has as its main issue the search for arguments that demonstrate what the real nature of the mind is and what relationship exists between the mind and the brain, as well as the advent of artificial intelligence, a new way of studying the mind through computational models has also emerged, which has been called the functionalist theory of mind" (GAVA, 2017, p. 8).

The discussion that permeates AI goes through philosophical issues related to the conceptual aspects of man, intelligence, the world, knowledge and technology, causing great impact and valuable insights into this debate (SILVEIRA, 2017). In this context, it is worth noting that Nakabayashi (2009) studied AI in the Philosophy of Mind as a science belonging to the nature of cognition and, in view of the findings - a bot (robot) capable of playing the role of a tutor in virtual teaching environments - he considered chatbots to be a very interesting alternative for use in education because their form of communication is based on natural language.

.

5.2Engineering

Engineering, especially those branches more closely related to electronics and telecommunications, has made indispensable contributions to the materialization of the concepts produced in the sphere of computing. It is through the data processing, storage, transmission and sharing devices produced by this area that computer systems in general can become a reality (RUSSEL & NORVIG, 2021).

On the other hand, engineering has been influenced by AI through the use of systems to collaborate in production, the stock sector, estimating demand, organizing orders and quality control, making it essential (SLACK, CHAMBER, HARDLAND, HARRISO & JOHNSTON, 2012). Another point is automation, which tries to improve the performance of industrial machinery and human-like techniques, with research into behavioral conduct and understanding human reasoning at its core (MITHIDIERI, BELIZARIO, SILVA, SANTOS & LIMA, 2018).

Recent research by Cardozo (2022) sought to capture information from a domestic environment using sensors and a fuzzy logic algorithm. It was possible to monitor luminosity, the presence of gases, people or animals in the environment, either locally or remotely, and it was found that the implementation of this system contributed to a reduction in electricity costs.

..

5.3Psychology

Debates on the relationship between psychology and artificial intelligence are not new (BURANI & VIEIRA, 2021). In fact, Tiberius (2016) correlated computing (for great revolutionary machines) with cognitive psychology (focused on the brain) through the conceptualization of memory, intelligence and the performance of computer systems. However, with the COVID-19 pandemic, these ties have been tightened, especially in the educational aspect, facilitating learning processes (BURANI & VIEIRA, 2020).

​​ "It's natural to think of AI research, on the reproduction of automated reasoning or intelligent behavior, as something based heavily on the technology that makes it possible to build these artifacts, with no commitment to a more reflective approach on aspects of how the mind works. However, AI can be thought of as an area of research that emerges from the activity of incessant scrutiny of the human mind, which began in Ancient Greece and spans centuries of history, when inserted into the present time with the manifestation of high technology. In this 21st century, humanity is able to artificially reproduce the natural processes of the human mind, externalizing intelligence. Perhaps this is the most peculiar aspect of AI. While philosophy and psychology seek to understand the aspects of the intelligent mind as internal manifestations and enclosed within the limits of being, AI, in turn, externalizes them" (MEDEIROS, MOSER & GARCIA, 2018, p. 91-92).

A recent study by Santamaría & Sánchez-Sánchez (2022) looked at the use of new technologies for the psychological assessment of professionals and patients. The authors found that the interaction between technology and psychology is still far from being a reality; they also recognized that a closer correlation between the fields must take into account the training of professionals in order to solve problems properly, so that bad practices don't occur:

"Artificial intelligence is a technology that lies halfway between science and art. Its aim is to build machines that, when solving problems, appear to think. A good example is the chess-playing machine" (TEIXEIRA, 2014, p. 7).

According to Teixeira (2014), robopsychiatry and robopsychology are set to revolutionize the understanding of mental illnesses, since the focus is on understanding robots with various particular and distinct characteristics or psychoses.

.

5.4Biology

Nature-inspired computing - as part of Natural Computing - uses ideas from reality to create artificial systems whose purpose is to solve complex problems, encompassing the fields of physics, chemistry, engineering, biology and computer science. Indeed:

"Nature-inspired systems have broken paradigms with regard to conventional technological solutions, which follow a strict set of rules and therefore often fail to produce satisfactory results in solving complex problems. This fact, combined with the infinite possibilities for applications of bioinspired systems, means that these systems are gaining more and more ground, either in solving new problems or in improving or replacing traditional models" (GOEDERT, FILHO & BLANCO, 2017, p. 40).

Among the main techniques involved are "Biologically Motivated Computing", "Biological Metaphor Computing" and "Bioinspired Computing" (BREVE, 2021). The latter, "Bioinspired Computing", one of the main methods investigated, employs different computational approaches, mainly artificial neural networks. These techniques are aimed at research focused on biology in its practical aspects, such as biological classifications and groupings related to DNA, as well as testing and comparison between species (GUARIZI & OLIVEIRA, 2014).

It is also worth highlighting, in the sphere of intersections between biology and AI, the potential of the concept of artificial life, in the following terms:

​​ "The concept of artificial life is still recent within studies of the human body, precisely because it is still closely linked to research into robotics and crude computing. There are, however, some researchers who are already talking about the post-human, which means that the exchange of information between the artificial and the biological can be thought of as part of studies into improving the human body. From this perspective, it seems that the human body would be just one of the models to be reproduced by artificial life, and possibly the most challenging and unattainable. However, it is worth looking at it from another angle, one in which artificial life promotes not only the means to reproduce biological life in an artificial environment, but also ends up driving technological advances to promote changes in human bodies too, so that the concept disconnects from raw science and begins to dialog with other areas of knowledge" (MARQUES & KRÜGER, 2019, p. 38 and 39).

In this sense, Reis (2017) analyzed cases resulting from the junction between computing, mathematics, biology and neuroscience and realized that lives can be germinated and proliferated in a natural environment (by bringing artificial life closer to natural life) and that biocybernetics can involve living beings (neuroscience), environments (biology) and objects (computing).

.

5.5Logic

Logic plays an extremely important role in the field of AI, developing reasoning and premises for solving problems (PEREIRA, 2001). Six objectives of AI are therefore described in relation to logic:

"Firstly, AI aims to mechanize logic, an absolutely essential tool for so many rational activities; Secondly, AI intends to make the subconscious logic we use explicit and well-defined, testing it objectively through its automation; Thirdly, AI employs logic as a generic language of communication, knowledge and procedures, between humans and computers, as well as between computers themselves; Fourthly, in AI, even when procedures and artifacts are not implemented using logic directly, logic can take on the role of a precise language for specifying the requirements of those procedures and artifacts, as well as a formalism through which to study their semantic properties; Fifthly, AI has contributed significantly to equating and examining the problem of identifying the limits of proper symbolic reasoning embodied in computers, as well as assessing whether these limits apply to human beings; Finally, and sixthly, AI has helped researchers to explore new questions and new methods of reasoning, as well as to combine disparate modes of reasoning into a uniform and unified framework, in order to deal with incomplete, imprecise, contradictory and changing information" (PEREIRA, 2001, p. 48). 48).

Logic and probability collaborate directly with the theories and measurements used by quantitative science; roughly speaking, it can be said that the evolution of computing and consequently its use by AI, has required a closer relationship between mathematics, logic and probability to solve problems (RUSSEL & NORVIG, 2021; SIQUEIRA-BATISTA & SILVA, 2019).

.

5.6Linguistics

Computer science related to linguistics - computational linguistics - is nothing more than a multidisciplinary branch of study focused on the syntactic, semantic and logical learning of natural language, with the aim of promoting improved technology in aspects involving communication (OLIVEIRA & FRESCHI, 2020).

"Although existing systems are far from reaching human capacity, they have countless possible applications, the aim is to create software products that have some knowledge of human language. These products will change our lives. They are urgently needed to improve human-machine interaction, because the main obstacle in human-computer interaction is simply a communication problem" (UZSKOREIT, 1997, p 2).

Computational linguistics is subdivided into two subfields: (1) corpus linguistics, which is concerned with language and its various linguistic forms; and (2) natural language processing, linked to the development of virtual assistance systems (chatbots), translators, voice recognition, among others (OTHERO & MENNUZZI, 1978; FINATTO et al., 2018).

An important piece of software used by linguists to work with corpus linguistics is WordSmith Tools, which has an arsenal of language patterns in various languages (SCOTT, 1996). Berber-Sardinha (2005) studied how to find the most important keywords in a corpus using WordSmith Tools and concluded that in order to have the best chance, the first keywords in the list, ordered by keyness, should be chosen.

Recent research has studied the use of AI through chatbots in the customer service process and has found several benefits related to cost, productivity, flexibility, quality and innovation (SCHUNK, 2020). Currently, many companies use chatbots (robots) to "triage" situations that can be solved through "robotic" customer service, differentiating them from those that will require an interlocution with a human.

In 2022, Meta, a giant technology company, unveiled a new AI model capable of translating 200 different languages with 44% accuracy (META, 2022). Microsoft recently released VALL-E, an AI capable of transforming a three-second audio into another audio with a different voice - I say "good morning" and the platform can generate a "good night" with my voice (WANG et al., 2023). At the same time as being an innovative tool, it can cause many problems, since creating audios not said by people and faithfully replicating them with their voices can contribute to the creation of Fake News.

.

6.Applications of AI

AI has a huge number of applications and, based on this, Russell & Norvig (2021) categorized the area according to the following segments: (i) Natural Language Processing (NLP), (ii) Automated Reasoning, (iii) Knowledge Representation, (iv) Computer Vision, (v) Robotics and (vi) Machine Learning (ML).

-Natural Language Processing (NLP) encompasses computational models capable of performing multiple tasks, such as organizing, translating and creating documents from a database, among others. The ways of carrying out this work can be based on rules (symbolic) or quantitative data (statistical) (ANDREATA, 2017).

-Automated Reasoning uses a logical inference mechanism that acts on facts and rules related to a given domain to derive conclusions or recommendations (RUSSEL & NORVIG, 2021).

-Knowledge representation is characterized by the ability of a computer to express information so that it can be used by an AI system, and this involves various forms and patterns of representation (CÂMARA, 2001).

-Computer Vision, according to Hollerweger (2019, p. 7), "comprises the set of processes (acquisition, processing, analysis and comprehension) applied to digital images, with the aim of automating tasks that require the perception provided by human vision".

-According to Mataric (2014, p. 21), robotics is "the study of robots, which means the study of their ability to sense and act in the physical world autonomously and intentionally".

Machine Learning (ML) is a sub-area of AI that works with factors and algorithms capable of learning to perform some task (ROZA, 2016).

It is worth noting that, in common, the tasks presented require skills that are considered intelligent in order to be carried out, which involve, for example, perception, reasoning, learning, planning, adaptation, prediction and language.

For the purposes of this article, it was decided to explore AM concepts and techniques in detail, focusing mainly on the supervised learning paradigm. This approach has a more "controlled" learning capacity and generally provides significant results in predictive tasks.

.

7.Machine learning

AM uses concepts from neuroscience, biology, statistics, mathematics and physics to make computers "learn". Imagine you're playing against a computer, 10 games have passed and you're still winning, but at a certain point, the computer starts to beat you. There are two options: you could be getting worse or the computer has learned to win. This means that if it has learned to win, it will be able to use the same strategies with other participants; this is a simple example of AM (MARSLAND, 2015).

Among the various possible explanations for AM, some stand out:

-"Machine Learning is about making computers modify or adapt their actions (to make predictions or control robots) to get more accuracy" (MARSLAND, 2015, p. 4).

-"Machine learning is the science (and art) of programming computers so that they can learn from data" (GÉRON, 2019, p. 2).

-"The ability to improve performance in performing some task through experience" (MITCHELL, 1997, p. 2).

-"Machine learning techniques are mainly used to solve problems involving phenomena for which there are no known analytical models that adequately represent them" (SIQUEIRA-BATISTA & SILVA, 2019, p. 46).

A variety of AM techniques can be applied in everyday situations, whether it's to improve healthcare (diagnosing diseases, making prognoses, monitoring patients, indicating medication, monitoring medical records...), translations, speech recognition, safer transportation, control and detection of possible fraud (e.g. credit card scams), among many other benefits. Despite their extraordinary growth, these AM techniques can also have negative points such as reducing human resources in companies, exacerbating social inequalities, accentuating discrimination and loss of privacy, among others (LUDEMIR, 2021).

In the context of ML, learning consists of building models capable of performing different tasks, which are basically divided into predictive and descriptive. Predictive models are directly related to a paradigm called supervised learning, while descriptive tasks are related to the unsupervised learning paradigm. Figure 2 highlights these relationships (FACELI et al., 2017).

 

Figure 2 - Learning paradigms and their tasks. Source: Faceli et al. (2010) adapted by the authors.

.

AM also considers other paradigms such as semi-supervised, active and reinforcement learning. In this article, however, only details related to the supervised paradigm are covered, with emphasis on its application in classification tasks.

.

7.1Supervised Learning

Supervised ML is applied to predictive tasks which can be of two types: classification and regression. In both cases, each element that takes part in the learning process is described by a set of descriptor attributes and there is also a target attribute, which establishes the label associated with that element. In cases where this label takes on a categorical value, there is a classification task, while when it takes on a numerical value, there is a regression task (RUSSEL & NORVIG, 2021).

Regardless of the type of task, the aim of supervised learning is to build predictive models with good generalization capacity, i.e. capable of predicting the target value for new elements that did not take part in the learning process, but which have similarities to those (FACELI et al., 2017).

.

7.1.1Regression

Regression models attempt to discover the behavior of one variable in relation to the changes undergone by others (CHEIN, 2019). In the meteorological field, it is very common to use regression models to outline temporal analyses. Another example is trying to predict age from an image. Figure 3 illustrates a learning problem involving a regression task. In it there is a data set in which each element is made up of two attributes, a descriptor x (independent variable) and a target y (dependent variable). The regression model learned corresponds to a linear estimator. In general, regression models related to real problems require the construction of much more complex estimators, which are highly non-linear and involve many descriptor attributes.

 

Figure 3 - Linear regression model obtained from a data set. Source: Prepared by the authors

.

 

7.1.2Classification

Classification models try to separate a set of elements according to previously established categories. Sorting the answers to a survey into domains and categorizing them based on certain criteria corresponds to a classification task. Another example of classification, presented in Santos (2016), refers to the granting of credit by a bank. For the bank to grant credit, it needs to have financial data on its customers and, according to their profile, classify them as to the risk of default in order to decide whether or not to grant credit. Figure 4 illustrates a learning problem involving a classification task. In it, the elements of the data set are represented by the descriptor attributes x and y and by the target attribute which establishes whether an element belongs to the class of "clouds" or "stars". The learned classification model is represented by a linear separator that establishes a decision boundary between the classes involved. As in the case of regression models, classification models associated with more complex problems require the construction of non-linear separators, which involve many descriptor attributes and many classes (FACELI et al., 2017).

 

Figure 4 - Classification model obtained from a data set. Source: Prepared by the authors.

.

There are various methods that can be used to build predictive classification models. In general, these methods are based on distances, probability, search, optimization, hybrid, among many others (FACELI et al., 2017).

.

7.1.2.1 Distance-based methods

These include techniques that use a proximity criterion to assign a label to an element that has not yet been classified. A well-known distance-based algorithm is k-NN (k- nearest neighbor), which assigns the most frequent label among the k nearest neighbors to a new element (FACELI et al., 2017). Figure 5 shows an example of a data set in which each element represents a message received by email and is described by two attributes.