128 32 5MB
English Pages 423 [424] Year 2023
Victor Hugo C. de Albuquerque, Pethuru Raj, Satya Prakash Yadav (Eds.) Toward Artificial General Intelligence
Also of interest Artificial Intelligence for Virtual Reality Edited by: Jude Hemanth, Madhulika Bhatia and Isabel De La Torre Diez, ISBN ----, e-ISBN ----
Quantum Computing and Artificial Intelligence. Training Machine and Deep Learning Algorithms on Quantum Computers Edited by: Pethuru Raj, Abhishek Kumar, Ashutosh Kumar Dubey, Surbhi Bhatia and Oswalt Manoj S, ISBN ----, e-ISBN ---- Artificial Intelligence of Things in Smart Environments. Applications in Transportation and Logistics Edited by: Mariyam Ouaissa, Zakaria Boulouard, Mariya Ouaissa and Yassine Maleh, ISBN ----, e-ISBN ---- Artificial Intelligence for Signal Processing and Wireless Communication Edited by: Abhinav Sharma, Arpit Jain, Ashwini Kumar Arya and Mangey Ram, ISBN ----, e-ISBN ----
Handbook of Augmented and Virtual Reality Edited by: Sumit Badotra, Sarvesh Tanwar, Ajay Rana, Nidhi Sindhwani and Ramani Kannan, ISBN ----, e-ISBN ----
Toward Artificial General Intelligence Deep Learning, Neural Networks, Generative AI Edited by Victor Hugo C. de Albuquerque, Pethuru Raj, Satya Prakash Yadav
Editors Prof. Victor Hugo C. de Albuquerque Department of Teleinformatics Engineering Federal University of Ceará Av. da Universidade 2853 60020-180 Fortaleza, Brazil [email protected] Pethuru Raj, Ph.D. Reliance Jio Platforms Ltd. D 002, Prospect Princeton Apartment 560068 Bangalore, India [email protected] Dr. Satya Prakash Yadav Department of Computer Science and Engineering GL Bajaj Institute of Technology and Management Kalam Road, Plot No.2 APJ Abdul, Knowledge Park 3 201306 Greater Noida, Uttar Pradesh, India [email protected]
ISBN 978-3-11-132356-5 e-ISBN (PDF) 978-3-11-132374-9 e-ISBN (EPUB) 978-3-11-132416-6 Library of Congress Control Number: 2023942201 Bibliographic information published by the Deutsche Nationalbibliothek The Deutsche Nationalbibliothek lists this publication in the Deutsche Nationalbibliografie; detailed bibliographic data are available on the Internet at http://dnb.dnb.de. © 2024 Walter de Gruyter GmbH, Berlin/Boston Cover image: Shutthiphong Chandaeng/iStock/Getty Images Plus Typesetting: Integra Software Services Pvt. Ltd. Printing and binding: CPI books GmbH, Leck www.degruyter.com
Preface Artificial intelligence (AI) has been the prominent field of study and research in educational institutions and research labs across the globe. Strategically speaking, AI has the wherewithal to bring a series of delectable and deft advancements across industry verticals. As widely accepted, AI has turned out to be the prime and pioneering enabler of business agility and automation. In other words, the real business transformation is to be achieved through the smart leverage of the distinct AI power. AI is all about swift and sagacious transitioning all kinds of intertwined digital data into timely and actionable insights. The knowledge discovered through the growing array of AI technologies and tools gets disseminated into all kinds of digital assistants, machineries at the manufacturing floors, defense equipment, medical instruments, handhelds, wearables, portables, mobile devices, robots, vehicles, drones, etc. Further on, business workloads and IT services are empowered through timely and actionable insights to make them competent, cognizant, and cognitive. AI intrinsically simplifies and speeds up the complicated process of fulfilling the mantra of data-driven insights and insights-driven decisions. In short, artistically replicating the contextually and comprehensively learning, articulating, and decision-making capabilities of the human brain in IT products, solutions, and services is the principal role and responsibility of the AI paradigm. AI is becoming penetrative, pervasive, and persuasive too. All kinds of digital systems (the digitized version of physical, mechanical, electrical, and electronics systems) are being empowered with AI phenomenon to exhibit adaptive and autonomous competency in their everyday operations, offerings, and outputs.
https://doi.org/10.1515/9783111323749-202
Contents Preface
V
List of authors About the editors
IX XIII
Mahesh Maurya, Vishal Gangadhar Puranik, A. Senthil Kumar, Balambigai Subramanian 1 Introduction to artificial intelligence 1 Shaik Mohammad Rafee, M. Prasad, M. Sunil Kumar, Balamurugan Easwaran 2 AI technologies, tools, and industrial use cases 21 G. Venkataramana Sagar, S. Ambareesh, P. John Augustine, R. Gayathri 3 Classification and regression algorithms 53 Ashwin M., Nazeer Shaik, Faiz Akram, Getachew Mamo Wegari 4 Clustering and association algorithm 87 Nazeer Shaik, S. Chandra Sekaran, Amit Mahajan, Balkeshwar Singh 5 Reinforcement learning 109 Syed Ibad Ali, Padmavathi Thasappan Kalaivaani, Ambigaipriya S., MD Rafeeq 6 Evaluation of AI model performance 125 K. Arthi, Veeramalai Sankaradass, Nikhat Parveen, Muralidharan J. 7 Methods of cross-validation and bootstrapping 145 S. Ramasamy, Kantharaju H. C., N. Bindu Madhavi, Haripriya M. P. 8 Meta-learning through ensemble approach: bagging, boosting, and random forest strategies 167 Abhisek Sethy, Nazeer Shaik, Pavan Kumar Yadavalli, S. P. Anandaraj 9 AI: issues, concerns, and ethical considerations 189 Abhisek Sethy, Nazeer Shaik, V. Saravanan, S. Karthika 10 The future with AI and AI in action 213
VIII
Contents
S. P. Sreekala, S. Revathy, S. Rajeshwari, B. Girimurugan 11 A survey of AI in industry: from basic concepts to industrial and business applications 233 S. Revathy, S. P. Sreekala, D. Praveenadevi, S. Rajeshwari 12 The intelligent implications of artificial intelligence-driven decisionmaking in business management 251 P. Krishna Priya, Karthik Rudra, Deepak Sai Kandula, G. Charan Teja, K. Venkata Koushik Reddy 13 An innovative analysis of AI-powered automation techniques for business management 269 B. Girimurugan, S. Rajeshwari, S. P. Sreekala, S. Revathy 14 The smart and secured AI-powered strategies for optimizing processes in multi-vendor business applications 287 S. Rajeshwari, D. Praveenadevi, S. Revathy, S. P. Sreekala 15 Utilizing AI technologies to enhance e-commerce business operations 309 Vijaya Gunturu, Charanjeet Singh, Nikhil S. Patankar, S. Praveena 16 Exploring the potential of artificial intelligence in wireless sensor networks 331 M. V. Jagannatha Reddy, J. Somasekar, Kaushalya Thopate, Sanjeev Kumar Angadi 17 Exploring artificial intelligence techniques for enhanced sentiment analysis through data mining 345 Senthil Kumar C, A. R. Arunarani, Piyush Charan, Sanjeev Kumar Angadi 18 Exploring the potential of artificial intelligence for automated sentiment analysis and opinion mining 361 Pramendra Kumar, Gunawan Widjaja, Nikhil S. Patankar, B. Suresh Kumar 19 A novel blockchain-based artificial intelligence application for healthcare automation 373 Soumya Mishra, Priyanka Sharma, Bhoopathy Varadharajan, Bhola Khan 20 Enhancing industrial efficiency with AI-enabled blockchain-based solutions 387 Index
401
List of authors Dr. Abhisek Sethy Department of CSE, Srinivasa Ramanujan Institute of Technology, Anantapur, Andhra Pradesh, India [email protected] Dr. S. Ambareesh School of Computer Science and Engineering, JAIN (Deemed-to-be University), Bangalore, Karnataka, India [email protected] Ambigaipriya S. Department of EEE, Mookambigai College of Engineering, Pudukkottai, Tamil Nadu, India [email protected] Dr. Amit Mahajan Centre for IT enabled services, University of Jammu, Jammu Tawi, Jammu and Kashmir, India [email protected] Dr. S. P. Anandaraj Presidency University, Bangalore, Karnataka, India [email protected] Dr. K. Arthi Data Science and Business Systems, School of Computing, SRM Institute of Science and Technology, Kattankulathur, Chennai, Tamil Nadu, India [email protected] A. R. Arunarani Department of Computational Intelligence, School of Computing, College of Engineering and Technology, SRM Institute of Science and Technology, Kattankulathur campus, Chennai, Tamil Nadu, India [email protected] Dr. Ashwin M. Department of Artificial Intelligence and Data Science, Koneru Lakshmaiah Education Foundation, Vaddeswaram, Andhra Pradesh, India [email protected] https://doi.org/10.1515/9783111323749-204
Dr. Balambigai Subramanian Department of ECE, Kongu Engineering College, Perundurai, Tamil Nadu, India [email protected] Dr. Balamurugan Easwaran Texila American University, Zambia [email protected] Dr. Balkeshwar Singh Department of Mechanical Engineering, Program of Manufacturing Engineering, Adama Science and Technology University, Adama City, Ethiopia [email protected] Bhola Khan Department of Economics, Yobe State University, Damaturu, Nigeria [email protected] Dr. Bhoopathy Varadharajan Department of Computer Science and Engineering, Malla Reddy College of Engineering, Secunderabad, Telangana, India [email protected] Dr. N. Bindu Madhavi KL Business School and Programme Coordinator (MBA), KL Centre for Distance and Online Education (CDOE), Koneru Lakshmaiah Education Foundation (Deemed to be University), Andhra Pradesh, India [email protected] Dr. S. Chandra Sekaran Department of Computer Science and Engineering, PSV College of Engineering and Technology, Krishnagiri, Tamil Nadu, India [email protected] G. Charan Teja Department of BBA, KL Business School, KL Education Foundation, Vaddeswaram, Andhra Pradesh, India [email protected]
X
List of authors
Charanjeet Singh Electronics and Communication Department, Deenbandhu Chhotu Ram University of Science and Technology, Murthal, Haryana, India [email protected] Deepak Sai Kandula Department of BBA, KL Business School, KL Education Foundation, Vaddeswaram, Andhra Pradesh, India [email protected] Dr. Faiz Akram Faculty of Computing and Informatics, Jimma Institute of Technology, Jimma University, Jimma, Oromia, Ethiopia [email protected] Dr. R. Gayathri Department of ECE, Rajalakshmi Engineering College, Chennai, Tamil Nadu, India [email protected] Dr. Getachew Mamo Wegari Faculty of Computing and Informatics, Jimma Institute of Technology, Jimma University, Jimma, Oromia, Ethiopia [email protected] Dr. B. Girimurugan KL Business School, Koneru Lakshmaiah Education Foundation, Guntur, Andhra Pradesh, India [email protected] Gunawan Widjaja Faculty of Law, Faculty of Public Health, Universitas 17 Agustus 1945, Indonesia [email protected] Dr. Haripriya M. P. Department of Computer Science, Kristu Jayanti College (Autonomous), Kothanur, K. Narayanapura, Bengaluru, Karnataka, India [email protected]
Dr. M. V. Jagannatha Reddy Department of CSE, Gandhi Institute of Technology and Management (GITAM Deemed to be University), Bengaluru Campus, Nagadenahally, Karnataka, India [email protected] Dr. P. John Augustine Department of Information Technology, Sri Eshwar College of Engineering, Kondampatti (post), Vadasithur (via), Kinathukadavu, Coimbatore, Tamil Nadu, India [email protected] Dr. Kantharaju H. C. Department of AI & ML, Vemana Institute of Technology, Koramangala, Bengaluru, Karnataka, India [email protected] Karthik Rudra Department of BBA, KL Business School, KL Education Foundation, Vaddeswaram, Andhra Pradesh, India [email protected] Dr. S. Karthika Product Development Intern (Virtual), Radicalx co, New York, USA [email protected] Dr. Kaushalya Thopate Computer Engineering Department, Vishwakarma Institute of Technology, Pune, Maharashtra, India [email protected] Dr. P. Krishna Priya KL Business School, KL Education Foundation, Vaddeswaram, Andhra Pradesh, India [email protected] Dr. Mahesh Maurya Associate Professor, K. C. College of Engineering, Mith Bunder Road, Kopri, Thane, Maharashtra, India [email protected]
List of authors
XI
Dr. Muralidharan J. Department of ECE, KPR Institute of Engineering and Technology, Avinashi Road, Arasur, Coimbatore, Tamil Nadu, India [email protected]
Dr. M. Prasad School of Computer Science and Engineering, Vellore Institute of Technology, Chennai, Tamil Nadu, India [email protected]
Nazeer Shaik Department of CSE, Srinivasa Ramanujan Institute of Technology, Anantapur, Andhra Pradesh, India [email protected]
S. Praveena Department of ECE, Mahatma Gandhi Institute of Technology, Hyderabad, Telangana, India [email protected]
Dr. Nikhat Parveen Department of Computer Science and Engineering, Koneru Lakshmaiah Education Foundation, Guntur, Andhra Pradesh, India [email protected] Nikhil S. Patankar Department of Information Technology, Sanjivani College of Engineering, Kopargaon, Maharashtra, India [email protected] Dr. Padmavathi Thasappan Kalaivaani Department of ECE, Vivekanandha College of Engineering for Women (Autonomous), Tiruchengode, Namakkal, Tamil Nadu, India [email protected] Dr. Pavan Kumar Yadavalli Department of Special Needs and Inclusive Education, Institute of Education and Behavioral Science, Ambo University, Ambo, Ethiopia [email protected] Dr. Piyush Charan Department of ECE, Manav Rachna University, Faridabad, Haryana, India [email protected] Pramendra Kumar Department of Computer Science and Engineering, Manipal University, Jaipur, India [email protected]
Dr. D. Praveenadevi KL Business School, Koneru Lakshmaiah Education Foundation, Guntur, Andhra Pradesh, India [email protected] Priyanka Sharma Department of Technical Education, Government Polytechnic, Barabanki, Uttar Pradesh, India [email protected] Dr. MD Rafeeq Department of CSE, CMR Engineering College, Kandlakoya, Medchal, Telangana, India [email protected] S. Rajeshwari School of Management, KPR College of Arts Science and Research, Coimbatore, Tamil Nadu, India [email protected] Dr. S. Ramasamy Department of Computer Science and Engineering, Hindusthan Institute of Technology, Coimbatore, Tamil Nadu, India [email protected] Dr. S. Revathy Department of Corporate Secretaryship, PSG College of Arts and Science, Coimbatore, Tamil Nadu, India [email protected]
XII
List of authors
Sanjeev Kumar Angadi Computer Science and Engineering, Nutan College of Engineering and Research, Pune, Maharashtra, India [email protected] Dr. V. Saravanan Department of Computer Science, College of Engineering and Technology, Dambi Dollo University, Dambi Dollo, Ethiopia [email protected] Dr. A. Senthil Kumar Artificial Intelligence Department, Sri Vishnu Engineering College for Women, Bhimawaram, Andhra Pradesh, India [email protected] Dr. Senthil Kumar C. Department of ECE, Sri Krishna College of Technology, Coimbatore, Tamil Nadu, India [email protected] Dr. Shaik Mohammad Rafee AI&ML Department, Sasi Institute of Technology and Engineering, Tadepalligudem, Andhra Pradesh, India [email protected] Dr. J. Somasekar Department of CSE, JAIN (Deemed-to-be University), Bangalore, Karnataka, India [email protected] Dr. Soumya Mishra Department of Electrical and Electronics Engineering, MVJ College of Engineering, Bangalore, Karnataka, India [email protected] Dr. S. P. Sreekala Research and Development Cell, School of Management, KPR College of Arts Science and Research, Coimbatore, Tamil Nadu, India [email protected]
Dr. M. Sunil Kumar Department of Computer Science and Engineering, School of Computing, Mohan Babu University (erstwhile Sree Vidyanikethan Engineering College), Tirupati, Andhra Pradesh, India [email protected] B. Suresh Kumar Department of EEE, Chaitanya Bharathi Institute of Technology, Hyderabad, Telangana, India [email protected] Dr. Syed Ibad Ali Department of Electronics and Telecommunication, GH Raisoni Institute of Engineering and Technology, Nagpur, Maharashtra, India [email protected] Dr. Veeramalai Sankaradass Department of Computer Science and Engineering, Chennai Institute of Technology, Kunrathur, Chennai, Tamil Nadu, India [email protected] K. Venkata Koushik Reddy Department of BBA, KL Business School, KL Education Foundation, Vaddeswaram, Andhra Pradesh, India [email protected] Dr. G. Venkataramana Sagar Department of ECE, G. Pulla Reddy Engineering College (Autonomous), Kurnool, Andhra Pradesh, India [email protected] Prof. Vijaya Gunturu Department of ECE, School of Engineering, SR University, Warangal, Telangana, India [email protected] Dr. Vishal Gangadhar Puranik School of Electronics and Telecommunication Engineering, MIT Academy of Engineering, Alandi, Pune, Maharashtra, India [email protected]
About the editors Prof. Dr. Victor Hugo C. de Albuquerque Department of Teleinformatics Engineering (DETI), Federal University of Ceará, Brazil Laboratory of Computer Systems Engineering (LESC) Laboratory of Automation, Control and Robotics (GPAR) Primary email ID: [email protected]
Prof. Dr. Victor Hugo C. de Albuquerque (Senior Member of IEEE) is a Professor and Senior Researcher at the Department of Teleinformatics Engineering (DETI)/Graduate Program in Teleinformatics Engineering (PPGETI) at the Federal University of Ceará (UFC), Brazil. He has earned a PhD in Mechanical Engineering from the Federal University of Paraíba (UFPB, 2010) and an MSc in Teleinformatics Engineering from the PPGETI/UFC (UFC, 2007). He completed a BSE in Mechatronics Engineering at the Federal Center of Technological Education of Ceará (CEFETCE, 2006). He has experience in Biomedical Science and Engineering, mainly in the research fields of Applied Computing, Intelligent Systems, and Visualization and Interaction, with specific interest in Pattern Recognition, Artificial Intelligence, Image Processing and Analysis, as well as Automation with respect to biological signal/image processing, biomedical circuits, and human/brain-machine interaction, including Augmented and Virtual Reality Simulation Modeling for animals and humans. Prof. Victor is a Full Member of the Brazilian Society of Biomedical Engineering (SBEB). He is Editor-in-Chief of the Journal of Artificial Intelligence and Systems and Journal of Biological Sciences, as well as Associate Editor of the IEEE Journal of Biomedical and Health Informatics, Computers in Biology and Medicine, Frontiers in Cardiovascular Medicine, Computational Physiology and Medicine, Applied Soft Computing, IEEE Access, Frontiers in Communications and Networks, Computational Intelligence and Neuroscience, Measurement, IET Quantum Communication. Also, he has been Lead Guest Editor of several highly reputed journals and TPC member of many international conferences. Pethuru Raj Chelliah Chief Architect and Vice President, Reliance Jio Platforms Ltd., Avana Building, Sarjapur Road, Bangalore 560103, Karnataka, India Daytime Tel: +91 866 069 7438 Email: [email protected] LinkedIn: https://www.linkedin.com/in/sweetypeter/ Twitter: https://twitter.com/peterindia ResearchGate: https://www.researchgate.net/profile/Pethuru-Raj-Chelliah-2/2 Google Scholar: https://scholar.google.com/citations?user=yaDflpYAAAAJ&hl=en
Dr. Pethuru Raj is the chief architect in the Edge AI division of Reliance Jio Platforms Ltd., Bangalore, India. He focuses on emerging technologies such as the Internet of things (IoT), artificial intelligence (AI) model optimization, big and streaming data analytics, blockchain, digital twins, cloud native, serverless and edge computing, reliability engineering, microservices architecture (MSA), event-driven architecture (EDA), 5G/6G, and metaverse. He previously worked at IBM global Cloud Center of Excellence (CoE), Wipro Consulting Services (WCS), and Robert Bosch Corporate Research (CR). He has published over 30
https://doi.org/10.1515/9783111323749-205
XIV
About the editors
research papers in peer-reviewed journals, authored and edited 42 books, and contributed 54 book chapters. He is a professional member of Association of Computing Machinery (ACM). He received his PhD from Anna University, Chennai, India, and did his postdoctoral research in Computer Science and Automation (CSA) at the Indian Institute of Science, Bangalore, and at the Nagoya Institute of Technology, Kyoto University and at the University of Tsukuba, Japan. Dr. Satya Prakash Yadav GL Bajaj Institute of Technology and Management, Greater Noida, Plot No. 2, Knowledge Park, III, Greater Noida 201306, Distt. G. B. Nagar, Uttar Pradesh, India Federal Institute of Education, Science and Technology of Ceará, Brazil Contact email: [email protected] /[email protected] Phone: +917678647454 Dr. Satya Prakash Yadav is an Associate Professor in the Department of Computer Science and Engineering, GL Bajaj Institute of Technology and Management (GLBITM), Greater Noida, India. He has completed his Postdoctoral, under the supervision Prof. (Dr.) Victor Hugo C. de Albuquerque, from Federal Institute of Education, Science and Technology of Ceará, Brazil. He has been awarded his PhD degree from Dr. A.P.J. Abdul Kalam Technical University (AKTU) (formerly UPTU), Lucknow, India. Currently, four students are working for PhD under his guidance. A seasoned academician having more than 17 years of experience, he has published four books: Programming in C, Programming in C++, and Blockchain and Cryptocurrency, published by I.K. International Publishing House Pvt. Ltd., and Distributed Artificial Intelligence: A Modern Approach, published December 18, 2020, by CRC Press. He has undergone industrial training programs during which he was involved in live projects with companies in the areas of SAP, Railway Traffic Management Systems, and Visual Vehicles Counter and Classification (used in the metro rail network design). He is an alumnus of Netaji Subhas Institute of Technology (NSIT), Delhi University, India. A prolific writer, Dr. Satya Prakash Yadav has published two patents and authored many research papers in web of science indexed journals. Additionally, he has presented research papers at many conferences in the areas of Digital Image Processing, Feature Extraction, Information Retrieval, C, Data Structure, C++, C#, and Java. Also, he is the Editor-in-Chief in Journal of Cyber Security in Computer System (MAT journals); Series Editor at Bentham Science (UAE); and Lead Editor at CRC Press, Taylor and Francis (USA), Tech Science Press (Computer Systems Science and Engineering), Springer (Human-Centric Intelligent Systems), Science Publishing Group (USA), and Eureka Journals (India).
Mahesh Maurya, Vishal Gangadhar Puranik, A. Senthil Kumar, Balambigai Subramanian
1 Introduction to artificial intelligence Abstract: A subfield of computer science called artificial intelligence (AI) aims to build intelligent machines that can think and act like people. AI entails developing computer programs and algorithms with some autonomy in how they reason, behave, and learn. It is used in a number of sectors, including retail, robotics, healthcare, and finance. AI technologies have the potential to transform numerous sectors and improve the accuracy and efficiency of a wide range of tasks. AI can help automate and streamline processes, improve customer service, optimize pricing strategies, do data analysis, manage and predict inventory, detect fraud, optimize marketing campaigns, and more. Keywords: Artificial, intelligence, healthcare, finance, technology
1.1 Definition of artificial intelligence The science and engineering of creating intelligent machines, particularly intelligent computer programs, is known as artificial intelligence (AI). It is related to the more general field of computer science and deals with the development of algorithms, software, and peripheral technologies that are used to create intelligence in machines and to simulate intelligent behavior in robots and other electronic devices. Artificial neural networks (ANNs), robotics, machine learning, computer vision, and natural language processing (NLP) are just a few of the many technologies that go under the umbrella term “AI” [1].
1.2 Brief history and evolution of AI Beginning in the early 1950s, academics like Alan Turing and Marvin Minsky first started to investigate the potential of machines exhibiting intelligent behavior. The early days of AI focused on creating computer programs that could “think” like humans, but it was soon realized that machines had their own unique abilities. Over time, AI progressed through four distinct stages: the “good old-fashioned AI” (GOFAI) era, the “expert systems” phase, the “cognitive computing” period, and the current “deep learning” era. In the GOFAI era, the focus was on hand-programmed rules and logic to enable machines to perform complex tasks [2]. Expert systems emerged in the 1980s and marked a shift toward the use of stored knowledge to answer questions https://doi.org/10.1515/9783111323749-001
2
Mahesh Maurya et al.
and present options. Cognitive computing then developed in the 1990s to give machines the ability to self-learn by recognizing patterns and adjusting behaviors. In the 2010s, ANNs – which mimic the neurons and synapses of the human brain – emerged as the most sophisticated form of AI [3]. Deep learning has enabled machines to achieve superhuman levels of accuracy in various tasks, including speech and image recognition, language translation, and game playing (e.g., AlphaGo).
1.3 Importance and applications of AI in various fields AI is an important technology that can be used to solve complex problems across many industries. AI can autonomously analyze data and perform tasks that would otherwise need to be done manually, such as medical diagnosis, NLP, and autonomous vehicle navigation. AI is most commonly applied in areas such as robotics, automation, healthcare, customer service, e-commerce, media, finance, and agriculture. In robotics, AI is used to help robots acquire and improve precision, safety, and productivity and can be used to automate and control various machines. AI is also used to enable autonomous vehicle navigation and decision-making in real-world scenarios, such as offering directions in a city or driving across dangerous terrain. AI has become increasingly common in healthcare, offering rapid and accurate medical diagnosis, drug discovery, and drug testing. AI can be used to speed up medical testing and eliminate the need for manual work. AI can also be used to analyze medical records for diseases and treatments and alert healthcare providers when patients might need to be monitored more closely. Applications for customer service can also be powered by AI. Real-time customer assistance can be offered and inquiries can be answered by chatbots that are driven by AI. Alexa and Google Assistant are examples of AI-based automated assistants that can be used to organize appointments and make product suggestions [4]. AI can also be applied to e-commerce to identify fraudulent transactions and offer personalized recommendations to clients. The identification and classification of photos and videos in media also makes use of AI and machine learning (ML). AI can be used in the financial sector to spot fraud, streamline bank processes, and offer client guidance. AI can be used in agriculture to increase crop yields, detect weeds, and track livestock.
1.4 Taxonomy of AI AI taxonomy is a system for categorizing AI technology. ML, knowledge representation and reasoning, NLP, computer vision, robotics, and human-computer interaction
1 Introduction to artificial intelligence
3
(HCI) are some of the categories into which it is typically separated. Then, in order to provide a more thorough classification, each of these categories can be further broken down into subcategories. For instance, supervised learning, unsupervised learning, and reinforcement learning are three categories of ML [5]. Similarly, knowledge representation and reasoning can be divided into rule-based systems, expert systems, ontologies, and logic [6]. By classifying AI technologies in a taxonomic manner, it becomes easier to compare and contrast different AI technologies, identify which technologies work better for which tasks, and identify areas for improvement which is shown in Tables 1.1–1.5. Table 1.1: Classical AI versus modern AI. Classical AI
Modern AI
Classical AI is a type of AI based on logic and reasoning. It uses sets of rules or algorithms to solve problems in the same way a person with expertise in the topic might do.
Modern AI is a type of AI that uses more varied and complex methods to solve problems. It includes models such as deep learning, neural networks, symbolic programming, genetic algorithms, and machine learning.
Classical AI works by breaking down a problem and The solutions it learns and adjusts can be based on using data to attempt to formulate a response. changing data sets and feedback from previous tasks. This sort of AI is more independent and adaptive than classical AI. It often uses trial-and-error reasoning to figure out the best solution. This type of AI is useful for sorting, searching, language processing, and other tasks that require specific heuristics.
It is better equipped to handle difficult tasks like image identification, self-driving cars, and even natural language processing. Modern artificial intelligence is quickly taking over sectors like healthcare, finance, and e-commerce.
Table 1.2: Narrow AI versus general AI. Narrow AI
General AI
Weak AI, sometimes referred to as narrow AI, refers to artificial intelligence systems that have been created to do a very specific task without the assistance of a person.
Strong AI, sometimes referred to as general AI, is a type of artificial intelligence that can comprehend and reason about a variety of activities.
These AI systems use a limited set of data to solve a particular problem.
General AI has more advanced learning capabilities than Narrow AI, and is able to improve its performance based on progressive feedback.
Examples of narrow AI include machine learning (ML) algorithms, automated computer systems, and robots used to assemble a car or carry out an industrial process.
Some examples of general AI systems include facial recognition, natural language processing, virtual assistants, and self-driving cars.
4
Mahesh Maurya et al.
Table 1.3: Weak AI versus strong AI. Weak AI
Strong AI
Weak AI, sometimes referred to as narrow AI, is a type of artificial intelligence that is concentrated on carrying out particular tasks autonomously.
A type of artificial intelligence called strong AI, commonly referred to as artificial general intelligence (AGI), is able to learn, reason, plan, and solve problems in a way that is similar to how humans do. It is regarded as artificial intelligence technology‘s ultimate objective.
Weak AI systems are limited in their scope and are not capable of learning and adapting.
Strong AI requires advanced algorithms and computational power, and is still in its infancy in terms of development.
Examples of weak AI include image recognition, virtual personal assistants, medical diagnosis, and facial recognition technology.
Examples of strong AI include self-driving cars, intelligent robots, and computer vision.
Table 1.4: Symbolic AI versus sub-symbolic AI. Symbolic AI
Sub-symbolic AI
Symbolic AI works by representing knowledge in terms of symbols or objects, then using algorithms to manipulate those symbols and objects to “reason” out answers to problems.
Sub-symbolic AI does not rely on symbol manipulation, instead derives conclusions and answers from data analysis without explicit knowledge or evidence. Sub-symbolic AI does not use a “top-down” approach; instead, results are derived from “bottom-up” or inductive reasoning, based on analyzing data points and patterns.
Artificial intelligence that places a major emphasis on the interpretation and manipulation of symbols is referred to as symbolic AI, also known as “strong” or “good” AI.
Sub-symbolic AI, commonly referred to as “weak” or “bad” AI, is a classification of rtificial intelligence built on probabilistic algorithms and statistical models.
Table 1.5: Machine learning (ML) versus rule-based AI. Narrow AI
General AI
A branch of artificial intelligence known as machine learning (ML) gives computers the ability to learn from data, recognize patterns, and make choices on their own.
Rule-based artificial intelligence (AI) is a more traditional approach to AI and relies on a predetermined set of rules to allow computers to make decisions.
1 Introduction to artificial intelligence
5
Table 1.5 (continued) Narrow AI
General AI
It is based on the idea that a computer can learn from data and automatically create computer algorithms to identify patterns and predict outcomes.
Rule-based AI uses predefined sets of rules in order to generate an expected outcome, based on the available input data. This approach requires a lot of hand-coded rules and is less flexible than ML algorithms that can adapt to changing conditions by adapting to new data.
1.5 Research aspects of AI Research aspects of AI are the efforts of scientists, AI experts, and computer scientists to develop better methods and algorithms for AI applications. This research includes the study of existing literature, the development of new theories and algorithms, and experiments that test the efficacy of different methods. Research aspects of AI are aimed at improving the existing capabilities of AI applications, expanding their usefulness, and developing new ones altogether. These research efforts could include developing more powerful deep learning models, evaluating new architectures, advancing NLP, and exploring ways to create more intelligent AI agents.
1.5.1 Machine learning and deep learning algorithms ML is a branch of AI that focuses on creating methods and algorithms that let computers understand and use data to learn. The algorithms provide computers the ability to make decisions without explicit programming by allowing them to learn from the facts at their disposal and from their prior experiences. ML methods include, for instance, decision trees, neural networks, support vector machines, linear regression, logistic regression, and k-means clustering. The use of neural networks, on the other hand, is the main component of deep learning, a subset of ML that enables computers to learn from complex data [7]. Deep learning algorithms use a variety of methods, including hierarchical feature extraction and representation learning to help machines learn considerably more effectively. Convolutional neural networks, recurrent neural networks, and generative adversarial networks are a few examples of deep learning methods.
6
Mahesh Maurya et al.
1.5.2 Natural language processing (NLP) The AI technology known as NLP makes it easier for computers to comprehend and manipulate human language. This technology is widely employed in many different applications, including text summarization, speech recognition, machine translation, document classification, and question answering. In order to identify correlations between words and phrases, NLP technology analyzes the grammar, syntax, semantics, and structure of written or spoken language. NLP has many uses, especially in industries like healthcare, business, and education. In healthcare, for example, NLP is used for clinical decision-making, medical information retrieval, research, and personalized healthcare advice [8]. In business, NLP can be used to analyze customer comments and questions, generate customer feedback, analyze customer sentiment, recommend products, and identify sales trends. In education, NLP can be used to help automate assessment, improve search accuracy in libraries, and detect plagiarism in student work. Natural language understanding (NLU) and natural language generation (NLG) are two more technologies that are frequently used in conjunction with NLP. NLU assists in decoding the same input to determine the user’s intent and meaning, whereas NLG assists in creating natural-sounding words and sentences from the data input [9]. Together, these technologies allow computers to generate more accurate and meaningful results from their input. NLP is a powerful tool that can be used to process and understand natural language in a more efficient and accurate way. As more applications for NLP are developed, its applications are likely to expand to a variety of industries. This technology can be used to create smarter and more efficient applications that deliver more personalized experiences for users.
1.5.3 Computer vision and image recognition Computer vision and image recognition have become increasingly significant technologies in today’s society. Computer vision involves using software to recognize objects within images and videos that can be further processed by a computer. It is a method of “teaching” a computer to distinguish different objects within an image, such as a bike, a car, or a person. Image recognition helps computers to determine characteristics of a scene or object without any human input. By recognizing certain objects, robots can then complete related tasks without the need for human help. Recently, computer vision and image recognition technologies have seen a significant uptake within the automotive and medical industries due to their high precision and accuracy in detecting and understanding objects. In cars, computer vision and image recognition can be used for automated emergency braking, active cruise control, parking assistance, lane change assistance, and driver monitoring. In medicine, these technologies can be used to help quickly detect cancerous cells, precancerous cells, and any abnormalities in X-rays or MRIs. Besides, computer vision and image recognition tech-
1 Introduction to artificial intelligence
7
nologies are also widely used for facial recognition, security, augmented reality (AR)/ virtual reality (VR) applications, robotics control, autonomous drones, deep learning networks, and NLP. For example, facial recognition software uses image recognition technology to recognize facial features and accurately identify an individual. Additionally, AR and VR applications use image recognition to accurately sense and display overlay information [10]. Computer vision and image recognition are highly capable technologies that enable a machine to understand the nuances and characteristics of an image or video. These technologies have already been incorporated into many industries such as the automotive and medical industries, as well as facial recognition applications, security, and AR and VR applications [11]. As technology advances, computer vision and image recognition will continue to expand and develop, making these technologies increasingly more powerful and useful.
1.5.4 Robotics and autonomous systems The present focus of research and development in the field of machine automation and decision-making is on robotics and autonomous systems (RAS). The aim is to create machines that can perform tasks autonomously without the need for human input. RAS have revolutionized the way many complex tasks and processes are completed, allowing organizations to automate labor-intensive tasks and to free up resources for higher order activities. Research in this area has made great strides in recent years, and it continues to remain an area of significant investment. One of the key advancements made in the field of robotics was the introduction of autonomy. This refers to a machine’s capacity to function without human input. Decisions are made and tasks are completed based on the programmed logic, rather than by following instructions from a person. Autonomous robots can sense their environment, analyze it, and react to it in order to plan and adjusts its actions accordingly [12]. This has allowed for a wide range of robotics purposes, including surveillance, space exploration, service robots, manufacturing, and more. Another major breakthrough has been the development of AI, which allows robots to learn from their experiences and apply what they have learned in order to complete tasks. AI allows robots to develop better algorithms and methods to approach problems [13]. AI can also be combined with sensors, image recognition, and other capabilities to allow robots to complete more complex tasks. Robots are also becoming increasingly better at cooperating with people. Human-robot collaboration, in which robots and people can work together to achieve a goal, is becoming increasingly popular and is being used in many aspects of life [14]. This includes things like industrial robots used in the manufacturing sector, domestic robots used in the home, and medical robots used in the healthcare environment. RASs represent a remarkable technological achievement, and their possibilities are endless. They will continue to have a profound impact on the way tasks are performed and decisions made in the near future, and the opportunities are exciting.
8
Mahesh Maurya et al.
1.5.5 Expert systems and knowledge representation Expert systems, which are knowledge-based systems intended to mimic the decisionmaking of human experts and deliver solutions in the form of recommendations, models, and predictions, are one of the cutting-edge AI technologies. Expert systems store and manage knowledge in a very specific and predetermined representation, known as knowledge representation. Knowledge representation on expert systems is usually formalized and displayed through ontologies, taxonomies, and frames, which are structured according to the contextual data of the expert system. Ontologies are a type of knowledge representation on expert systems composed of concepts, relationships between them, and their properties that define and differentiate them from each other. The concept of ontologies differentiates itself from taxonomies in that it not only identifies hierarchical relationships between objects, but can also store the characteristics that differentiate each object from the others and represent that knowledge in the form of statements. An example of an ontology in expert systems is an ontology that represents the characterization and classification of the objects experienced by the system. For example, an expert system for stock market trading might have an ontology representing different types of stocks and their variation in parameters such as price, volume, and volatility. A type of knowledge representation very similar to ontologies is frames. Frames are graphical representations of objects that use slots, which imply fixed characteristics, and the amount of slots vary depending on the context. Frames are used in knowledge-based Systems because they provide a comprehensive representation of individual concepts or objects, covering different characteristics of them. An example of a frame within an expert system would be a frame representing a stock or bond, which contains information about the stocks price, volatility, yield, and market capitalization. Taxonomies are another type of knowledge representation used in expert systems. Taxonomies are a way of representing knowledge by categorizing objects into groups. The taxonomy of an expert system contain nodes which functions as a hierarchical structure that organizes concepts into categories based on their similarity or shared characteristics. An example of a taxonomy on an expert system could be a taxonomy that categorizes stocks by their industry or sector into groups such as energy, materials, and technology [15]. The expert systems use knowledge representations such as ontologies, taxonomies, and frames to efficiently organize and store information about objects experienced by the system. Each of these representations has distinct characteristics and is used to display different aspects of the knowledge stored by the system.
1.5.6 Reinforcement learning Machines and software agents can interact with their environment using reinforcement learning (RL), a type of AI to learn how to optimize their reward. It is a branch
1 Introduction to artificial intelligence
9
of ML and is an area of deep learning. It is an iterative process of learning that allows machines to take appropriate actions in an environment to achieve maximum reward. RL works by maximizing the expected future reward from any given situation. The machine’s action is expected to have an impact on the future, and to this end the machine will experiment with different actions or strategies and will gradually learn to make optimal decisions to maximize the future reward. Self-learning agents, which are at the core of RL algorithms, can change their behavior in response to rewards and penalties associated with prior choices, which enables them to grow more adept at navigating or engaging in strategy games. Robotics, finance, and marketing are just a few of the industries that can benefit from the use of RL. RL algorithms can be used in robotics to train robots to navigate unfamiliar or uncertain settings, while they can also be employed in finance to improve investment decisions [16]. The various fields in which RL can be used are wide and diverse. RL has a lot of potential applications, and it is being integrated into many different areas. As companies continue to explore the technology, we will likely see more and more uses for it. For instance, virtual agents, personalized robots, and automated services are just a few of the many possibilities that are now being explored.
1.5.7 Neural networks and artificial neural networks The two most popular techniques to ML and AI are neural networks and ANNs. Neural networks are predicated on the notion that the brain is made up of a significant number of interconnected neurons that cooperate to produce a desired result. ANNs are created to replicate the biological structure and behavior of the neurons found in the brain in computer code. A “back-propagation algorithm” is used in a neural network to find patterns in the data and change weights and other parameters to achieve the desired outcome. This outcome is often a prediction, such as the optimal course of action in a certain circumstance. An input layer, hidden layers, and an output layer make up a neural network. The data is processed using the nodes and connections between them, which produce outputs according on the parameters. A wide range of tasks can be carried out by ANNs. For example, ANNs are often used for face recognition, image classification, fraud detection and NLP. By making use of the vast amounts of data available, ANNs can learn a variety of tasks and can continue to learn and improve themselves over time. ANNs have a number of advantages over other ML algorithms. For example, ANNs have greater flexibility and can be adapted to the type of problems more easily. Additionally, ANNs can be trained more quickly and accurately than other algorithms. However, ANNs do have some drawbacks. Training and evaluating an ANN can require a great deal of information and layering in order to achieve accurate results. Additionally, ANNs can suffer from overfitting, which is when the algorithm fails to generalize and performs worse given new data. For ML and AI, neural networks and ANNs are effective and practical methods. They offer a
10
Mahesh Maurya et al.
fantastic approach to swiftly and accurately analyze data and generate results [17]. While they can suffer from some issues, such as overfitting, there are ways to avoid these problems. With the right training and evaluation, ANNs can provide an invaluable tool for many different tasks.
1.6 Challenges in AI The biggest challenge in AI is to develop artificially intelligent systems that can effectively combine the use of existing AI technologies with human knowledge and intuition to make decisions and solve problems. Additionally, AI must be able to operate in complex and unpredictable environments. Achieving these aims becomes more crucial as AI continues to permeate our daily lives, both for the benefit of the AI development community and for the society as a whole. One of the biggest AI challenges is creating environments that are easy for both machines and humans to understand and interact with. As AI progresses, developers need to not only develop smarter machines that are able to make decisions more effectively, but also create environments in which machines can easily parse and comprehend the abundance of data available today. In addition, AI must become better at understanding both spoken and written language, enabling it to interact with humans in more natural conversation. AI must also become better at reasoning and problem solving, as well as understanding and learning from humans. All these capabilities should ideally be combined with a selfregulating mechanism, enabling AI to improve itself.
1.6.1 Ethical considerations in AI development and deployment As AI continues to grow in complexity and effectiveness, a complex set of ethical considerations emerge. AI technology has the potential to revolutionize countless industries, medical practices, and even our daily lives. The implications for our future are tremendous, but these potential innovations come with ethical considerations. Since AI is being used to make ever-increasing decisions that require greater levels of trust, the need for ethical considerations in its development and deployment has never been more urgent. At the development level, programmers and engineers must consider how AI technology should be used responsibly and how to build with realism in mind. This includes ensuring the accuracy and accountability of data used for developing AI-based systems, as well as taking steps to protect user data from misuse and malicious abuse. Furthermore, AI developers must take into consideration the potential for bias in the development of AI-based applications. Algorithmic bias is an important issue, as algorithms and the data they use may reflect the biases of the developers, potentially leading to undesirable outcomes. At the deployment level, ethical considerations become
1 Introduction to artificial intelligence
11
even more important to consider. It is essential that businesses and organizations using AI have a set of guidelines and policies in place to ensure ethical data practices, create transparency and accountability throughout the AI process, and adhere to international regulatory frameworks. The effects of implemented AI-based systems on vulnerable individuals or populations must also be taken into account [18]. To ensure ethical use of AI, businesses should create and adhere to a set of values that decrease the potential for bias or other unethical outcomes. AI holds the potential for far-reaching innovation and progress, and its responsible development and deployment is of paramount concern. For AI to work in the long term, such considerations must be taken seriously and addressed directly.
1.6.2 Bias and fairness in AI algorithms Various areas, including healthcare, banking, and education use AI algorithms. They have the ability to completely alter how we live and work. They do, however, carry a chance of bias and unfairness. AI algorithms can replicate and even exacerbate preexisting biases in the data they are trained on since they are only as good as the input they are given. This may result in arbitrary decisions regarding loan applications, job applicants, and college admissions. In order to prevent bias from entering into AI algorithms, data needs to be examined and preprocessed to ensure it is structured in ways that do not inherently include bias. Data sets must be diversified to ensure the algorithm is exposed to a variety of experiences and perspectives. There must also be diversity in the people designing the algorithm and ethical considerations should be built into the system. The algorithmic transparency must be maintained. The methods of the algorithm should be easy to understand and available for scrutiny. Measures must be put in place to ensure responsible adjustment for any bias that is found in the decisions made. Finally, an open and continuous dialogue should be held between AI systems and stakeholders to ensure a sense of fairness and trust in the system. The potential of AI is immense and must be regulated to ensure that bias and fairness are accounted for. With proper training and oversight, AI algorithms can provide better and fairer decisions without sacrificing accuracy.
1.6.3 Privacy and data security concerns The proliferation and development of AI is leading to a new wave of innovation, automation, data privacy, and security concerns. AI’s potential for helping humans complete their daily tasks and providing life-saving services is clear. Yet the lack of transparency and the massive amounts of data collected also raise considerable data security and privacy issues that governments and organizations must address earnestly. Governments, organizations, and individuals are all responsible for ensuring the protection of data
12
Mahesh Maurya et al.
privacy and security. Governments must develop regulations and laws that protect individuals’ data from unwanted access and only allow verified and trusted organizations to access such data. Organizations are responsible for following regulations and best practices put in place by governments to ensure the highest level of data security and privacy. They must also make sure that data collected is used for the specific purpose it was collected for and that it is not exposed to unauthorized users. Last but not least, people need to be aware of how their data is being used and make sure that critical information like social security numbers and credit card information is not shared with anyone. The algorithms used to handle and interpret data are one of the main privacy and data security problems in AI. These algorithms can be used to find trends and make judgments about people’s behavior and traits, putting the privacy of people’s personal information at stake. Algorithms should be examined closely and comply with laws and regulations to ensure that no individual’s data is used without his/her permission [19]. While AI offers many potential benefits to society, governments, organizations, and individuals must take the necessary steps to ensure that data privacy and security are not compromised. Governments must put in place regulations and organizations must take proactive measures to make sure that all information collected is used responsibly and securely. Finally, individuals must be aware of their data usage and be vigilant against malicious actors who may be trying to take advantage of them.
1.6.4 Interpretability and explainability of AI systems As AI becomes more complicated, it is becoming more and more vital to worry about the interpretability and explainability of AI systems. Understanding how an AI system makes decisions is crucial for building trust in the system and its outcomes as complicated AI systems become more prevalent. The ability of humans to comprehend the inner workings and logic of AI systems is referred to as interpretability [20]. This is important for providing an understanding of the problems the AI system is trying to solve, and the reasoning and decisions it has made. Explainability, on the other hand, relates to the ability of AI systems to provide an explanation of their behavior. Explainability includes providing meaningful explanations of why a system behaves the way it does and should ideally provide insight into the reasoning and logic of the system. An essential topic of AI research is the creation of interpretable and explicable AI systems [18]. AI researchers will need to make use of computer-readable representations and comprehensible methods like symbolic and statistical AI, NLP, and ML in order to develop these systems. While AI systems are currently incredibly powerful, their lack of interpretability and explainability is preventing them from being used in many industries. By creating interpretable and explainable AI systems, humans will be able to understand and trust how these systems make decisions, paving the way for broader applications of AI.
1 Introduction to artificial intelligence
13
1.6.5 AI and job displacement AI has been developing quickly in recent years, enabling machines to take on an increasing number of tasks that were previously performed by people. Such developments have prompted a number of inquiries, the most important of which is whether and to what degree AI will result in job displacement. According to some analysts, AI may result in the loss of millions of jobs because many functions can be done by machines more effectively and cheaply than by humans. As AI becomes more advanced, redundant or low-skill occupations may be phased out, as machines become more viable in such roles. For example, customer service and administrative support roles are increasingly being automated with AI-driven machines and chatbots that provide basic customer service interactions. Other experts argue that while some jobs may disappear, AI will actually lead to the increase in new types of jobs. As AI proliferates, it will create an increasing demand for experts who understand how it works, and how it can be used to advance a business. Additionally, AI systems will need to be maintained, monitored, and kept updated, which requires a dedicated and knowledgeable workforce [21]. Still, it is certain that AI will leave an indelible impact on jobs and how works gets done in the future. It is therefore incumbent on governments and businesses to plan and prepare for the job displacement that is expected to occur. Governments must implement policies that encourage training and re-education for displaced workers, increasing their employability in the AI economy. Businesses need to take responsibility for providing current employees with the skills needed to succeed in the AI workplace. There is little doubt that AI has the ability to bring about employment displacement as well as significant growth and prosperity. By being agile and preparing for the changing job market, we can ensure we maximize the benefits and mitigate the risks of further advancing AI.
1.6.6 Regulation and policy challenges The development of AI systems is creating various policy and regulatory challenges for governments and organizations. As AI algorithms become increasingly complex and autonomous, policymakers must grapple with regulatory questions related to the design, use, and effects of the technology. This essay will examine the key regulatory and policy challenges posed by AI systems through an analysis of issues relating to privacy, liability, trust, security, and ethical considerations. Privacy is perhaps the most important regulatory concern posed by the increasing use of AI algorithms. As AI algorithms become more widely used, they rely on access to large amounts of data. People frequently submit this data, maybe unaware of the consequences of doing so when supplying it to AI systems. To ensure that individuals maintain control over their personal data, adequate privacy protections should be established [22]. This may involve constraint on the type of data that can be collected, limits on how long data can be stored, and limits on
14
Mahesh Maurya et al.
how personal data can be used. Liability is also an important regulatory issue in the context of AI algorithms. As algorithms become increasingly autonomous, it is possible that they will be held responsible for their own decisions or actions. This raises questions over who should be held liable in case of an algorithm failure or mistake. This may involve issues such as whether algorithm developers or users should be liable or whether liability should be assumed by the government. Trust is another important policy consideration when it comes to AI systems. Trust in AI algorithms is essential for their use, as the public must be able to rely on the accuracy and reliability of the decisions made by the algorithms. To ensure trust, it will be necessary for AI systems to be governed by a framework of accountability and transparency. This may involve developing technical standards for AI systems, as well as certification and audits for developers and users of the technology. Security is also an important regulatory concern, as AI algorithms have the potential to be used maliciously or for malicious ends. The security of AI systems must be assured through rigorous security measures. These may involve controls over the implementation and use of the algorithms, as well as measures to prevent malicious manipulation of the data used for training the algorithms. The ethical considerations are another key regulatory challenge posed by AI algorithms. AI algorithms must be designed in such a way that they respect fundamental ethical principles and values. This may involve, for example, ensuring that the algorithms are not used to discriminate against certain groups or that they adhere to appropriate privacy protections [23–25]. A variety of regulatory and policy concerns are brought about by AI systems. These include concerns with respect to privacy, liability, trust, security, and morality. To guarantee that AI algorithms are used safely, responsibly, and in a way that respects people’s rights and values, policymakers must address these issues.
1.6.7 AI and societal impact The popularity and advancement of AI have risen dramatically in recent years. AI describes a machine’s capacity to learn and reason similarly to a person. This technology has the potential to completely transform people’s lives all around the world and can be applied in a range of contexts, including industrial automation and medical diagnosis. Although AI may have certain advantages, it also contains some risks and has the potential to have a big impact on society. One of AI’s most significant effects is that it allows machines to perform tasks that were previously thought to be performed only by humans. As AI technology advances, machines can increasingly be used to replace human labor. This could lead to massive job losses and an increase in income inequality, for example. Moreover, AI machines may be vulnerable to attack from malicious actors, putting important data and systems at risk. Another major concern is the ethical implications of AI technology [26]. AI is being used to make decisions that affect human lives, such as medical diagnoses and loan approvals. The decisions made by AI systems may be based on biased or inaccurate data, creating ethical and social dilemmas. The
1 Introduction to artificial intelligence
15
potential for abuse and misuse of AI technology is also a major concern [27]. The effects of AI technology over the long run must be taken into account. AI machines have the ability to surpass human intelligence, which could result in technological advancements that are difficult to be foreseen. This could have a profound effect on human society, potentially leading to the displacement of millions of people, or the creation of new classes of people who possess great power and wealth [28]. The AI technology has a variety of potential impacts on society. It can be used to replace human labor, create ethical dilemmas, and potentially lead to technological advances that could drastically change the way people interact with each other. As AI technology continues to advance and become more prevalent, it is important to consider these potential societal impacts so that they can be better managed and addressed.
1.7 Conclusion AI has the potential to revolutionize the world with its numerous innovative applications. This technology has grown at a rapid rate and can be seen as a major stepping stone in the future of technology. AI is capable of problem-solving, decision-making, and finding subtle patterns in information that humans cannot detect. It is expected to have a wide range of applications in the future, ranging from healthcare, finance, transportation, security, and many more. Using NLP technology to help doctors better comprehend the problems their patients are facing is a significant application of AI in the healthcare industry. AI can help doctors make better judgments and deliver more individualized care by analyzing the talks between the patient and the doctor to gain insights into the patient’s mental health, exhaustion, or other difficulties. AI is also being used in the finance sector to evaluate data more effectively. Applications like fraud detection, risk analysis, portfolio management, and personal finance management fall under this category. Financial institutions may make judgments more rapidly by using AI to spot trends in data. AI is being used in the transportation industry as well to enable autonomous vehicles to drive safely and efficiently in any situation. This technology can detect and react to changes in the environment quickly and respond appropriately. AI can also be used to optimize traffic flow and public transportation around cities to reduce congestion. AI is also becoming essential to enhancing security. AI can be used in criminal identification, facial recognition, and military applications. AI can detect suspicious activity quickly and provide more accurate detection than humans are able to. AI is important because it enables us to process large amounts of data and gain insights that were not possible before. It can also provide solutions to problems faster and more accurately than humans can. We must keep innovating in this area if we are to gain from AI, which will play a significant role in our future. The question of how AI will affect humanity’s future is still up for debate. AI has already made significant breakthroughs in the world of technology and robot-
16
Mahesh Maurya et al.
ics, allowing us to automate and optimize processes and operations, as well as make predictions and automate decisions. Ultimately, the future of AI will depend on how it is used and applied. If AI is used responsibly, with a focus on creating positive and beneficial solutions for society, it can help build a better future for humanity. To bring responsible AI into action, governments, businesses, and AI developers need to collaborate, with an emphasis on transparency, accountability, and ethical principles. By regularly researching the use of AI in the real world and adjusting policies and regulations accordingly, we can build an AI-driven future that puts humans first.
1.8 Future work Many experts believe that AI will continue to be a driving force in the development of innovative technologies and processes that can help us improve our lives and businesses. The most pressing question that emerges from this discussion is whether AI will have a positive or negative effect on humanity in the future. There are legitimate concerns that AI could cause significant disruption in areas such as labor, healthcare, transportation, and even politics. On the other hand, there are a number of exciting possibilities that could come from harnessing the power of AI, such as greatly improving the quality and effectiveness of healthcare, understanding the complexity of the natural environment and the universe more deeply, and building a more prosperous and equitable future for humanity.
1.8.1 Emerging trends and advancements in AI The emerging trends and advancements in AI are revolutionizing the way we approach data and computation. In recent years, AI has progressed beyond traditional AI techniques, such as ML, to include deep learning algorithms, NLP, and robotics. AI is used for a variety of tasks from image recognition to voice recognition to robotic process automation. Deep learning algorithms debuted with great success, as it gave AI the ability to produce accurate results without needing vast amounts of training data. NLP saw success with its ability to parse out meaning from large quantities of text and audio data. Automation saw progress with the ability to automate mundane and repetitive tasks without human intervention. Today, AI is rapidly advancing to meet the growing demand for more efficient and effective data processing. AI-based software can identify patterns in large data sets, predict customer behavior, and even autonomously control robots. The rapidly developing technology of AI is also being used to create AR and VR experiences. AI is being applied to a wide variety of areas, ranging from healthcare to banking to retail. As AI continues to advance, the possibilities for its implementation will expand rapidly. The advancements in AI have made it
1 Introduction to artificial intelligence
17
easier than ever for businesses and organizations to access powerful data analysis tools. AI-driven insights can drive performance and productivity improvements, allowing organizations to discover previously hidden trends and insights. Cloud-based AI solutions are also becoming increasingly popular, giving organizations the ability to quickly implement AI solutions without significant upfront investment. The advancements in AI-related technologies are continually driving progress. AI-based robots are becoming smarter and more capable every day, and advancements in ML and NLP are making it easier for AI systems to understand their surroundings and interact with humans. In addition, AI-driven systems are increasingly able to replace humans in a variety of roles, driving efficiency and cost savings. As AI continues to evolve, its applications will become even more diverse. AI-driven solutions that automate mundane tasks have the potential to reduce human workload, while AI-driven insights can inform decision-making and improve results. Regardless of the industry or field, AI is sure to continue to drive innovation in the years to come.
1.8.2 Potential areas for future research and development The development of AI and autonomous robotics is already speeding up advances across various industries and fields of research. AI is enhancing capabilities and solutions in sectors like healthcare, banking, traffic control, automotive, and more. The potential to integrate AI into products and services will improve user experience and lead to a more efficient utilization of resources. Autonomous robotics is leading to a dramatic shift in industrial production. Already, autonomous machines are being used in areas such as agriculture, manufacturing, and construction for improved productivity and efficiency. The field of technology is rapidly advancing and the areas of potential research and development are vast. Some potential areas of research and development that stand out include further development of AI, autonomous systems, robotics, additive manufacturing, quantum computing, NLP, ML, biomedical technology, autonomous vehicles, robotics, and the Internet of things (IoT). In addition, development of new materials, advances in nanotechnology, and improved renewable energy production and storage could be beneficial for advancing various technological areas.
1.8.3 Impact of AI on different industries and society The emergence of AI in recent years has been the most significant technology development since the invention of the internet. AI is already having an incredible impact on multiple industries and is projected to revolutionize the way our society functions. AI’s potential for improvement and disruption is unprecedented as its applications are endless. The most direct and visible impact of AI is on industries, such as healthcare, fi-
18
Mahesh Maurya et al.
nance, and transportation. For example, AI has enabled medical breakthroughs, such as more accurate diagnoses and treatments, which have enabled the healthcare industry to move faster and more efficiently than ever before. AI has also had an impact on finance, with the development of algorithmic trading, which has enabled banks and asset managers to process trades at a much faster rate. Additionally, AI has been used to optimize transportation services, such as ride-sharing apps, to better manage fleets of vehicles and optimize route planning for drivers. AI is also transforming the way we interact with technology. Through AI-driven NLP technology, users can increasingly interact with machines in a more natural and intuitive manner. This technology has enabled more effective task automation and improved customer service, which is expected to play a major role in economic sectors such as retail, food services, and hospitality. AI is also likely to have a profound impact on society, potentially replacing human labor and ushering in a new era of automation. It is expected that much of the busy work involved in everyday life, such as the automation of manufacturing, will be replaced by AI. This automation is expected to reduce errors and improve the overall efficiency of organizations. In addition, AI-driven analytics are expected to revolutionize the way we interact with businesses, enabling consumers to access more accurate and personalized services. Ultimately, the impact of AI on industries and society is farreaching and potentially revolutionary. AI has already enabled major improvements in industries such as healthcare, finance, and transportation, and is expected to reshape the way we interact with technology and with each other. We are only just beginning to understand the potential of AI and the opportunities that it presents, but it is clear that it will have a significant impact on our lives in the years to come.
1.8.4 Ethical considerations and responsible AI practices In recent years, advances in AI have deeply impacted society. AI is used widely in a variety of fields such as biomedicine, law, education, and finance in which ethical considerations and responsible AI practices are of paramount importance. Ethical considerations are important in AI as the technology has the potential to affect individuals in significant ways. AI is often used for decision-making in areas such as healthcare and financial services. In such cases, detrimental effects may occur when AI-driven decisions are made with inaccurate data or without considering the diverse characteristics and values of individuals. Thus, it is important to review and audit the ethical considerations of AI-driven decisions. Responsible AI practices can ensure that AI is developed to consider ethical issues as well as human biases and values. In addition, responsible AI practices can be used to reduce potential risk associated with AIdriven decisions. For example, it is important to take into consideration the privacy of individuals when designing AI systems. Responsible AI practices can help ensure that data collected from individuals is used in a transparent and secure manner. Furthermore, AI systems used for decision-making should be designed to be explainable and
1 Introduction to artificial intelligence
19
free of algorithmic bias or other discriminatory effects. Ethical considerations and responsible AI practices should be an integral part of AI development. Ensuring that ethical considerations are taken into account when building AI systems and that they adhere to responsible AI practices will help ensure that the technology is used safely and without detriment to individuals.
References [1]
Ongsulee, P. (2017, November). Artificial intelligence, machine learning and deep learning. In 2017 15th International Conference on ICT and Knowledge Engineering (ICT&KE) (pp. 1–6). IEEE. [2] Norvig, P. R., & Intelligence, S. A. (2002). A modern approach. Upper Saddle River, NJ, USA: Prentice Hall; Rani, M., Nayak, R., & Vyas, O. P. (2015). An ontology-based adaptive personalized e-learning system, assisted by software agents on cloud storage. In Knowledge-Based Systems Vol. 90 (pp. 33–48). [3] Alimadadi, A., Aryal, S., Manandhar, I., Munroe, P. B., Joe, B., & Cheng, X. (2020). Artificial intelligence and machine learning to fight COVID-19. Physiological Genomics, 52(4), 200–202. [4] Awasthi, S., Naruka, M. S., Yadav, S. P., & De Albuquerque, V. H. C. (Eds.). (2023). AI and IoT-based intelligent health care, & sanitation. India: Bentham Science Publishers. [5] Brewka, G. (1996). Artificial intelligence – a modern approach by Stuart Russell and Peter Norvig, Prentice Hall. Series in Artificial Intelligence, Englewood Cliffs, NJ. The Knowledge Engineering Review, 11(1), 78–79. [6] Gerke, S., Babic, B., Evgeniou, T., & Cohen, I. G. (2020). The need for a system view to regulate artificial intelligence/machine learning-based software as medical device. NPJ Digital Medicine, 3(1), 53. [7] Chauhan, S. S., Yadav, S. P., Awashthi, S., & Naruka, M. S. (2023). Sentimental analysis using cloud dictionary and machine learning approach. In Cloud-based Intelligent Informative Engineering for Society 5.0 (pp. 157–169). India: Chapman and Hall/CRC. [8] Li, Y., Kumar, R., Lasecki, W. S., & Hilliges, O. (2020, April). Artificial intelligence for HCI: A modern approach. In Extended Abstracts of the 2020 CHI Conference on Human Factors in Computing Systems (pp. 1–8). [9] Hyder, Z., Siau, K., & Nah, F. (2019). Artificial intelligence, machine learning, and autonomous technologies in mining industry. Journal of Database Management (JDM), 30(2), 67–79. [10] Siau, K., & Wang, W. (2018). Building trust in artificial intelligence, machine learning, and robotics. Cutter Business Technology Journal, 31(2), 47–53. [11] Krishna, A. B., Tanveer, A., Bhagirath, P. V., & Gannepalli, A. (2020). Role of artificial intelligence in diagnostic oral pathology-A modern approach. Journal of Oral and Maxillofacial Pathology: JOMFP, 24(1), 152. [12] Das, S., Dey, A., Pal, A., & Roy, N. (2015). Applications of artificial intelligence in machine learning: Review and prospect. International Journal of Computer Applications, 115(9), 31–41. [13] Saravanan, V., Rajeshwari, S., & Jayashree, P. (2013). SECURITY issues in protecting computers and maintenance. Journal of Global Research in Computer Science, 4(1), 55–58. [14] Pollack, M. E. (1995). Artificial intelligence – A modern approach – A review. Ai Magazine, 16(3), 73–73. [15] Khanzode, K. C. A., & Sarode, R. D. (2020). Advantages and disadvantages of artificial intelligence and machine learning: A literature review. International Journal of Library and Information Science (IJLIS), 9(1), 3.
20
Mahesh Maurya et al.
[16] Woschank, M., Rauch, E., & Zsifkovits, H. (2020). A review of further directions for artificial intelligence, machine learning, and deep learning in smart logistics. Sustainability, 12(9), 3760. [17] Yadav, S. P., & Yadav, S. (2019). Mathematical implementation of fusion of medical images in continuous wavelet domain. Journal of Advanced Research in Dynamical and Control System, 10(10), 45–54. [18] Sultan, A. S., Elgharib, M. A., Tavares, T., Jessri, M., & Basile, J. R. (2020). The use of artificial intelligence, machine learning and deep learning in oncologic histopathology. Journal of Oral Pathology and Medicine, 49(9), 849–856. [19] Bini, S. A. (2018). Artificial intelligence, machine learning, deep learning, and cognitive computing: what do these terms mean and how will they impact health care?. The Journal of Arthroplasty, 33(8), 2358–2361. [20] Žarković, M., & Stojković, Z. (2017). Analysis of artificial intelligence expert systems for power transformer condition monitoring and diagnostics. Electric Power Systems Research, 149 125–136. [21] Asemi, A., Ko, A., & Nowkarizi, M. (2020). Intelligent libraries: a review on expert systems, artificial intelligence, and robot. Library Hi Tech, 39(2), 412–434. [22] Jakhar, D., & Kaur, I. (2020). Artificial intelligence, machine learning and deep learning: Definitions and differences. Clinical and Experimental Dermatology, 45(1), 131–132. [23] Jones, L. D., Golan, D., Hanna, S. A., & Ramachandran, M. (2018). Artificial intelligence, machine learning and the evolution of healthcare: A bright future or cause for concern? Bone and Joint Research, 7(3), 223–225. [24] Nithya, C., & Saravanan, V. (2018). A study of machine learning techniques in data mining. International Scientific Refereed Research Journal, 1 31–38. [25] Aruna, R. D., Surendran, S. D., Yuvaraj, N. D., & Debtera, B. (2022). An enhancement on convolutional artificial intelligent based diagnosis for skin disease using nanotechnology sensors. Computational Intelligence and Neuroscience, 2022, 1–13. [26] Silvia Priscila, S., Sathish Kumar, C., Manikandan, R., Yuvaraj, N., & Ramkumar, M. (2022, February). Interactive artificial neural network model for UX design. In Arulmurugan Ramu, Chow Chee Onn & M.G. Sumithra (Eds.). International Conference on Computing, Communication, Electrical and Biomedical Systems (pp. 277–284). Cham: Springer International Publishing. [27] Adams, J. A. (2001). Multiagent systems: A modern approach to distributed artificial intelligence. AI Magazine, 22(2), 105–105. [28] Ghahramani, Z. (2015). Probabilistic machine learning and artificial intelligence. Nature, 521(7553), 452–459.
Shaik Mohammad Rafee, M. Prasad, M. Sunil Kumar, Balamurugan Easwaran
2 AI technologies, tools, and industrial use cases Abstract: Artificial intelligence (AI) technology refers to various technologies and tools used in the development of systems that can carry out tasks that normally require human intelligence. AI systems can process large datasets and communicate in natural languages, taking decisions, recognizing objects, detecting anomalies, and managing data to provide personalized recommendations. AI technologies can be used in almost any field, including healthcare, finance, e-commerce, marketing, robotics, Internet of things, and more. Popular AI technologies and tools include computer vision, natural language processing (NLP), machine learning (ML), deep learning (DL), sentiment analysis, voice recognition, and robotics. Computer vision is used to identify objects in pictures or videos. NLP is used to understand language. ML is used for data modeling and to find patterns in data. DL techniques are used to process large amounts of data and automate decisions. Sentiment analysis is used to analyze customer feedback. Voice recognition is used to recognize and understand spoken words. Robotics technology is used to build and program robots. Businesses are increasingly using AI technologies and tools to improve customer experience, increase efficiency, automate tasks, increase scalability, and provide personalized recommendations and services. Keywords: Artificial, intelligence, healthcare, finance, technology
2.1 Overview of AI technologies and tools Artificial intelligence (AI) is a rapidly growing field of computer science that has the potential to revolutionize virtually every sector. AI technologies and tools are being applied to a wide range of ways, from diagnosing medical illnesses to increasing industrial productivity and to providing assistance in daily life tasks. AI technologies are divided into three main classifications: reasoning/inference, learning, and perception. Reasoning/inference involves the ability to draw logical conclusions from facts and evidence. AI systems can draw logical conclusions based on structured data, unstructured data, and even from input from a user [1]. This type of technology is used in computer vision, natural language processing (NLP), and robotics. Learning machines use algorithms to repeatedly analyze data and update their knowledge without relying on human assistance. Examples of AI applications in this domain include selfdriving cars, recommendation algorithms, and automated image and speech recognition. Perception involves the ability to observe and process real-world inputs in order https://doi.org/10.1515/9783111323749-002
22
Shaik Mohammad Rafee et al.
to make decisions or recommendations. AI enables machines to identify objects in a frame such as persons or items. It can also help machines detect and distinguish objects that are otherwise indistinguishable to humans such as a sound or vibration. Tools are how AI technologies can be integrated into existing systems. AI technologies are often used as part of larger systems, but can also be used on their own. Examples of AI tools used in this way include NLP tools, image recognition, and natural language generation (NLG) tools. AI technologies are becoming increasingly popular across many industries [2]. Many companies are investing heavily in the development of new technologies and tools that can help them generate insights, which can be used to improve their services and products. AI technologies have proven to be a powerful tool for organizations to become more efficient and productive while also providing customers and employees with more personalized experiences. AI technologies and tools are revolutionizing the way humans interact with computers, and are becoming increasingly popular across many industries. AI technologies can be used in a variety of ways such as developing reasoning and inference capabilities, learning from data, and improving perception. AI tools are widely used in various industries, from product development to customer service, and are continually evolving. As AI technologies continue to develop, new applications and tools may arise, allowing organizations to take advantage of this exciting technology’s potential [3].
2.2 Importance of AI in industrial applications AI has become increasingly important for many industrial applications in this day and age. AI has enabled the industrial sector to carry out complex tasks with greater precision and speed than ever before. AI helps increase safety in production and storage, and it is even used to automate complex processes. At a high level, AI can be used to optimize production and storage by reducing costs and increasing quality. AI can be used in various ways in the industrial field. For instance, AI algorithms can be used to detect and diagnose potential problems with machinery or other industrial processes. AI can also be used to control robots in automated manufacturing processes. AI-based solutions can help with both the planning and optimization of production processes. By taking into account a number of factors, such as cost, resources, environment, and regulations, AI-based solutions can come up with the optimal production and storage process. AI can be used to predict demand, and help with the inventory and supply chain processes. AI-based solutions are used in predictive maintenance to diagnose and solve potential problems before the machine or equipment ends up breaking down [4]. AI has become vital to various industrial processes. With its ability to help with planning, automation, optimization, prediction, and diagnosis, AI-based solutions provide an incredible amount of value to industrial sectors. AI is expected to play an even more significant role in the future, and will be an important factor in creating a more efficient and reliable industrial sector.
2 AI technologies, tools, and industrial use cases
23
2.3 Machine learning algorithms Machine learning (ML) algorithms are a set of programming techniques used to identify patterns and behavior within data sets. They are used to detect anomalies, classify data, and to make predictions. They are the basis of many AI applications, including computer vision, NLP, robotics, and gaming. ML algorithms are divided into two major categories: supervised learning and unsupervised learning. Supervised learning algorithms are given a dataset that contains labeled input and output, so they can learn from it [5]. Unsupervised learning algorithms are given a dataset that contains unlabeled data and they must create their own labels to decipher and identify patterns and behavior. Supervised learning algorithms are the most common type of ML algorithms and are used for predictive analytics. These algorithms include linear regression, logistic regression, support vector machines, and decision trees. Supervised learning algorithms can be used to predict values for new data that is fed to the algorithm. They can also be used to identify certain patterns within the data, such as identifying fraud or a customer’s purchasing habits [6]. Unsupervised learning algorithms can find structure in data that is not labeled, and are used for clustering analysis. These algorithms include k-means clustering, hierarchical clustering, and self-organizing maps. Unsupervised learning algorithms can be used for clustering and segmenting customers or to identify outliers or data points that do not fit with the rest of the data. Reinforcement learning (RL) algorithms are a type of ML algorithms that interact with their environment to identify the best action in order to maximize a specific reward. These algorithms can be used to train robots or autonomous vehicles, and are used for gaming. ML algorithms are constantly evolving as they are trained with new data. Algorithms are modified to become better and more accurate as they receive new data, and it is heavily dependent on the quality of the data it receives. Algorithms must be chosen carefully so that it suits the task at hand and can meet the expected results [7].
2.3.1 Supervised learning algorithms Supervised learning algorithms are algorithms that use labeled data to learn from. Labeled data is data that has been marked or identified in some way, making it easier for the algorithm to understand the patterns that exist in the data. Supervised learning algorithms can be used to solve a wide range of tasks, such as predicting stock market trends, determining the customer segments that are likely to purchase new products, or classifying images and other types of data. By using labeled data, supervised learning algorithms are able to build models that are better able to make
24
Shaik Mohammad Rafee et al.
predictions or decisions compared to algorithms that use only raw, unsegmented data. When creating supervised learning algorithms, data engineers will typically first split the data into two main categories: training and testing data. The training data is used to build the supervised learning model, while the testing data is used to evaluate its performance [8]. The goal is to create a model that performs accurately on both the training and testing datasets. Once a model is built, it can then be used to make predictions or decisions. This requires a careful evaluation of the model’s performance on both the training and testing datasets. If the model performs well on the training dataset, but not on the test dataset, then it is an indication that the model has not learned the underlying patterns in the data and needs to be refined or modified. Supervised learning algorithms are powerful tools used in data engineering to create models that can make accurate predictions or decisions. These algorithms make use of labeled data, which provides a way of understanding the underlying patterns in the data. Such algorithms are often used in predictive applications, such as stock market analysis, customer segmentation, and image recognition. By carefully evaluating the performance of the model on both the training and testing datasets, data engineers can ensure that the model accurately captures the underlying patterns in the data [9].
2.3.2 Unsupervised learning algorithms Unsupervised learning algorithms are an important class of ML algorithms used to explore and discover patterns in datasets without the use of any labels or output. Unlike supervised learning algorithms, which use labeled data points to learn how to classify data, unsupervised learning algorithms do not have any labels. This means they must find patterns automatically without any help. Unsupervised learning algorithms can be divided into three general categories: clustering, association, and anomaly detection [10]. – Clustering algorithms: Clustering algorithms are used to group together similar data points by discovering underlying patterns in the data. Clustering is useful for discovering natural divisions or groupings in a dataset. – Association algorithms: Association algorithms are used to discover associations between data points. Association rules finding can be used to uncover correlations in a dataset. These correlations can then be used to make predictions about new data points. – Anomaly detection algorithms: Anomaly detection algorithms are used to find data points that exhibit unusual or unexpected behavior. Anomaly detection is useful for finding fraudulent or unusual patterns that may be obscured or difficult to find.
2 AI technologies, tools, and industrial use cases
25
Unsupervised learning algorithms are useful for exploring data, uncovering hidden patterns, and making predictions. They can be used in many applications, such as customer segmentation, disease diagnosis, fraud detection, and more. As ML algorithms become increasingly sophisticated, the potential uses for unsupervised learning algorithms are expected to grow [11].
2.3.3 Reinforcement learning algorithms RL is a type of ML algorithm that enables artificial agents to learn from their environment and optimize their actions by trial and error. Unlike traditional supervised learning algorithms, RL algorithms learn from rewards and punishments, allowing the agent to learn more quickly and adapt to changing conditions. RL algorithms are composed of several components, including an agent, a learning environment, a reward function, and a set of actions. The agent learns by observing the environment and then deciding which action to take based on the rewards and punishments it receives. The environment provides feedback to the agent in the form of rewards or punishments, which influence the learning process. When the agent makes a “correct” decision, it is rewarded with a positive reward. If it makes a “wrong” decision, it is punished with a negative reward [12]. This reward system encourages the agent to learn a “correct” policy that maximizes the rewards it receives. The goal of RL algorithms is to maximize the long-term rewards an agent receives. As the agent takes actions, it accumulates rewards in a cumulative fashion so the longer it engages in the task, the higher the rewards it receives. Many different RL algorithms exist, each with a slightly different approach or a set of parameters. Examples of popular RL algorithms include Q-learning, SARSA, and evolutionary algorithms. No matter which algorithm is used, RL can provide powerful solutions to complex problems such as robotic navigation, finance, and games. It can also be used in simple problems such as in navigation, where the agent needs to find the most efficient path to an endpoint [13]. Regardless of the type of problem, RL algorithms are a powerful tool that can help agents to learn a “correct” policy, while maximizing rewards.
2.3.4 Deep learning algorithms Deep learning (DL) algorithms are a branch of AI algorithms modeled after the structure of the human brain. DL algorithms are composed of artificial neural networks and are capable of unsupervised learning, supervised learning, and RL. They are used mainly for tasks such as image recognition, NLP, and self-driving cars. The typical structure of a DL algorithm consists of several layers of processing units, which are usually called neurons. Each neuron within the network passes data to the nest neuron, and all of these neurons interact with each other and with the original input
26
Shaik Mohammad Rafee et al.
data. In each layer, the neurons connect to each other to form an artificial neural network. Optimizing the weights of the connections between the neurons allows the model to generalize useful patterns in data [14]. In some cases, DL algorithms are trained using millions of examples of data. For example, a DL algorithm, trained on a dataset of millions of images of cats and dogs, can be used for a task like image recognition. The algorithms build up more and more complex levels of information from the dataset until it can recognize and distinguish cats and dogs. DL algorithms are becoming increasingly powerful as the number of neurons and layers within the networks increase, and are now being used in many real-world applications, such as medical diagnosis, drug discovery, and automated language translation [15]. With the increasing availability of extremely large datasets, DL algorithms will become even more powerful and will eventually become a major part of our lives.
2.3.5 Ensemble learning algorithms Ensemble learning is a type of ML that combines multiple different algorithms to create a more powerful and accurate model. It is made up of multiple learning algorithms that are trained with the same training set, whose results are combined to give a better result. This method improves accuracy, precision, and generalization in ML models. Ensemble learning algorithms leverage the diversity of weak learners to create a more robust model. This diversity is typically achieved through several algorithms that look at different aspects of the data and come up with different outputs. The individual outputs are then combined, typically using a weighted average, to form the final result. By creating an ensemble learning system, we can get better prediction accuracy, even if some of the algorithms may be weak. For instance, a simple ensemble method might be to combine the output of two different algorithms into one prediction. The first algorithm might be a traditional supervised learning algorithm that uses a training set to create the output, while the second algorithm might use features generated from a clustering algorithm to create an output [16]. By combining the two algorithms, we can create a better prediction than when either algorithm was used solely. Ensemble learning can also be applied to more complex methods such as Neuroevolution, which uses evolutionary algorithms to evolve neural networks. By using multiple neural networks, with each learning different aspects of the data from the same training set, a more complex model can be created. As more algorithms are added and trained, the variance of the model decreases and the accuracy of results increases. The ensemble learning algorithms can be used to tackle difficult problems such as object recognition, where many different algorithms are needed to properly identify different types of objects. By combining the outputs of all of the algorithms into a single model, the accuracy of the model can increase and the object recognition rate can be improved. Ensemble learning is a powerful method of ML that combines multiple algorithms in order to improve accuracy, precision, and
2 AI technologies, tools, and industrial use cases
27
generalization. By leveraging the diversity of its weak models, it can create a more robust and accurate model. It is becoming increasingly popular and is now commonly used in a variety of applications such as object recognition and Neuroevolution.
2.4 AI development frameworks and libraries The development of AI technology has advanced significantly over the past decade, and many businesses are relying on AI to help automate and optimize their processes. AI frameworks and libraries are the backbone of any successful AI system, providing the necessary resources to build powerful predictive models and automate tasks. These frameworks and libraries provide tools for developers to create innovative solutions while making AI development more accessible to those with less technical experience [17]. There are various AI development frameworks and libraries to choose from, primarily depending on the type of AI project. Popular frameworks and libraries include TensorFlow, Google’s AI platform, Microsoft’s Cognitive Toolkit, and Torch. Each of these frameworks and libraries come with their own set of advantages and features that make them suitable for various applications. To choose the best option, it is important to consider the project’s scope and purpose as well as the expertise and resources available. TensorFlow is a powerful open-source framework developed by Google and is used for high-performance numerical computation, which is often used for ML and DL, particularly the latter. TensorFlow offers a wide array of features, including GPU acceleration, which is essential for building DL models. It also provides visualization and debugging tools, which are helpful during development and when monitoring performance [18]. Microsoft Cognitive Toolkit is a library from Microsoft, which was originally launched as CNTK and was then rebranded to accommodate its larger scope, beyond solely DL. Like TensorFlow, Cognitive Toolkit offers a range of features such as GPU acceleration as well as an advanced optimization engine that helps developers focus on training models quickly and with better results. Torch is an ML library that is built on the open-source Lua scripting language. It is designed to be user friendly, making it ideal for research experiments and for developers just starting out with AI development. Torch contains several features, such as an easy-to-use neural network library, an imperative python-like language for writing high-level code, and support for CUDA, which allows GPU acceleration. These frameworks and libraries are a great starting point for AI development. While each offers its own set of features and tools, keep in mind that there are many other options available, depending on the project. Moreover, pay close attention to the ongoing development of AI technology to ensure that the frameworks and libraries chosen stay up-to-date with the industry. By using the right AI development frameworks and libraries, businesses can reap the full potential of AI technology and create groundbreaking applications that can revolutionize the way they operate [19].
28
Shaik Mohammad Rafee et al.
2.4.1 TensorFlow TensorFlow is an open-source software library created by Google that helps developers to create ML models. It is a library for building, training, and testing models of neural networks in a way that is both fast and highly scalable. TensorFlow makes it possible for developers to create models that are dynamically updatable and can respond to real-time data. This makes it ideal for working with large datasets. The TensorFlow library contains an array of tools to create, train, optimize, and deploy neural networks. It also has an extensive library of pre-trained models that developers can use out-of-thebox. This makes it easier for those new to ML to quickly put together models and start iterating [20]. The library also features DL APIs such as Keras, TFLearn, and Sonnet. These APIs make it simpler for developers to write high-level ML code for a variety of tasks such as object recognition and NLP. TensorFlow also supports GPU acceleration, which is useful for large-scale DL tasks. GPUs can greatly speed up the training of neural networks and allow developers to quickly iterate on models in order to obtain higher accuracy. TensorFlow offers a user-friendly experience, making it easier for developers to create, develop, and deploy ML models. It also has a growing community of developers and users who share tips, tricks, and best practices [21]. TensorFlow is a powerful tool for anyone looking to create and work with ML models. Its easy-to-use APIs and wide variety of tools make it suitable for both new and experienced developers. With its GPU acceleration and user-friendly experience, TensorFlow is a great choice for those who need to quickly create and deploy models for large-scale projects.
2.4.2 PyTorch PyTorch is an open-source ML library for Python, used primarily for DL. It is developed by Facebook’s AI Research Group and is designed to be fast and flexible. PyTorch allows users to easily and quickly create and deploy DL models on a variety of devices, including GPUs, CPUs, and even mobiles. PyTorch supports automatic differentiation, a technique that allows the library to compute gradients and backpropagation. This means that users can easily and quickly train DL models, making it ideal for applications such as computer vision and NLP. PyTorch also offers native support for popular DL frameworks such as TensorFlow, Caffe2, and Theano. PyTorch also offers several advantages over other DL libraries. For example, it offers high performance with both CPUs and GPUs [22]. This allows users to quickly develop and deploy DL models to run on a variety of devices. Additionally, the library is designed to be userfriendly and easy to learn, making it suitable for both experienced and novice users. PyTorch offers excellent support for distributed training, thanks to its distributed package. This allows users to train models across multiple machines and GPUs, which can potentially help reduce training time. This is especially beneficial for training large models, which might otherwise take a long time to train on a single machine.
2 AI technologies, tools, and industrial use cases
29
PyTorch is an excellent DL library with plenty of features, advantages, and support. It is easy to use, allowing users to quickly create and train models, without needing to spend a lot of time learning the ins and outs of the library. This makes it a great choice for both experienced and novice users.
2.4.3 Keras Keras is an open-source DL library for Python. It is a high-level neural networks API that provides an easy-to-use and powerful interface for applications such as NLP, computer vision, and text analysis. The library was developed by François Chollet, a researcher in AI, and released in 2017. Keras is written in Python and is built on top of either Theano or TensorFlow. Its main advantages are flexibility and fast prototyping. Keras can run on both CPU and GPU. Its simple and intuitive interface allows users to quickly build and train powerful DL models [23]. Additionally, it is highly modular, allowing users to rapidly and easily construct complex model architectures. Keras is widely used to build DL models in a variety of fields, including computer vision, NLP, time series, and audio/video analysis. In addition, many popular pre-trained DL models are available in Keras, including InceptionV3, VGG19, MobileNet, ResNet50, and more. Users of Keras can benefit from a large and active community of developers as well as an extensive library of tools and tutorials available online. The development team works to ensure the library remains reliable, secure, and up-to-date with the latest research trends. With a wide range of applications and user-friendly interface, Keras has become an essential component for many DL projects.
2.4.4 Scikit-learn Scikit-learn is a widely-used open-source ML library for the Python programming language. The library provides a wide range of efficient and reliable algorithms for data mining, predictive analytics, classification, clustering, and regression, making it an essential tool for data scientists and ML professionals. Its user-friendly interface makes it easy to integrate into any existing Python code, and its wide range of features make it suitable for a variety of tasks. Scikit-learn includes many commonly used algorithms, such as linear and logistic regression, decision trees, k-nearest neighbors, random forests, and support vector machines (SVMs). It also offers a variety of built-in tools for preprocessing, feature engineering, and model selection [24]. With its extensive range of features, it allows users to quickly and easily build powerful models that are optimized for specific tasks. Furthermore, Scikit-learn also allows users to easily visualize results to help validate hypotheses and gain more insights. In addition to the many built-in algorithms, Scikit-learn also provides a number of powerful techniques for optimizing model parameters. For example, it uses cross-validation to ensure that
30
Shaik Mohammad Rafee et al.
the optimal parameters are identified, and that overfitting is avoided when building models. Moreover, its practical tutorials allow developers to quickly become familiar with the library and its various features. The Scikit-learn is an incredibly valuable tool for data scientists and ML professionals. It provides a wide range of robust algorithms, powerful optimization techniques, and intuitive visuals for gaining insights. And with its user-friendly interface, it makes ML tasks easier, faster, and more reliable.
2.4.5 Theano Theano is an open-source computing library for Python, developed at the LISA lab at the University of Montreal. Theano is designed specifically to handle complicated numerical calculations and to optimize code efficiently when multiple CPU cores and GPUs are available. Theano can be used for a wide range of applications, such as ML, computer vision, and NLP. At its core, Theano provides a basic set of functions and operations that handle complex mathematical problems and reduce problems to their most basic form. This enables ML models to run faster and more efficiently, as Theano performs computations for algorithms in parallel [25]. Theano also simplifies the development of DL networks by allowing network layers to be built quickly and easily. DL networks require complex calculations to determine the weights and biases of each layer. Theano can optimize these calculations and speed up learning. Theano also supports distributed computing, which means that it can be used on a range of different computers and machines to speed up computations even further. Other powerful DL libraries, such as TensorFlow, have adopted Theano’s distributed computing architecture for their own libraries. Theano offers debugging tools and data visualizations to aid in debugging, and help understand the model and its performance. This allows developers to ensure they are making progress in their DL models and provides them with useful insights into how the model is being trained. Theano is a powerful Python library for developing DL models, optimizing complex mathematical equations and performing distributed computing. It is an invaluable tool for any ML enthusiast.
2.4.6 Caffe The Caffe tool is an open-source AI software library, designed to support DL and deep RL models in ML applications. It is a framework for complex modeling and design of various DL models. Developed by a team at UC Berkeley’s AMPLab and funded by the Berkeley AI Research (BAIR) project, the Caffe tool is the brainchild of Professor Yann LeCun and Professor Berkeley AI Research Group and is maintained by the Berkeley Vision and Learning Center. The Caffe tool is written primarily in C++ and is made
2 AI technologies, tools, and industrial use cases
31
available under the BSD 2-Clause license. It consists of two libraries: Caffe and Caffe2. The Caffe library provides a general interface for DL networks. Through the use of this library, users can easily create, modify, and deploy neural networks without having to worry about manually coding up the model architecture and training algorithms [26]. Caffe2 is an extension to Caffe that includes a number of pre-trained models as well as supports distributed training. This makes it easier for users to experiment with various DL models and architectures. The library also provides a set of helpful APIs and tools to facilitate DL research and applications. Using Caffe, users can quickly identify objects and features from images and videos, and can also generate predictions about future data set. By making use of this library, researchers and developers can easily build and deploy complex neural networks and DL models.
2.4.7 MXNet MXNet is an open-source DL framework developed by a collective of researchers, engineers, universities, and other practitioners. It was first released in 2015 and has since had over 1,250 contributors from over 200 organizations. MXNet has been used for many applications, ranging from image recognition and object detection to NLP and time series analysis. MXNet is designed to offer a multidimensional array processing backend, allowing developers to write code that can be understood by the machine as well as by humans. This makes it easier to generalize models across multiple languages and backends. The framework is also extensible, allowing developers to create custom components to add to the existing library. This gives developers greater flexibility and control over how their models are built and deployed [27]. MXNet also offers several tools for developing and deploying large-scale projects. It includes libraries for DL, including autoML and Gluon, and a flexible library for training models. Other features, such as the MXNet Model Server, allow users to easily deploy trained models to any environment, such as the cloud, desktop, or embedded devices. With its many features, MXNet has become an essential tool for developers of DL and AI.
2.5 Natural language processing (NLP) tools NLP is a branch of AI, related to the interaction between computers and humans in a natural language. It mainly deals with the automatic analysis, understanding, and generation of human language such as English, French and Spanish. Through NLP, computers can understand various aspects of a language, such as syntactic, semantic, and pragmatics relationships. NLP is used in a variety of applications, including machine translation, automated question answering, information extraction, sentiment analysis, speech recognition, and automatic summarization. In order to better under-
32
Shaik Mohammad Rafee et al.
stand human language, NLP uses advanced tools such as tokenizers, part-of-speech taggers, and semantic role labelers. These tools are used to identify and then analyze the meanings of words and phrases in a given sentence. It also enables computers to understand the context in which a sentence is used. NLP also relies on algorithms such as support vector machines, decision trees, and neural networks for natural language understanding. These algorithms are trained using large amounts of data so that they can identify patterns and interpret natural language [28]. In addition, NLP also incorporates methods such as ontologies to organize complex data, and fuzzy logic to interpret various levels of ambiguity. NLP contains several powerful tools that make it possible to interact with computers in natural language. These tools include natural language parsing, text summarization, and text-to-speech conversion. Natural language parsing is used to identify and understand the meaning of a given sentence. Text summarization enables computers to understand and summarize long texts to a shorter form. Text-to-speech conversion is used to transform text into voice messages. NLP tools are becoming increasingly popular and are used in many areas such as AI, e-commerce, marketing, government, and healthcare. Financial institutions, in particular, have largely benefited from the use of this technology in the form of automated customer service, fraud prevention, and credit risk assessment. The development of NLP tools has enabled automated customer service to answer customer queries quickly and accurately. Fraud prevention services detect unusual behavior and prevent fraudulent transactions. NLP tools represent a revolutionary advancement in the field of AI. Automation has enabled companies to better understand the language of their customers, detect fraud, and manage risk. NLP tools have been used to develop better customer service, research, and analysis. It is no wonder that this technology is revolutionizing the way many businesses interact with their customers.
2.5.1 Tokenization and text preprocessing Tokenization and text preprocessing are essential components of NLP tools. Tokenization is the process of dividing a large body of text into smaller tokens such as words, subwords, characters, phrases, and sentences. It is the fundamental step for any NLP task as it makes the text ready to be processed by further algorithms. Text preprocessing consists of several operations to clean the collected text data and make it suitable for the application. These operations include removing special characters, whitespaces, HTML tags, stop words, non-English words, punctuation marks, etc. It depends on the application and the use case of the NLP model for which the text preprocessing step may be added, reduced, or skipped. Tokenization is a critical step as incorrect tokenization can lead to incorrect results; thus, special attention should be paid to this step. After tokenization, preprocessing operations help in reducing the noise present in the text and make it more suitable for further processing [29]. They can be used to normalize the text, convert words and abbreviations, lemmatize and stem the
2 AI technologies, tools, and industrial use cases
33
words, etc. Corrupt or incomplete text data causes problems in training and evaluation of the NLP model; therefore, it is important to clean them before training, to get accurate results. Both tokenization and text preprocessing are important NLP tools with many of its applications such as text classification, sentiment analysis, text summarization, machine translation, etc. They help in optimizing the data by removing unnecessary information and make it more suitable for NLP-related tasks. Therefore, they are often used as essential preprocessing steps in NLP models.
2.5.2 Named Entity Recognition (NER) Named entity recognition (NER), also known as entity extraction, is an important element of NLP that identifies and categorizes specific items mentioned in text-based documents, such as organizations, locations, time mentions, and quantity mentions. This information can be used to better understand context, text relationships, and search query intent. Examples of NER machines would include search engines categorizing names, topics, products, and more in order to provide more informative search results. NER has a long history and has been studied by many NLP researchers. In the early days, researchers had to manually annotate data and apply rule-based systems in order to identify certain entities. Modern approaches to NER are mostly focused on DL-based solutions such as recurrent neural networks, long short-term memory networks, and convolutional neural networks as well as legacy-style systems such as support vector machines (SVM). Generally, these solutions can learn from annotated data that had been previously collected or labeled, and can extract information from unlabeled data [30]. NER is mainly used to extract names, dates, email addresses, locations, etc. from text. It is becoming increasingly popular in a variety of applications such as text summarization, opinion mining, and question answering. For instance, NER can be used to extract the names of people and organizations from news articles in order to summarize the article in a brief manner. Additionally, NER can be used to build retargeting models and can aid in option mining in order to identify what people are saying about products or services. NER is a valuable technology that is constantly being improved and refined in order to provide better accuracy and more efficient solutions. It has been used by various industries such as healthcare, finance, and security, and is always making new advances and helping facilitate the processing of text in a wide range of NLP applications.
2.5.3 Sentiment analysis Sentiment analysis is the process of determining the emotional tone behind a series of words and is used to gain an understanding of the attitudes, opinions, and emotions expressed within an online mention. It is commonly used to measure the sentiment of
34
Shaik Mohammad Rafee et al.
customer feedback, reviews, tweets, and other social media content in order to gain insights into customer experiences. It can also be used to gain insights into what people think about products, services, and more. This type of analysis enables businesses to track the level of customer satisfaction and identify areas for improvement [31]. By analyzing sentiment, organizations can better understand how customers perceive their offerings and adjust their strategies in order to maximize customer satisfaction.
2.5.4 Text classification and clustering Text classification and clustering are two of the most popular methods of text mining. Text classification is the process of assigning categories or labels to written documents, based on their content. It is used to enhance the accuracy of web search engines, categorize documents, and to cluster similar documents together. Text clustering is the process of assigning clusters of similar documents together. It is a mixture of unsupervised and supervised ML algorithms. The main goal of text clustering is to identify and categorize data points, usually words and sentences, into meaningful groups. This is an important tool for analyzing large amounts of data with minimal effort. Clustering can help reveal patterns and trends in documents as well as identify groups of related words. Clustering also helps to reduce the noise and sparsity of the data, and allows researchers to more easily identify the underlying structure of documents. Both methods are used for understanding and analyzing textual data. The use of text classification and clustering determines the similarities and patterns of each text document. With text classification, the goal is to assign labels and categories to documents and groups of documents, while with text clustering, the goal is to group together documents that are similar in content. Text classification and clustering are both powerful methods for understanding and analyzing textual data in both computational linguistics and NLP.
2.5.5 Language modeling and generation Language modeling and generation is a key component of NLP tools. It is a complex set of tasks covering many aspects of the natural language data processing. A model is created to capture the underlying structure of the data, and then used to generate predictions and responses. The process of language modeling and generation starts with data capture, which involves gathering the data in a suitable format. This could include text documents, emails, webpages, audio recordings, and more. This data is then processed, usually through a set of algorithms, to extract the relevant features. These features are then used to create the model for the language. The model can take the form of a statistical model, a knowledge graph, a network model, or other types. Once the model is created, the next step is to use it for prediction of new language data. This can be done through various methods such as NLP algorithms, ML
2 AI technologies, tools, and industrial use cases
35
algorithms, and other statistical techniques. The data is then used in the application of language modeling and generation tasks, such as language translation, question answering, summarization, and summarization detection. There are several types of language models that can be used for language modeling and generation. The most common are RNN, CNN, LSTM, and GRU networks. RNNs are the most basic type of recurrent neural networks that use temporal information to capture the structure of the language. CNN models are used for detecting local patterns in text and for sequence-to-sequence learning tasks. LSTM models, on the other hand, add a memory factor to perform tasks such as text classification and sentence completion. Finally, GRU networks are the most advanced type of language modeling and generation networks, and they can perform tasks such as text summarization and text generation. The data used in the language modeling and generation process can be varied. It can include text documents, emails, webpages, audio recordings, and even images. This is important because it ensures that the model can accurately capture the structure of the data and generate responses.
2.5.6 Machine translation Machine translation is a type of NLP tool that enables automated translation between two languages. It is used mainly in commercial applications and often provides a way of translating large amounts of content quickly and accurately. Machine translation is useful in a variety of areas, such as translating documents, web content, and conversation. Businesses and organizations abroad can translate customer and employee emails or texts, helpdesk requests, product descriptions, and other web content quickly and precisely; and research organizations can use machine translation for accurate information gathering from large numbers of local documents. One of the key advantages of machine translation is increased speed, as software can handle large quantities of content quickly in comparison to manual translation. It can translate hundreds of documents in a fraction of the time it would take human translators, meaning businesses can serve their customers more efficiently and keep up with the competition. However, machine translation is still limited by the language used and context in which it is being used. Machines do not have the ability to understand the nuances of language in the same way humans can, so it is important to ensure you use a reputable provider of machine translation, and make sure the scope and context of the documents is fully understood. Machine translation is a valuable tool for companies and organizations that need to quickly and accurately translate large amounts of content and documents in multiple languages. While machine translation does involve some limitations, the benefits outweigh the costs and the resulting speed and accuracy can improve customer service, market efficiency, and growth.
36
Shaik Mohammad Rafee et al.
2.6 Computer vision and image recognition tools Computer vision and image recognition are two of the most powerful tools in the field of technology today. The ability to use these tools to make sense of and interpret the world around us has led to the development of many different applications. Computer vision is used to analyze digital images and video to detect objects, identify the objects, and describe them with accurate measurements. It is also used to identify patterns in an image, such as facial features, and to determine motion from one frame to the next. The technology utilizes algorithms and ML methods to extract the most relevant features and characteristics of an object so that it can be identified. Image recognition or image classification is the process of automatically identifying objects in an image. The most common instance of image recognition is facial recognition, where a system is able to recognize and identify a person’s face from a digital image. Image classification can also be used to classify images into different categories such as landscapes, animals, and plants. This type of recognition is useful in applications such as search engines, medical imaging, and security systems. Both of these technologies have seen incredible growth in recent years, with advances in AI and big data leading to increased accuracy and speed of analysis. There are numerous applications for these tools, ranging from self-driving cars and visual search engines to medical imaging and security systems. The potential for these technologies is tremendous, and in the future, as these technologies become more widely adopted, we can expect to see them in use everywhere, from analyzing satellite images to helping detect cancer. With the help of computer vision and image recognition tools, we will be able to make better and more informed decisions, and open the doors to more capable and efficient services everywhere.
2.6.1 Image preprocessing and augmentation Image preprocessing and augmentation are two key steps in the field of computer vision. Preprocessing typically refers to a set of techniques used to prepare an image for further processing, such as feature extraction, object detection, or content classification. Image augmentation, on the other hand, is a process of creating a new version of an existing image by applying certain procedures such as shifting, flipping, or resizing. Preprocessing involves applying mathematical algorithms and other modifications to an image with the aim of improving its performance in further stages of the imaging process. Preprocessing often includes correcting for any anomalies in the image, adjusting color balance, sharpening edges, and increasing contrast or saturation. This is done to make the image more accurate and consistent, and improve its visual appeal. Augmentation, in contrast, is the process of creating new images from existing ones. By applying techniques such as shifting, scaling, zooming, or flipping, new images can be created from existing images or without any distortion or impact
2 AI technologies, tools, and industrial use cases
37
to the original image. This process is widely used in ML and DL to increase the quantity of training data sets as well as to increase the diversity of each set. Image preprocessing and augmentation are essential steps for any imaging project. The former allows for minor adjustments to existing content, while the latter stimulates the creation of new image content. By properly executing these two stages, images can be improved or expanded in a relatively easy and cost-effective manner.
2.6.2 Object detection and tracking Object detection and tracking are two important types of computer vision tasks that have been widely adopted in a variety of applications, such as surveillance, automotive, robotics, and entertainment. Object detection is the process of locating and recognizing the presence of objects in an image or video. This could be as simple as detecting faces or identifying objects present in the scene. On the other hand, tracking involves the process of locating and tracking objects in a video or image sequence, such as following a person in a video or tracking the movement of a particular object. Object detection and tracking have become increasingly popular over the years as they improve the accuracy and efficiency of computer vision applications. Object detection algorithms allow for improved accuracy in recognizing the presence and type of objects in video and images, which can then be used for tracking that object’s movements. Similarly, object tracking algorithms enable systems to precisely estimate the position of an object and follow it over time. Object detection and tracking have numerous applications, including security surveillance for detecting movement in a restricted area or recognizing faces in a crowd. Tracking algorithms can be used to track the movement of cars in real time, which are important for developing autonomous vehicles. Additionally, object detection and tracking are utilized in entertainment applications, such as augmented reality, gaming, and virtual reality, where users need to interact with virtual objects in the environment. Overall, object detection and tracking are two important computer vision tasks that have led to countless advances in robotics, automotive engineering, security surveillance, and entertainment. As the development of object detection and tracking algorithms improve, their applications will only continue to grow.
2.6.3 Image classification and segmentation Image classification and segmentation are important areas of computer vision since they are used to further understand the content of images. Along with detection, these methods form the cornerstones of machine vision, which in turn serves as the foundation for ML. Image classification is the process of assigning a label to an image, based on the content of the image. This is done by either applying a set of predefined
38
Shaik Mohammad Rafee et al.
thresholds or by training an ML model. In either case, the model is trained to recognize patterns in the images and assign labels based on their similarity. Image segmentation is the process of partitioning an image into different segments or regions. This is usually done by identifying objects and their boundaries in the image. For example, in an image of a person, segmentation can be used to identify the person’s head, torso, legs, and arms. By segmenting an image into different regions, it is possible to identify different elements in a scene, such as objects, background, and boundaries. Together, image classification and segmentation provide powerful tools for analyzing images and understanding their contents, which is the basis for many applications such as object recognition, facial recognition, and autonomous navigation. By harnessing these methods, it is possible to create computer vision systems that are capable of understanding and interpreting images, just like humans do.
2.6.4 Facial recognition Facial recognition technology is one of the most important emerging technologies of the twenty-first century. The technology uses special computer algorithms to identify a person from their facial features and is incredibly accurate in identifying a person in almost any situation. The main purpose of facial recognition is to identify potential security threats by detecting faces of people who have been flagged by the intelligence community or law enforcement. It can also be used for identifying potential criminals and those on watch lists. For example, it has been used to identify people in surveillance footage and in passport control or security protocols. Facial recognition has also been used in some retail stores for purposes such as tracking customer loyalty and dealing with shoplifting. In addition, it can be used in medical research to diagnose and investigate various diseases as well as helping to identify people who may have Alzheimer’s. Facial recognition technology is not without its issues. The primary concern is that it can be used to infringe the basic civil liberties of individuals, with some privacy experts arguing that there is no need for this type of technology to be used on the general population. In addition, there are ethical questions raised around its use on vulnerable individuals, such as children or those with mental illnesses. Facial recognition technology has both potential benefits and risks. It should certainly be used carefully and only when absolutely necessary to protect the privacy of individuals. In the future, it is likely that safety and security departments will increase their use of facial recognition technology, while at the same time developing measures to ensure that ethical issues are still adequately addressed.
2 AI technologies, tools, and industrial use cases
39
2.6.5 Image generation and style transfer The advent of image generation and style transfer technology has revolutionized the way we interact with digital media. Image generation refers to the process of creating a unique, synthetic image from computer-generated data. Style transfer is a process in which the “style” – or visual elements – of one image is applied to the content of another image, creating a unique effect. In the past, generating a synthetic image would have required a skilled artist or graphic designer. Through the use of computer algorithms, however, an AI system is able to do the work of many people – in the fraction of the time. This is because, unlike humans, computers are able to use a large number of preexisting inputs to generate new images. Style transfer algorithms are also able to create new and unique images from existing ones. One common style transfer algorithm is called “style transfer networks.” These algorithms take a preexisting image or “style” and blend it with the content of another image to create a new, hybrid image. For example, a style transfer network may take the style of a Monet painting and apply it to a photograph of a castle, creating a completely new image with both elements from the original two. Image generation and style transfer technology have a variety of applications, from marketing and advertising to virtual reality and gaming. By adding variety and creativity to otherwise mundane images, this technology can help bring ideas and products to life, enhancing visual storytelling and providing an immersive experience for viewers. It can also be used to restore antiquated art, allowing it to be digitally transformed and preserved for future generations to enjoy. The implications of image generation and style transfer technology are numerous and profound. By leveraging the power of computer algorithms, designers and visual artists can focus more of their efforts on artistic expression and creativity. In turn, these technologies can help us explore different cultures, tell unique stories, and reimagine the world around us.
2.6.6 Video analysis and understanding Video analysis and understanding focuses on techniques to analyze and gain insight from different types of videos. It covers a variety of methods, both qualitative and quantitative, including visual analysis, facial recognition, sound analysis, and object detection. The analysis begins with a study of the video. This can include aspects such as frame rate, brightness, contrast, focus, and composition. Other aspects of the video that can be analyzed are sound, color, and motion. Audio can be analyzed to determine if there are any repeating patterns or know the types of sounds being used. Color analysis can help to identify objects and people in scenes as well as provide information about the time and place being filmed. Motion analysis can provide insight into how people or objects are moving within the scene. Once the video is analyzed, the data can be used to build visual models. This involves creating a 3D model of the
40
Shaik Mohammad Rafee et al.
scene to show how people and objects interact in the environment. The models can be used to create animations that better explain the data. The models can also be used to track people and objects or to track the changes in the environment over time. The video analysis and understanding can also help with facial recognition and object detection. Facial recognition can be used to identify a person from a certain group of people or to verify the identity of a person. Object detection can be used to track objects in the scene. This is useful in cases such as facial recognition, where the same person could have multiple appearances. Video analysis and understanding is an important tool for researchers and businesses to gain new insight from videos. It can be used to identify trends, detect anomalies, and guide decision-making. It also helps to protect against fraud, since videos can be used to detect unusual behaviors or activities. By using the techniques of video analysis and understanding, researchers and businesses can gain new insights from videos and create better strategies for the future.
2.7 Robotics and autonomous systems Robotics and Autonomous Systems (RAS) are becoming increasingly important in the modern world. In recent years, robots and machines have been used to perform a variety of tasks from highly precise manufacturing to mundane labor. Autonomous systems are also becoming a powerful tool in areas like medical diagnostics, space exploration, security, and natural disaster response. Robots are allowing us to do things that were once thought impossible. For example, many robots are now used in surgery to perform delicate and precise movements. Additionally, robots are being used in factories to reduce costs and increase efficiency. As robots become more advanced, they will be able to take on more tasks with greater precision and accuracy, reducing waste, labor, and errors. Autonomous systems are also increasingly popular. Autonomous systems such as self-driving cars and drones are now capable of navigating complex environments and reaching destinations without any human intervention. This is a major breakthrough for transport systems and is expected to reduce accidents and emissions. Autonomous systems are also being used in agriculture to monitor and control processes such as crop health and yield. In the future, robots and autonomous systems will continue to play a major role in our lives. Self-driving cars will become more common, and more sophisticated tasks, like self-tidying homes, may become reality. Autonomous systems will also become more involved in disaster relief, search and rescue, medical diagnostics, and in other areas of exploration and research. We can look forward to increased efficiency, fewer errors, and more safety. Robotics and autonomous systems are playing an ever-growing role in our lives. This technology is revolutionizing how we live, work, and interact with each other. From
2 AI technologies, tools, and industrial use cases
41
space exploration to medical diagnostics, these systems will continue to be beneficial and necessary for the advancement of humanity.
2.7.1 AI in robotics and automation The integration of AI into robotics and automation is revolutionizing the production industry, and will be key to the future of many industries. AI-driven technologies, such as robotics and automation, have already proven to be a valuable asset in many different sectors. The increased use of AI in robotics and automation allows for a more efficient and accurate production process. Robots and automated systems have already found a number of different applications, ranging from assembly lines to hazardous waste management. Autonomous robots are able to quickly and efficiently process and sort, assemble, and move parts and components in a controlled manner, drastically increasing productivity and precision, while simultaneously reducing labor costs. AI-enabled robots can also carry out a myriad of other tasks through voice commands and have the ability to learn from their experiences. This learning capability is creating unprecedented opportunities for the use of robots and simple automation systems to perform more complex tasks and to operate in unpredictable and unfamiliar environments. AIenabled robots and automation systems also promise to create greater efficiency and safety at the workplace. In addition to reducing the need for manual labor in hazardous and difficult-to-access environments, AI-enabled automation also reduces the risk of human error. Furthermore, systems designed with AI are capable of monitoring their processes in real time and providing feedback to the system operator, allowing for greater oversight and accuracy. The introduction of AI in robotics and automation also enables a new level of customization to the production process. AI-enabled systems have the ability to recognize patterns, learn from experience, and respond accordingly. This adaptation and customization makes robots and automation systems more resilient and able to accommodate changes in the output requirements and demands in real time. The integration of AI into robotics and automation is already having a profound impact on the production industry. By reducing the dependence on labor-intensive processes, AI-enabled systems provide greater efficiency, accuracy, and safety, while simultaneously reducing production costs. Through their ability to adapt to changes and provide real-time feedback, AI-enabled robots and automation systems are also unlocking new opportunities for customizing production processes to meet ever-changing customer demands.
2.7.2 Perception and sensing technologies Perception and sensing technologies are the foundation of any modern technological infrastructure, and can be used to improve the efficiency, safety, and effectiveness of
42
Shaik Mohammad Rafee et al.
any system. They allow systems to detect environmental data, monitor activities, interact with users, and identify user behavior. When these technologies are used to their maximum potential, they can provide immense benefits. The technology works by capturing data through any number of sensors, whether they are sound, vision, or motion. This data can be used to create a representation of the reality at hand, which can then be further analyzed to determine patterns of behavior or identify anomalies. This allows systems to quickly and accurately identify emerging issues, and respond appropriately. For example, if a system can detect motion, it may be able to detect suspicious behavior or the presence of a security threat. On the other hand, cameras can be used to quickly scan an area to ensure that the environment is safe and secure. This data can also be used to identify unusual events or behaviors, alerting users to potential issues before they become serious [32]. Perception and sensing technologies can also be used to engage with users and influence their behavior. By providing personalized feedback and interactive experiences, companies can influence consumer behavior in ways that can have a significant impact on their bottom line. By monitoring user behavior, companies can detect user preferences and habits, and tailor their products and services accordingly. Perception and sensing technologies are quickly becoming the norm in modern technology. They are being integrated into more and more systems, with the potential to significantly enhance security, safety, and efficiency. As these technologies continue to develop, they will likely become a cornerstone of many industries and applications.
2.7.3 Motion planning and control Motion planning and control is an important aspect of robotic engineering and design. These two fields are concerned with how robots move and interact with their environment, and how to control their movements. Motion plans are created in order to tell a robot how to maneuver in a given environment, with the least amount of time and energy. This involves the use of various algorithms, sensors, and feedback from the environment. Control is then applied to ensure proper execution of the motion plans. Motion planning is a complex process that relies on complex algorithms. It has evolved over time, from the development of models and simulations to the use of ML and AI. In the past, motion planning was done manually with trial and error. However, researchers have developed more efficient algorithmic approaches that enable robots to craft plans based on observational data from their environment. Control is the mechanism for executing motion plans. It incorporates feedback from sensors and the environment to ensure that the robot behaves as desired. This approach can be used to process information from different sources to make decisions on how to move in order to reach a given goal. Motion planning and control are closely related to many fields of engineering, from robotics to autonomous driving. A combination of motion planning and control enables robots to maneuver safely and reliably in differ-
2 AI technologies, tools, and industrial use cases
43
ent environments while minimizing the energy needed for movement. As the technology continues to develop, motion planning and control will enable robots to navigate through more complex and unpredictable environments with greater accuracy and speed.
2.7.4 Human-robot interaction Human-computer interaction is a rapidly growing field that explores how humans interact with and use computers, robots, and other types of technology. It seeks to understand how humans process, comprehend, and use the systems to interact and collaborate with computers, robots, and other technological devices. With robots being increasingly used in many aspects of daily life, it is becoming increasingly important to understand the interaction between humans and robots. Robots can bring a lot of benefits to humans, from taking on labor-intensive tasks to providing physical assistance in industrial and medical settings. This is why there is a growing interest in studying the human-robot interaction, as the prospect of robots and humans working in tandem is an exciting discussion in AI research. As robots become increasingly sophisticated, there are added complexities to consider when it comes to studying the human-robot interaction. For instance, the way in which a robot responds to a given command will depend on the programming, data, and even the way in which the human responds. It is thought that in the long-term, a robot can eventually possess enough complexity to think for itself and make decisions independently [33]. The implications of this for human-robot interaction are both interesting and challenging. There are already some ethical issues surrounding the idea that a robot could take decisions on its own, which could have an impact on humans. The importance of understanding how humans interact with robots extends further than just providing robots with human-level intelligence. By understanding how humans interact with robots, we can develop better ways in which robots and people can be more productive and connected. With ongoing research, innovation, and progress, human-robot interaction could be a completely new field of collaboration and understanding in the future. The human-robot interaction is an important field of study that seeks to understand how humans interact with robots, and intends to provide a bridge between humans and robots that will enable the two to be more productive and connected. As robots become increasingly sophisticated, there are a variety of factors to consider, which makes it a fascinating and intriguing area of research.
2.7.5 Industrial automation and use cases Industrial automation is the use of technology, systems, and software to control and monitor industrial processes, machines, and equipment. It has enabled greater pro-
44
Shaik Mohammad Rafee et al.
ductivity and efficiency in industrial operations, with increased safety, reliability, and accuracy. Industrial automation has revolutionized many industries, from automotive manufacturing to food production. Industrial automation is most commonly seen in manufacturing processes, where tasks are repeated and automated using machines and robots [34]. Automation makes it possible to increase the speed and accuracy of the production process. It also improves the quality of the product. Industrial automation also helps in eliminating human error and reducing costs. Industrial automation can be used in a variety of industries and fields. Some of the most common use cases are found in the manufacturing industry, where tasks like quality control and inspection are often automated. Automation also makes it possible to create complex products and reduce the number of workers on the assembly line. In the healthcare sector, industrial automation is used to streamline processes and tasks, such as analyzing and interpreting patient data. In the agriculture industry, robotic operations are used to automate the planting, nurturing, and harvesting of crops. In the transportation industry, automated trucks are used to deliver goods as well as improve safety and reduce emissions. Industrial automation has ushered in a new era of efficiency and productivity in many industries. It helps to reduce costs, increase safety and accuracy, and improve product quality. With the continued development of technology and software, there will likely be new use cases of industrial automation in the future.
2.7.6 Autonomous vehicles and drones In recent years, advances in technology have created a dynamic landscape in which autonomous vehicles and drones are now commonplace. Autonomous cars and drones, although relatively new in the transportation industry, have already become integral parts of the future of transportation. Autonomous vehicles operate without direct human input, using computers and AI systems to plan their own routes and navigate obstacles. Similarly, drones are small, unmanned aerial vehicles that can be programmed to fly around objects autonomously. Autonomous vehicles and drones have several advantages over traditional vehicles, including increased safety and efficiency, reduced environmental impact, and potential cost savings. Autonomous cars respond more quickly to changes in the environment and can effectively navigate obstacles, significantly reducing the amount of time it takes to reach a destination. Similarly, drones are much more agile than traditional aircraft and have minimal noise pollution due to their small size [35]. Furthermore, autonomous vehicles and drones are more efficient than vehicles with drivers, and can result in lower energy costs as well as fewer traffic problems. The introduction of autonomous vehicles and drones also has potential to revolutionize industries that require precise tracking capabilities, such as deliveries, inspection, and surveying. Autonomous vehicles allow for faster and more cost-efficient deliveries, as they automatically navigate the shortest possible route to the destination and can be operated 24/7. Similarly, drones can be used for inspection and surveying
2 AI technologies, tools, and industrial use cases
45
tasks, such as surveying large geographic areas or inspecting construction sites from a bird’s eye view. Although these technologies have practical applications, there are still a variety of safety and ethical concerns associated with them. Autonomous vehicles and drones need to be thoroughly tested to ensure they are safe for use on public roads and operated according to legal restrictions. Additionally, there is a fear that autonomous vehicles and drones could increase the risk of surveillance, making it easier for companies and governments to track the movements of citizens. Finally, there are concerns related to the potential for autonomous vehicles and drones to be hacked and used for malicious purposes. Autonomous vehicles and drones are reshaping the landscape of transportation and have the potential to revolutionize industries [36]. In spite of potential drawbacks, the advantages of autonomous vehicles and drones suggest that they will continue to play a crucial role in the future of transportation.
2.8 Industrial use cases AI is already having a major impact on how industrial systems are being designed and operated today. It is being used to automate mundane processes, optimize complex systems, and make them more efficient. AI is being used to create new opportunities for businesses in the industrial sectors. One of the most popular AI-powered industrial use cases is predictive maintenance. Predictive maintenance uses AI-driven analytics to predict when machinery or equipment will need to be serviced or replaced before a breakdown can occur. This helps keep machines running more reliably and reduces the likelihood that a breakdown will occur in the first place. Another way AI is being used in industry is for robotic process automation (RPA). This technology is used to automate tedious tasks such as data entry or customer support. RPA can also be used to recognize images or objects, which can help improve safety and accuracy in industrial settings. AI also has a lot of potential for the manufacturing industry. AI-based systems can be used to manage inventory, deliver real-time feedback on production efficiencies, and detect malfunctions in machines. AI-based facial recognition can also be used to identify workers, helping to improve security in factories. AI is also being used in industries such as finance and healthcare. AI can be used to analyze large amounts of data, identify patterns, and uncover previously unknown insights. This can help financial institutions make faster, more informed decisions, and can also help healthcare providers diagnose illnesses faster and more accurately. AI can also be used to improve customer service in industries such as retail and hospitality. AI-powered chatbots and voice assistants can help customers find what they’re looking for quickly and easily [37]. AI can also be used to track customer preferences and deliver targeted, personalized offers and recommendations. AI is already having a major impact on the industrial sector. It is being used to automate processes, optimize systems, and uncover new opportunities for busi-
46
Shaik Mohammad Rafee et al.
nesses. As AI technology continues to advance, the potential applications of AI in industry will only become more numerous and more powerful in the years to come.
2.8.1 AI in healthcare and medical diagnosis AI has become an increasingly prominent part of healthcare and medical diagnosis. Technological advances have enabled AI to take on a larger role in the medical field by automating mundane tasks and aiding in the diagnosis of diseases. The increased usage of AI has had, and will continue to have, a major impact on the way healthcare is provided and accessed. One of the most popular uses of AI technology in healthcare is in the field of medical diagnosis. AI algorithms have been widely adopted due to their ability to analyze large amounts of data and medical records to detect patterns and generate insights. AI algorithms can effectively identify medically relevant data to enable clinicians to accurately diagnose diseases. For instance, AI algorithms can be used to quickly identify areas of concern within patient databases and provide better treatment solutions. Additionally, AI can also be used to detect possible problems or diseases before they become serious. AI can also be harnessed to automate mundane tasks such as verifying and filling prescription orders as well as keeping up with medical patients’ records. This can not only save time and money, but also reduce potential medical errors. In addition, AI can be used to predict potential medical outcomes and help doctors make informed decisions based on patient data [38]. For example, AI can provide healthcare professionals with data-driven insights that may be used to develop personalized treatment plans and avoid unnecessary treatment procedures. AI has drastically changed and improved the way healthcare is provided and accessed. It has enabled healthcare providers to diagnose diseases more accurately and quickly and automate mundane tasks. Furthermore, AI can be used to analyze large amounts of data to give doctors and healthcare professionals better insights into potential medical outcomes. The increased use of AI will continue to revolutionize the healthcare industry and make healthcare accessible to more people.
2.8.2 AI in finance and banking The role of AI in finance and banking has seen a rapid growth in recent years. AI is a technology that enables machines to think, learn, and act like humans. It is being implemented in financial services, across various areas including banking, investments, insurance, and credit services. AI can help financial services companies to identify patterns in the customer’s data and provide better assistance to customers. For instance, AI can detect changes in customer spending and alert the customers for possible frauds. AI can also help in analyzing customer behavior, predicting customer’s future needs and suggesting appropriate products or services. AI can be used to auto-
2 AI technologies, tools, and industrial use cases
47
mate mundane tasks in financial services, like payments, document handling, and compliance with regulatory requirements. It can be used to detect anomalies in customer’s transactional activities and stop suspicious activities quickly. AI can also be used to detect frauds, mitigate financial risks, and prevent money laundering activities. AI can also be used in financial services to create customer profiles for better segmentation and target marketing. AI-based customer acquisition strategies can also be used to identify potential customers and reach out to them. AI can be used to offer personalized advice to individual customers and suggest suitable products that may suit their needs. AI can also be used to provide automated and optimized customer service. AI has become essential in the finance and banking industry. It provides companies the power to identify patterns in data, automate mundane tasks, detect frauds, prevent money laundering activities, create customer profiles, and offer personalized advice. AI is playing a significant role in transforming the financial services industry and aiding its growth.
2.8.3 AI in manufacturing and supply chain management AI has revolutionized many aspects of how businesses run their operations. It has been especially noteworthy in the fields of manufacturing and supply chain management, and the benefits of using AI to manage and optimize manufacturing processes and supply chain operations are felt across every stage of the production cycle. First and foremost, AI helps to minimize costs by improving efficiency and quickly detecting errors in production processes. Automating processes that were done manually in the past, and implementing automated quality control systems, allow managers to quickly identify inefficiencies in production and fix them with minimal disruption. For supply chain management, AI quickly identifies any potential problems in the supply chain process, such as supply shortages or production delays, and provides real-time solutions that can be implemented quickly and easily. AI also helps to increase product quality and customer satisfaction. By accurately predicting customer demand, AI can help businesses identify which products are in high demand, and ensure the necessary supplies and manpower are available at the right time to produce these items. AI can also help to keep track of the accuracy and quality of products, and alert managers when something is not up to standard. AI helps to reduce the environmental footprint of manufacturing. Data-driven production and AIassisted supply chain management allow for optimized production processes that minimize wastage. This is done by optimizing for things like material utilization and reducing waste in production, as well as optimizing shipping and delivery to minimize fuel costs. AI has had a transformative effect on the manufacturing and supply chain management processes. By reducing costs, increasing product accuracy, and reducing environmental waste, AI helps businesses to be both cost-effective and sustainable. With AI, businesses can streamline their operations, improve quality, and increase customer satisfaction.
48
Shaik Mohammad Rafee et al.
2.8.4 AI in customer service and chat bots The AI technology revolution is rapidly transforming the customer service field as more and more companies are looking to incorporate AI into their service operations. AIpowered chatbots are especially becoming popular as they automate various customer service processes, starting from simple tasks to complex ones. Chat bots are able to understand customer inquiries, respond effectively, and even simulate a human-like conversation with customers, when necessary. They are less expensive than human labor and can provide faster response times. Furthermore, they have the potential to ensure customers are getting the best possible customer service experience, since they are able to process multiple requests simultaneously without having to remember individual information. AI-powered chatbots can also be used to quickly respond to common queries and help customers get answers to their questions quickly. For example, if a customer wants to avail a service or ask for assistance, a chatbot can provide answers to the customer quickly and accurately. They can also suggest other services that the customer may want to explore, based on their inquiry. In addition to responding to customer inquiries and resolving issues quickly, AI-friendly chatbots can also provide proactive customer service. They can remind customers about existing orders or inform them when new products become available. Chatbots can also be used to anticipate needs and offer additional services, such as discounts or vouchers. This can further improve the customer service experience and boost customer loyalty. AI technology is having a huge impact on the customer service field. AI-enabled chatbots are proving to be beneficial in reducing costs, providing faster service resolution, and providing proactive services for customers. Companies should continue to explore how this technology can improve their customer service operations.
2.8.5 AI in energy and utilities The use of AI in energy and utilities is transforming how companies in the sector manage resources. AI is being used to optimize energy consumption and production, identify problems quickly, and improve customer service. AI can optimize energy consumption and production by using advanced analytics to discover new ways to save energy and improve efficiency. It can analyze data from different sources, such as customer bills and weather forecasts, and suggest cost-effective solutions. AI can identify variations in energy usage that could potentially indicate waste or an energy leak and alert the utility company to take action. AI can also predict how usage patterns may change in the near future, and adjust operations accordingly.
2 AI technologies, tools, and industrial use cases
49
2.8.6 AI in agriculture and farming People everywhere are facing the challenge of feeding the world’s growing population while using fewer resources. To help meet this challenge, many people have started to look to AI for assistance in agriculture. AI is an ever-evolving field that has revolutionized various industries such as healthcare, finance, and education. Now, AI is beginning to take on a major role in agriculture and farming. AI has the potential to monitor environments and crops better than manual labor. AI applications now allow farmers to accurately recognize weeds and pests, count individual crops, and identify soil health issues. Farmers are using AI to create a “digital twin” of their farm – an exact 3D representation of the farm and its inputs, which can be used to make better informed decisions and optimize operations. With AI, fertilizers, irrigation, and pesticides can be applied more precisely and in smaller doses [39]. AI can help farmers maximize the use of their resources by enabling them to identify the optimal amounts and combinations of inputs for any given crop. AI can also streamline the farming process by automating repetitive activities such as seeding and harvesting. AI is also being used for predictive analysis. AI can analyze a variety of variables such as weather patterns, crop production, soil data, and market prices. With this data, farmers can make more informed decisions on where to invest their resources for maximum efficiency. However, there are many obstacles that must be overcome for AI to be adopted widely in agriculture. AI technologies are still expensive and require experts to ensure the systems provide accurate results. Accessibility to the internet is a major issue for many rural farmers, a factor which limits the effectiveness of AI technologies. In spite of the challenges, AI has the potential to revolutionize agriculture. It has the potential to make farming more efficient, accurate, and resourceful. As the technology matures, we will see even greater advancements in AI’s role in farming and agriculture.
References [1] [2] [3]
[4]
Patel, P., Ali, M. I., & Sheth, A. (2018). From raw data to smart manufacturing: AI and semantic web of things for industry 4.0. IEEE Intelligent Systems, 33(4), 79–86. Augustine, P., Raj, P., & Munirathinam, S. (Eds.). (2022). Enterprise digital transformation: Technology, tools, and use cases. New York: CRC Press. Alexopoulos, K., Nikolakis, N., & Chryssolouris, G. (2020). Digital twin-driven supervised machine learning for the development of artificial intelligence applications in manufacturing. International Journal of Computer Integrated Manufacturing, 33(5), 429–439. Heier, J. (2021, July). Design intelligence-taking further steps towards new methods and tools for designing in the age of AI. In Artificial Intelligence in HCI: Second International Conference, AI-HCI 2021, Held as Part of the 23rd HCI International Conference, HCII 2021, Virtual Event, July 24–29, 2021, Proceedings (pp. 202–215). Cham: Springer International Publishing.
50
[5] [6]
[7]
[8] [9] [10] [11]
[12]
[13] [14]
[15]
[16]
[17] [18]
[19]
[20] [21]
Shaik Mohammad Rafee et al.
Whittaker, M., Crawford, K., Dobbe, R., Fried, G., Kaziunas, E., Mathur, V. . . . Schwartz, O. (2018). AI now report 2018. (pp. 1–62). New York: AI Now Institute at New York University. Kakani, V., Nguyen, V. H., Kumar, B. P., Kim, H., & Pasupuleti, V. R. (2020). A critical review on computer vision and artificial intelligence in food industry. Journal of Agriculture and Food Research, 2 100033. Bumblauskas, D., Mann, A., Dugan, B., & Rittmer, J. (2020). A blockchain use case in food distribution: Do you know where your food has been?. International Journal of Information Management, 52 102008. Giordani, M., Polese, M., Mezzavilla, M., Rangan, S., & Zorzi, M. (2020). Toward 6G networks: Use cases and technologies. IEEE Communications Magazine, 58(3), 55–61. Chui, M., Roberts, R., & Yee, L. (2022). Generative AI is here: How tools like ChatGPT could change your business. Quantum Black AI by McKinsey, 1–5. Kuzior, A., Sira, M., & Brozek, P. (2022). Using blockchain and artificial intelligence in energy management as a tool to achieve energy efficiency. Virtual Economics, 5(3), 69–90. Chou, J. S., Tai, Y., & Chang, L. J. (2010). Predicting the development cost of TFT-LCD manufacturing equipment with artificial intelligence models. International Journal of Production Economics, 128(1), 339–350. Schlögl, S., Postulka, C., Bernsteiner, R., & Ploder, C. (2019). Artificial intelligence tool penetration in business: Adoption, challenges and fears. In Knowledge Management in Organizations: 14th International Conference, KMO 2019, Zamora, Spain, July 15–18, 2019, Proceedings 14 (pp. 259–270). Spain: Springer International Publishing. Shaw, J., Rudzicz, F., Jamieson, T., & Goldfarb, A. (2019). Artificial intelligence and the implementation challenge. Journal of Medical Internet Research, 21(7), e13659. Praghash, K., Arshath Raja, R., Chidambaram, S., & Shreecharan, D. (2023, May). Hyperspectral image classification using Denoised Stacked Auto Encoder-Based Restricted Boltzmann Machine Classifier. In Ajith Abraham,Tzung-Pei Hong, Ketan Kotecha, Kun Ma, Pooja Manghirmalani Mishra, & Niketa Gandhi (Eds.). Hybrid Intelligent Systems: 22nd International Conference on Hybrid Intelligent Systems (HIS 2022), December 13–15, 2022 (pp. 213–221). Cham: Springer Nature Switzerland. Bello, O., Holzmann, J., Yaqoob, T., & Teodoriu, C. (2015). Application of artificial intelligence methods in drilling system design and operations: A review of the state of the art. Journal of Artificial Intelligence and Soft Computing Research, 5(2), 121–139. Lee, D., & Yoon, S. N. (2021). Application of artificial intelligence-based technologies in the healthcare industry: Opportunities and challenges. International Journal of Environmental Research and Public Health, 18(1), 271. Davenport, T. H. (2018). From analytics to artificial intelligence. Journal of Business Analytics, 1(2), 73–80. Sivapriya, T. R., & Saravanan, V. (2011). Automatic brain MRI mining using Support Vector Machine and Decision tree. CiiT International Journal of Artificial Intelligent Systems and Machine Learning, 3(2), 109–116. Kohli, M., Alkasab, T., Wang, K., Heilbrun, M. E., Flanders, A. E., Dreyer, K., & Kahn, C. E., Jr (2019). Bending the artificial intelligence curve for radiology: Informatics tools from ACR and RSNA. Journal of the American College of Radiology, 16(10), 1464–1470. Peres, R. S., Jia, X., Lee, J., Sun, K., Colombo, A. W., & Barata, J. (2020). Industrial artificial intelligence in industry 4.0-systematic review, challenges and outlook. IEEE Access, 8 220121–220139. Bhende, M., Thakare, A., Pant, B., Singhal, P., Shinde, S., & Saravanan, V. (2022). Deep learningbased real-time discriminate correlation analysis for breast cancer detection. BioMed Research International, 2022, 1–12.
2 AI technologies, tools, and industrial use cases
51
[22] Ren, L., Zhang, L., Wang, L., Tao, F., & Chai, X. (2017). Cloud manufacturing: Key characteristics and applications. International Journal of Computer Integrated Manufacturing, 30(6), 501–515. [23] Romanovs, A., Pichkalov, I., Sabanovic, E., & Skirelis, J. (2019, April). Industry 4.0: Methodologies, tools and applications. In 2019 Open Conference of Electrical, Electronic and Information Sciences (eStream) (pp. 1–4). IEEE. [24] Praghash, K., Peter, G., Stonier, A. A., & Priya, R. D. (2023, May). Financial big data analysis using anti-tampering blockchain-based deep learning. In Hybrid Intelligent Systems: 22nd International Conference on Hybrid Intelligent Systems (HIS 2022), December 13–15, 2022 (pp. 1031–1040). Cham: Springer Nature Switzerland. [25] Hansen, E. B., & Bøgh, S. (2021). Artificial intelligence and internet of things in small and mediumsized enterprises: A survey. Journal of Manufacturing Systems, 58 362–372. [26] Rabah, K. (2018). Convergence of AI, IoT, big data and blockchain: A review. The Lake Institute Journal, 1(1), 1–18. [27] Sircar, A., Yadav, K., Rayavarapu, K., Bist, N., & Oza, H. (2021). Application of machine learning and artificial intelligence in oil and gas industry. Petroleum Research, 6(4), 379–391. [28] Silvia Priscila, S., Sathish Kumar, C., Manikandan, R., Yuvaraj, N., & Ramkumar, M. (2022, February). Interactive artificial neural network model for UX design. In International Conference on Computing, Communication, Electrical and Biomedical Systems (pp. 277–284). Cham: Springer International Publishing. [29] Sayed-Mouchaweh, M. (Ed.). (2020). Artificial intelligence techniques for a scalable energy transition: Advanced methods, digital technologies, decision support tools, and applications. India: Springer Nature. [30] Avinash Khatri, K. C., Shah, K. B., Logeshwaran, J., & Shrestha, A. (2023). Genetic algorithm based techno-economic optimization of an isolated hybrid energy system. ICTACT Journal on Microelectronics, 8(4), 1447–1450. [31] Manikandan, R., Sara, S. B. V., Yuvaraj, N., Chaturvedi, A., Priscila, S. S., & Ramkumar, M. (2022, May). Sequential pattern mining on chemical bonding database in the bioinformatics field. In AIP Conference Proceedings (Vol. 2393, No. 1, p. 020050). AIP Publishing LLC. [32] Ramesh, G., Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). Prediction of energy production level in large PV Plants through AUTO-Encoder Based Neural-Network (AUTO-NN) with restricted Boltzmann feature extraction. Future Internet, 15(2), 46. [33] Kumar, A., Yadav, S. P., & Mittal, S. (2020). DAI for document retrieval. In : Satya Prakash Yadav, Dharmendra Prasad Mahato & Nguyen Thi Dieu Linh (Eds.).Distributed Artificial Intelligence (pp. 251–262). Boca Raton: CRC Press. [34] Tang, A., Tam, R., Cadrin-Chênevert, A., Guest, W., Chong, J., & Barfett, J., Canadian Association of Radiologists (CAR) Artificial Intelligence Working Group (2018). Canadian Association of Radiologists white paper on artificial intelligence in radiology. Canadian Association of Radiologists Journal, 69(2), 120–135. [35] Al-Turjman, F., Yadav, S. P., Kumar, M., Yadav, V., & Stephan, T. (Eds.). (2022). Transforming management with AI, Big-Data, and IoT. India: Springer. [36] Parunak, H. V. D. (1996). Applications of distributed artificial intelligence in industry. Foundations of Distributed Artificial Intelligence, 2(1), 18. [37] Alippi, C., Ferrero, A., & Piuri, V. (1998). Artificial intelligence for instruments and measurement applications. IEEE Instrumentation and Measurement Magazine, 1(2), 9–17. [38] Marr, B. (2019). Artificial intelligence in practice: How 50 successful companies used AI and machine learning to solve problems. USA: John Wiley & Sons. [39] Bravo, C., Saputelli, L., Rivas, F., Pérez, A. G., Nikolaou, M., Zangl, G. . . . Nunez, G. (2014). State of the art of artificial intelligence and predictive analytics in the E&P industry: A technology survey. SPE Journal, 19(04), 547–563.
G. Venkataramana Sagar, S. Ambareesh, P. John Augustine, R. Gayathri
3 Classification and regression algorithms Abstract: Machine learning algorithms for predicting or categorizing data include classification and regression techniques. Regression algorithms are used to forecast a continuous numerical value, such as the anticipated value of a stock price. In contrast, classification algorithms predict a discrete label, such as whether something is a cat or a dog. The accuracy of the algorithm’s predictions and classifications determines how effective a prediction or classification is. Decision trees, support vector machines, k-nearest neighbors, and naive Bayes are examples of standard classification techniques. Regression methods often used include generalized, polynomial, and linear regression. Keywords: Classification, regression algorithm, machine learning, prediction, data
3.1 Overview of classification and regression algorithms Machine learning (ML) algorithms are used to evaluate data and include both classification and regression methods. Based on present patterns, classification algorithms are used to create predictions about a data set. A classification method, for instance, could be used to determine whether or not an email is spam. On the other hand, data analysis uses regression methods to determine the nature of the relationship between the independent and dependent variables [1]. These algorithms are mainly used to forecast or estimate a value from a set of available data. Regression algorithms can be used, for instance, to indicate a person’s annual pay based on their age and gender. Within ML, some tasks require both classification and regression algorithms. Data is typically categorized using classification algorithms into predetermined groups or labels. This procedure involves classifying the data after learning the tags from a particular data set. On the other hand, regression algorithms are typically employed to forecast a continuous output from a given data set. Finding a function that converts the input data into the desired result is the first step in this procedure. ML algorithms are used to examine data, and two common types are classification and regression. Regression algorithms are used to forecast values from input data sets, whereas classification techniques are used to categorize data samples. Both of these algorithms play crucial roles in data analysis and predictive modeling.
https://doi.org/10.1515/9783111323749-003
54
G. Venkataramana Sagar et al.
3.2 Importance of classification and regression in machine learning Regression and classification are two of the most essential methods used in ML. An ML method called classification groups incoming data into predetermined classes. This method frequently identifies trends and divides data into subcategories, including sensitive information, spam, and fraud. Classifiers are employed in various structured and unstructured text analytics types, including face recognition, audio recognition, handwriting recognition, and others. A statistical method called regression fits a model to a set of data points. It determines how one or more predictor and dependent variables are related [2]. It is applied to data to identify trends and influences. By examining linear and nonlinear relationships in data, regression can be used to forecast future trends or values. It is impossible to exaggerate the significance of classification and regression in ML. The model is created and predictions are made using classification and regression methods. More complex algorithms are designed to accurately categorize and analyze data patterns due to the growing complexity of data. These methods are also employed in various fields, including forecasting, facial recognition, consumer segmentation, robotics, fraud detection, natural language processing (NLP), etc. Offering information on consumer behavior, market trends, and other topics, they assist firms in making more informed decisions. Regression and classification are essential to ML techniques. They are applied to data classification, trend analysis, and value forecasting. These methods assist businesses in comprehending consumer behavior and market trends. These two methods ultimately allow ML to find patterns and relationships in massive datasets.
3.3 Supervised learning basics In ML and artificial intelligence (AI), known as supervised learning, models are trained using algorithms on training datasets. Creating a model that can learn to predict outcomes from new data is the aim of supervised learning. The model is “trained” using historical data that has been correctly categorized. The model gains the ability to predict the outcome more precisely as it receives additional training and data. Supervised learning algorithms are rigorous mathematical procedures that analyze datasets labeled with results to find patterns and develop understanding. These formulas are based on models that relate inputs to anticipated or intended results [3]. Decision tree learning, neural networks, support vector machines (SVMs), and Bayesian networks are the most well-liked supervised learning methods.
3 Classification and regression algorithms
55
3.3.1 Overview of supervised learning To precisely predict the intended learning output, the supervised learning approach needs a “training dataset.” This training dataset consists of labeled samples, including the desired result and the input data. The labeled data is used to train the model on how to predict the intended outcome from new data. Following the supervised learning procedure’s completion, the model should be able to predict and categorize new data points correctly. This makes it possible to test the model using inputs that it has not previously been trained on. For computer scientists and data scientists, supervised learning is a potent tool that enables automatic predictions from new data or data the model has never seen before. The model can develop the ability to anticipate the desired outcomes more precisely as it continues to gather more data and is trained with additional labels.
3.3.2 Training and testing datasets ML projects need training and testing datasets because they give the model the knowledge and experience to make wise decisions and predictions. The testing dataset is used to assess the model’s correctness and ascertain how well it will make a prediction or decision on previously unknown data. The training dataset is used to create and train the model. Training datasets are gathered via prior experience and are used to instruct the model on how to correctly find patterns, correlations, and similarities among various parameters [4]. One or more associated factors, such as a client’s name, age, marital status, and spending patterns, may be included in a training dataset that the model can use to determine which good or service will most likely meet the consumer’s demands. Testing datasets differ from others because they contain information the model has never seen before. This information evaluates the model’s ability to conclude, make predictions, or compare data that it has never seen before. Testing datasets are also used to ensure the model is suitably extended to fit many circumstances rather than overfitting the training data. Training and testing datasets are necessary to build and assess an ML model. Using both datasets, the model can learn how to make accurate predictions and decisions based on experience and data it has never seen. ML projects need training and testing datasets because they give the model the knowledge and expertise to make wise decisions and predictions. The testing dataset is used to assess the model’s correctness and ascertain how well it will make a prediction or decision on previously unknown data. The training dataset is used to create and train the model. Training datasets are gathered via prior experience and are used to instruct the model on how to correctly find patterns, correlations, and similarities among various parameters [4]. One or more associated factors, such as a client’s name, age, marital status, and spending patterns, may be included in a training dataset that the model can use to determine which good or ser-
56
G. Venkataramana Sagar et al.
vice will most likely meet the consumer’s demands. Testing datasets differ from others because they contain information the model has never seen before. This information evaluates the model’s ability to conclude, make predictions, or compare data that it has never seen before. Testing datasets are also used to ensure the model is suitably extended to fit many circumstances rather than overfitting the training data. Training and testing datasets are necessary to build and assess an ML model. Using both datasets, the model can learn how to make accurate predictions and decisions based on experience and data it has never seen.
3.3.3 Evaluation metrics for classification and regression Metrics for evaluation are used to gauge a model’s efficacy. This is crucial when contrasting various models to determine which of them fits a particular issue the best. Two of the most popular ML approaches are classification and regression, and there are multiple metrics to gauge their effectiveness. Regression measures include mean absolute error, mean squared error, and R2 score, whereas classification metrics include accuracy, precision, recall, and F1 score. The most used classification metric is accuracy, which expresses how effectively a model predicts as a percentage of all instances [5]. Higher scores indicate more accurate predictions. Precision is calculated as the ratio of genuine positives to all anticipated positives. Higher scores in recall indicate fewer false negatives since it gauges the proportion of real positives to actual positives. The F1 score combines precision and recall, and the highest scores show the best coordination of these two metrics. The average discrepancy between the anticipated and actual values is the mean absolute error in regression. The mean squared error measures the average of the squares of the differences between the predicted and actual values. The percentage of data variation that the model successfully explains is the R2 score. This gauges how well the model can account for any observed data changes. Accuracy, precision, recall, and F1 score are assessment metrics for classification; mean absolute error, mean squared error, and R2 score are evaluation measures for regression. These measures can be used to compare the performance of many ML models and guarantee that the best model is selected for a given task.
3.4 Logistic regression When categorizing numerical data into various groups, the powerful and well-liked predictive modeling technique, logistic regression, is applied. It is a regression technique used to forecast the likelihood of an event or outcome. The most specific industries for its use are market research, healthcare, finance, and engineering. Using an iterative procedure, the supervised learning algorithm logistic regression estimates
3 Classification and regression algorithms
57
and categorizes the likelihood of a target outcome. The observed data are mapped onto a range of values between 0 and 1 using a logistic function. As it can forecast the likelihood that an event or consequence will occur, logistic regression is a valuable tool in many different domains [6]. A simple strategy that yields quick and precise results is logistic regression. It may be applied to various tasks, including predicting stock prices, predicting customer turnover, and classifying data into multiple groups. It is also very adaptable because it may be applied to both linear and nonlinear models. Precision, recall, accuracy, and F1 scores are frequently used to evaluate logistic regression models. These metrics gauge how well the model predicts the desired result or class. They also aid in locating potential biases and error-producing factors in the model. A versatile predictive modeling method that may be employed in many different disciplines is logistic regression. This flexible approach can handle both linear and nonlinear models. It is simple to use and yields quick, precise results. Additionally, its performance is frequently assessed using F1 scores, precision, recall, and accuracy measures.
3.4.1 Binary logistic regression A type of statistical analysis called binary logistic regression is used to forecast the likelihood that a dichotomous (two-category) event will have a particular outcome in the future. This type of regression analysis uses continuous independent variables and a categorical dependent variable. In light of the knowledge offered by the observed data, it is used to calculate the likelihood that a specific occurrence will occur. Binary logistic regression models are used for outcomes with just two possible values, such as true or false, male or female, pass or fail, yes or no, on or off, etc. Binary logistic regression is a powerful technique to model binary response data and comprehend correlations between independent factors and dependent variables. It can be used to calculate the likelihood that a specific outcome will occur, making it valuable for forecasting future behavior. It can also pinpoint crucial elements that affect how well a forecast turns out. To comprehend how different circumstances and events affect outcomes or results, it is frequently used in medical, psychological, marketing, and other disciplines [7]. The relationship between the dependent variable and one or more independent factors that affect the result is explained using the binary logistic regression model. This is accomplished by evaluating the impact of each independent variable on the likelihood of the outcome. In binary logistic regression, the independent variables are assumed to be unrelated and have a linear connection with the event’s probability. Use logistic regression to comprehend how several independent variables affect a decision’s outcome. It can be used, for instance, to investigate how various client characteristics may affect their propensity to sign up for a service. It can also assess how well a particular marketing effort attracted new clients. Binary logistic regression can be used in the medical industry to evaluate how risk factors
58
G. Venkataramana Sagar et al.
affect the likelihood of developing disease and dying. Binary logistic regression is a valuable technique for modeling, forecasting, and comprehending the possibility that specific events will occur in the future. It is a useful tool for many different fields and has a wide range of applications.
3.4.2 Multinomial logistic regression Multinomial logistic regression is a form of ML technique that works well for categorical outcome prediction. A set of explanatory variables, commonly called predictors, are used to train a logistic regression model to forecast the likelihood of a discrete result. In contrast to binary logistic regression, which is used to classify data with only two possible outcomes, multinomial logistic regression can be used to classify data with more than two possible outcomes. Multinomial logistic regression can analyze data effectively and make precise predictions, such as forecasting a customer’s response to an offer or placing them in a specific market group. A dataset can be examined for trends and associations using multinomial logistic regression, which can then be used to make decisions or develop marketing plans. Compared to other ML techniques, multinomial logistic regression models are thought to be more robust and better able to withstand overfitting [8]. This is because it uses the links between the explanatory variables to forecast the outcome rather than assuming that one variable is the only significant predictor of a given event. As a result, multinomial logistic regression frequently makes more precise forecasts. Multinomial logistic regression is typically implemented using software like SAS or R. With the appropriate datasets and model parameters, this kind of software is intended to support the multinomial logistic regression process. Additionally, the model’s outputs may be interpreted using the software, which helps comprehend the influence that each predictor has on the conclusion. The model can then be utilized to generate accurate predictions after it has been put into practice. Multinomial logistic regression is a robust statistical method that can be used to categorize data with several outcomes precisely. To create accurate predictions, it enables the evaluation of the associations between the explanatory factors and the result. Additionally, this ML method can be used to find patterns and insights in data, helping decision-makers.
3.4.3 Applications and use cases When one or more predictor variables (X) are about a binary class variable (Y), logistic regression is a statistical approach used to examine the data. It is one of the most popular predictive modeling approaches in the predictive analytics community. For classification issues where the goal variable is binary, logistic regression predicts cate-
3 Classification and regression algorithms
59
gorical data. Logistic regression is frequently employed in various disciplines, including marketing, medicine, finance, and political science. These applications include forecasting consumer propensity to buy a product and patient inclination to acquire a specific ailment [9]. Logistic regression is commonly used in applications such as: – Credit scoring: Applications like loan approval and credit score use this technique. Experts can determine the probability of a default occurring and the likelihood that a borrower would repay a loan by using logistic regression. – Healthcare: Logistic regression models are frequently utilized to determine the risk of various medical disorders. For instance, doctors use logistic regression to estimate the likelihood of a patient contracting a specific illness or condition. A treatment plan is then created using this information. – Marketing: Logistic regression is additionally used in marketing to assess consumer behavior and forecast the likelihood that consumers will purchase a product or take advantage of a deal. – Predictive maintenance: Predictive maintenance uses logistic regression to detect impending equipment failure or determine the potential dangers of changing a component in a complicated machine. This knowledge is useful in making care decisions. Logistic regression is also extensively used in fraud detection. It is employed to spot unusual activity on accounts or questionable conduct on the part of people.
3.5 Decision trees A supervised learning system known as a decision tree is used in many applications, such as image classification and stock market forecasting. They are well-liked because they are simple to understand and may be applied to decision-making without requiring sophisticated mathematics. A branching diagram with a tree-like appearance is the general structure of a decision tree. Decisions are made from the root nodes to the leaves or end nodes. We evaluate a collection of features at each internal node and take actions based on the input values of each part. Each tree branch represents a potential result or option. For instance, a decision tree can be used to determine if a tumor is benign or malignant. The cancer’s size, form, and color are evaluated at each internal node. After that, the input variables are contrasted with values known to indicate malignancy, such as bigger size or darker hue. Whether a tumor is benign or malignant is determined at the end nodes. Decision trees are helpful because they are simple to interpret [10]. The decisions can be clearly understood when a decision tree is visually represented. Decision trees are versatile and valuable in various applications because they are nonparametric models that can handle linear and nonlinear interactions between features. Decision trees are a useful tool for making judgments
60
G. Venkataramana Sagar et al.
in supervised learning applications. They can manage many relationships between input variables and offer an easy-to-understand depiction of the decision-making process. Decision trees are thus helpful for data scientists and executives who make business decisions.
3.5.1 Basics of decision trees Decision trees are a helpful tool that customers and business owners can use to understand their data and make informed decisions. A diagrammatic representation of the decision-making process, decision trees show each option and its effects. By quickly understanding the various possibilities for a decision, a firm can receive insight into where and how to move forward, avoiding costly mistakes that might occur if poor judgment is made. Making decision trees requires the use of input data. This information could consist of market research, financial data, or customer surveys, depending on the kind of decision that needs to be made. This creates a decision tree, with each branch denoting a potential outcome of the choice. Following each chapter, the decision-maker can determine which product is preferable. The nodes represent the choices that must be made, and the links between the nodes show the potential courses of action. The tree’s leaves or endpoints characterize the conclusion of the decision-making process.. The branches are scored or evaluated using the decisionmaker’s stated criteria, and the best outcome is chosen. Decision trees can tackle problems such as client segmentation, customer service, product testing, and financial analysis. It is important to note that decision trees should only be used for classification problems, not regression problems [11]. The decision tree is a reliable data analysis method that can streamline and simplify difficult decisions. By visually presenting data and potential courses of action, decision trees help business owners, decisionmakers, and customers quickly discover and choose the appropriate action.
3.5.2 Splitting criteria A supervised learning approach, decision tree learning, can be applied to classification and regression applications. It is an algorithm that can generate a predictive model from a data set by building a tree of rules. Decision trees can generally predict class labels or values in response to particular circumstances. As they make it simple to build complex models without asking the user to define all possible conditions, decision trees are an effective tool in ML. The criteria we employ to choose the feature that best separates the data samples for our decision tree are called the splitting criteria of decision trees. In most cases, we perform an entropy measurement to identify the attribute that will increase information gain the greatest. Your data’s entropy level indicates how
3 Classification and regression algorithms
61
disorganized or ensure it is. Finding the optimal point for categorizing our data into different classes is the goal of decision trees. Information gain is a metric representing the anticipated entropy reduction by dividing the data by an attribute. Typically, the point with the most significant information gain is preferred for the split. It can employ a calculation like the Shannon entropy, which considers the likelihood that each sample will belong to a specific class, to determine the information gain. The degree of uncertainty in the data increases with the Shannon entropy. The knowledge gain is higher, the less uncertain the situation. It also considers other aspects when selecting which attribute to choose for a split. In general, it favors details with a bigger sample size or more different values, enabling us to divide the data more effectively. Although various factors may affect the split, the computed information gain determines the preferable partition [12]. The feature that best distinguishes data samples from one another can be chosen using the splitting criteria of decision trees. We can locate the part with the most significant gain by computing the Shannon entropy, which allows us to quantify the information gain of each property. When choosing which feature splits the data most effectively, further factors may also be considered.
3.5.3 Handling categorical and numerical features Decision trees are one of the most widely used ML techniques for supervised classification tasks. They are regarded as one of the most precise and trustworthy prediction models. A decision tree is the most effective solution for classification and regression issues. A decision tree’s principal function is to segment a dataset into more digestible chunks so that predictions can be made. This is accomplished by managing the dataset’s categorical and numerical properties. In contrast to numerical features – variables with continuous values like height, weight, and salary – categorical features are variables with discrete values like gender, race, and age. The decision tree uses a divide-and-conquer strategy to classify the data when dealing with flat features [13]. According to the values of the categorical features, it will divide the dataset into subgroups. The decision tree then gives each of the subgroups a label or class. A decision tree will establish a split point at a specific value to divide the data into various branches while handling numerical features. The facts on the different components are then once more labeled. The entropy or information gain associated with each characteristic determines this approach’s split points and labels. The decision tree is likelier to incorporate a feature in its prediction when the entropy is higher. The decision tree will find the title or class that appears the most frequently among the branches to make a prediction, and it will then assign that label or type to the data. This is referred to as the majority voting strategy. When it comes to handling both category and numerical variables, decision trees are very effective. In addition, they are a preferred option among other ML algorithms due to their simplicity and precision.
62
G. Venkataramana Sagar et al.
3.5.4 Ensemble methods: random forests and gradient boosting Due to their better accuracy and resilience, ensemble methods – which incorporate numerous algorithms – have emerged as one of the most widely used methods for supervised ML. Gradient boosting and random forests are the most commonly used ensemble techniques. An ensemble technique called random forests creates numerous decision trees to generate predictions. Each decision tree employs a unique mix of characteristics and is trained using a different subset of the training data. As a result, separate decision trees are created, which are subsequently joined using a simple majority voting strategy. With the help of this averaging strategy, the model’s bias and variance are reduced, significantly improving its accuracy and robustness. An ensemble technique called gradient boosting joins several “weak” learners to create a “strong” learner. Different subsets of data are used to train each weak learner’s decision tree; however, unlike random forests, the vulnerable learners are not independent. The learners that follow are taught to concentrate on the data points the previous learner struggled to classify accurately. The performance of the weaker learners determines the weights for the weighted average strategy used to mix these learners. This method produces a model with much higher accuracy and robustness than a single decision tree. Gradient boosting and random forests are effective ensemble techniques that can boost precision and robustness. Although each strategy has pros and cons, they are both successful solutions for supervised learning issues.
3.6 Support vector machines Powerful and often-used SVMs are a subset of ML techniques. They are supervised learning models that can be applied to classification and regression issues. SVMs are founded on margin maximization, which aims to reduce the risk of overfitting by maximizing the distance between the classifier’s boundary and the training data points. As they can provide highly accurate models despite complicated interactions between features, SVMs are incredibly alluring [14]. SVMs can handle big datasets with a variety of features and are resistant to overfitting. Additionally, SVMs can frequently be combined with kernel functions to map nonlinear data into a space that can be divided linearly. SVMs are commonly chosen over conventional artificial neural networks (ANN) because they need less training data and develop accuracy more quickly. Additionally, they are less likely to overfit, which makes them better suited for complex tasks. SVMs are strong and well-liked ML techniques for classification and regression tasks. They are robust to overfitting and aim to maximize the margin between the boundary and the training data points, allowing them to build highly accurate models even with complex input data. Additionally, they perform well with nonlinear datasets thanks to their use of kernel functions.
3 Classification and regression algorithms
63
3.6.1 Linear SVM The linear SVM is a supervised ML model that recognizes and separates data points in feature space using several methods. This algorithm creates a hyperplane or collection of hyperplanes in a high- or infinite-dimensional room that can divide various object types. The algorithm can make a decision boundary that will assign classes to the input objects by identifying a group of hyperplanes that collectively divide the multiple categories of things. To maximize the distance between the two data types, the linear SVM algorithm first builds two parallel hyperplanes of maximum width, known as a decision surface. A unique optimization method known as the quadratic programming method, which optimizes an expression containing the geometric distance between the two classes of data, is used to find the best hyperplane (maximizing the width between the two data classes). The program will employ a margin measurement after creating the ideal hyperplane. The margin is the shortest distance between two data points, one from each of the two classes, in a linear SVM. The algorithm will determine the margin using the two locations closest to the hyperplane [15]. In a wide range of classification tasks, it has been discovered that the linear SVM method performs exceptionally well. It is now among the most used algorithms for handling supervised learning issues. The algorithm is well-optimized and has been proven to scale very well with massive datasets due to its speed and complete accuracy. The linear SVM algorithm is helpful for many applications since it is resilient against outliers and can handle high-dimensional datasets.
3.6.2 Nonlinear SVM with kernel functions A potent method for classification jobs, particularly for big and complicated data sets, is the nonlinear SVM with kernel functions. Algorithms for ML make extensive use of it. A hyperplane that optimizes the margin or separation between the two classes is used to map a data set into a higher dimensional space using a kernel function to separate the points. These models are also called “kernel SVMs” due to their nonlinear nature. The polynomial kernel, radial basis function (RBF) kernels, and sigmoid kernels are SVM’s most widely used kernel functions. Each operation on specific data sets outperforms linear kernels like the linear SVM. The polynomial kernel is frequently used when dealing with data sets that cannot be separated linearly, such as the XOR problem. It converts the provided data points into a higher dimensional space and divides the ends using a polynomial [16]. When the data cannot be split linearly, the RBF kernel is used. It is used for cases where the relationship between variables can be complex and is more flexible than a polynomial kernel. The RBF kernel can be used for various data types and supports different decision boundaries. When the data cannot be linearly separated, and the output can have a value of 0 or
64
G. Venkataramana Sagar et al.
1, the sigmoid kernel is used. It is used for issues like text classification, where giving the probability to particular classes is necessary. The sophisticated tools known as nonlinear SVMs with kernel functions can be used to categorize large, complex data sets. They are very helpful when handling massive data sets. The complexity of the data and the task at hand will determine the type of kernel employed.
3.6.3 Soft-margin SVM The soft-margin SVM is a powerful ML technique among many other applications. The SVM excels at pattern identification and classification problems due to its capacity to learn intricate nonlinear correlations between variables. In essence, the SVM creates a border between two different groups of points using a series of attributes in a dataset. The SVM can recognize and categorize new data points based on their position on this border by optimizing the parameters of this boundary. The SVM is particularly good at creating boundaries that accurately divide two different sets of data points. The soft-margin SVM incorporates a parameter called C that enables the model to categorize specific issues across the border to enhance overall boundary fit. This increases accuracy even further. This procedure adds error tolerance and improves the model’s ability to represent nonlinear interactions between variables. The softmargin SVM is a powerful and practical ML technique used significantly in several fields, including finance, face detection, image classification, and NLP [17]. The model is an excellent choice in various circumstances due to its capacity to recognize and categorize complicated patterns, its handling of noisy data, and its tolerance of errors.
3.6.4 Applications and use cases Since its emergence in the 1990s, SVMs have gained popularity as a tool for various applications. SVMs are supervised algorithms that categorize data by locating the hyperplane that most accurately distinguishes the two classes. Even when the data contains outliers or noise, it performs well with small datasets and has good accuracy. Regression, classification, novelty detection, and clustering are just a few of the tasks in which SVMs can be used. They are highly adaptable and may be used with organized and unstructured data. Image classification, text classification, object identification, facial recognition, protein folding, robotic navigation, and handwriting recognition are some of the applications for SVMs that are most frequently used. SVMs are frequently used in the classification of images. The SVM model can correctly categorize new photos of the same items after being trained on a large batch of labeled photographs of those same objects. Complex traits that conventional algorithmic methods might be unable to identify can be found using the SVM model.
3 Classification and regression algorithms
–
–
–
65
Text categorization is yet another essential SVM application. The classification of text documents according to their semantic significance is known as text categorization. After being trained on labeled documents, the SVM model can correctly categorize new documents it has not seen. This can group news stories, emails, blog posts, web pages, and other content into the appropriate categories. Object detection is yet another important use for SVMs. The SVM model is trained to identify individual objects in an image or video. This can be used to detect vehicles in a scenario involving automated driving or to detect items in a project involving 3D scanning. Using SVMs for facial recognition is another option. The SVM model can accurately recognize unfamiliar faces after being trained on a vast collection of face photos. This technology is frequently employed for security purposes such as face recognition to confirm an individual’s identification.
These are only a few of the numerous uses and applications for SVMs. SVMs are a valuable tool for any data scientist or ML engineer because of their adaptability and capacity to classify data effectively even when it contains noise and outliers.
3.7 Naive Bayes The popular ML algorithm naive Bayes is used in supervised learning problems. It is a probabilistic method that performs classification jobs based on the Bayes’ theorem. Naive Bayes models are popular in many applications and are straightforward yet effective. The technique is considered ‘naive’ since it assumes independence between several traits or variables. This is the reason the algorithm is occasionally called naive Bayes. The final probability may be calculated considerably more quickly and easily, thanks to the independence assumption. Naive Bayes performs better than other classification algorithms, using fewer resources, making it a good choice for small datasets [18]. The algorithm uses the Bayes’ theorem to determine the likelihood of an event based on historical data. Based on this probability, it then makes a classification decision for a new item. For instance, naive Bayes will examine all the words in the email and the likelihood that each dish will be linked with spam or ham (not spam) when classifying a new email as either ham (not spam) or spam. The algorithm will determine the likelihood that each word will be related to either one before using Bayes’ theorem to get the possibility that the entire email will be connected to either one. Naive Bayes is used in many applications because it is quick and straightforward, including search engines, document filtering, text prediction, and many others. It is also a fantastic place to start if you want to learn more about data science and ML.
66
G. Venkataramana Sagar et al.
3.7.1 Bayes’ theorem and probability basics A crucial tool in probability and statistics is the Bayes’ theorem. It is described as a formula for figuring out the conditional probability of an event using information about past conditions or circumstances. In other words, the Bayes’ theorem illustrates how new information can alter the likelihood of a specific event. It bears the name Thomas Bayes (1702–1761), a Presbyterian priest and English mathematician. The theorem is expressed in the following equation: PðAjBÞ = PðBjAÞPðAÞ = PðBÞ
(3:1)
where P(A|B) is the probability of event A occurring given that event B has already occurred. P(B|A) is the probability of event B occurring given that event A has already occurred. P(A) is the probability of event A occurring, independent of any other events. Finally, P(B) is the probability of event B occurring, independent of any other events. For example, let us say we want to use the Bayes’ theorem to calculate the probability of someone having the flu, given that they have a fever. We could express this problem in the following equation: PðflujfeverÞ = PðfeverjfluÞPðfluÞ =PðfeverÞ
(3:2)
Here, P(flu|fever) is the probability of having the flu given that the person has a fever. P(fever|flu) is the probability of having a fever, given that the person already has the flu. Without considering any other influencing circumstances, P(flu) and P(fever) represent the likelihood of getting the flu and a fever, respectively. Given the existence of a fever, we may determine the possibility of having the flu by plugging the actual probabilities of each event into the equation. This allows us to revise our assessment of an event’s likelihood when new information comes to light. As this example shows, the Bayes’ theorem is a helpful resource for performing probabilistic computations. Its fundamental ideas are simple, and they may be applied to make decisions about potential outcomes in various settings.
3.7.2 Naive Bayes assumption An ML method called the naive Bayes assumption forecasts potential outcomes based on a group of connected variables. This presumption, based on the Bayes’ theorem, asserts that all of the model’s linked variables are independently correlated [19]. This means that the value of a single variable will not impact the results of any other variables in the model. This supposition makes computation easier since it eliminates the need to study the intricate relationships between the variables and enables the model to generate predictions much more quickly. This assumption simplifies the calculating process, but it can be wrong if the data are not correctly analyzed; remembering is
3 Classification and regression algorithms
67
crucial. This assumption does not consider potential correlations between the variables, which means it might not accurately reflect their actual relationship. Understanding the connections between the dataset’s variables and using them to create a trustworthy algorithm is crucial for successfully using the Naive Bayes assumption. The model must be carefully examined and tested to ensure accuracy, like with any ML technique. Additionally, it is crucial to verify that the data meet the assumptions established in the model before relying on the results produced by this assumption because the independent nature of this assumption depends on the underlying data.
3.7.3 Gaussian naive Bayes The probabilistic classification algorithm, Gaussian naive Bayes (GNB), is also called a linear classifier. It is a well-known and effective classification technique based on Bayes’ theorem and employs a probabilistic classification strategy. It is frequently used in various applications, including spam filtering, text classification, and NLP. Given a set of input data, GNB calculates the probability of a specific event using statistics. It presumes that each characteristic has a normal distribution and that all features are independent. Due to its resistance to outliers, this classification approach is frequently employed in issues requiring numerous classes and is highly helpful for handling unbalanced or noisy datasets. A basic GNB classification algorithm has two parts: first, it determines the likelihood of each type for each feature, and second, it determines the posterior probability of each class for a given set of input values [20]. It also incorporates prior knowledge to make the model more accurate, such as what the data type should indicate or what the class labels should represent. The simplicity and execution speed of the GNB algorithm is among its benefits. It may be put into use very quickly, and compared to specific other approaches, its prediction time is quicker. Additionally, the technique is more generalizable and less prone to overfitting because it only needs a small number of training samples to determine the weights for the classes and features. In comparison to other probabilistic classification techniques, it requires less processing power. Due to the assumption that all data would have a normal distribution, the GNB algorithm has drawbacks when dealing with non-normal data. The GNB algorithm may also produce erroneous predictions if the prior data is inaccurate. Every classification algorithm, including GNB, is susceptible to overfitting, which can harm the algorithm’s effectiveness. An effective and practical technique for ML and data analysis is the GNB. It is a popular option in many sectors since it can be used to answer a wide range of classification problems accurately.
68
G. Venkataramana Sagar et al.
3.7.4 Multinomial naive Bayes The multinomial naive Bayes is an ML method used in classifying text and other data. It is based on the Bayes’ theorem and assumes that each feature contributes equally to classification. It differs from multivariate naive Bayes in that it relies on an individual probability distribution for each part rather than a joint probability distribution like multivariate naive Bayes does. The initial step in the algorithm’s operation is determining the probability of a particular class given the input features. Assuming that each part has a multinomial distribution, this is accomplished. The class with the highest chance is then chosen as the anticipated class based on the class-specific probability distribution. Numerous text classification tasks, including sentiment analysis, topic modeling, language detection, can be performed using the technique. Additionally, it aids in biological sequence recognition and medical diagnosis [21]. Multinomial naive Bayes is quick, easy, and accurate compared to other ML algorithms. It is effective with a limited dataset and applies when missing data exists. Additionally, it is resistant to the influence of unimportant traits. Also, there are several drawbacks to multinomial Naive Bayes, such as the assumption that each feature is independent of the others. In some circumstances, this presumption may be accurate only sometimes. Additionally, it ignores the potential significance of feature interactions. The multinomial naive Bayes algorithm is a potent tool for categorizing data. It can be used for many things, including language identification, sentiment analysis, and medical diagnosis. With a limited dataset, it can produce good results quickly and accurately. It is also reasonably resistant to the influence of unimportant factors.
3.7.5 Applications and use cases A robust ML technique called naive Bayes can forecast an outcome given a set of input variables. It is founded on the Bayes’ Theorem, a method for estimating the likelihood of certain events. Naive Bayes is a prominent and highly favored option for many data science projects because of its wide range of applications and use cases. – Spam filtering is one application where naive Bayes is frequently used. Based on previously collected data, Naive Bayes can be used to determine whether a given email is spam. A spam detector can use specific terms or phrases from the email and the sender’s information. If you have adequate data, you can use Naive Bayes to develop an efficient classifier for this issue. – The classification of texts is another use for naive Bayes. For NLP, categorical labels are frequently necessary for textbooks to be processed intelligently. For instance, you might want to group sentences or text documents into various categories. Naive Bayes can calculate the odds of each text falling into each class quickly and accurately.
3 Classification and regression algorithms
–
69
Predicting client churn can also be done with naive Bayes. By indicating the dangers, naive Bayes can help organizations identify which clients will most likely stop using their services. This forecast can be formed based on client demographic information, transaction histories, and other behavioral patterns. This aids companies in staying one step ahead and implementing the necessary adjustments.
Sentiment analysis frequently makes use of naive Bayes. The system can determine whether a sentence is a positive or negative ken. This is a helpful tool for businesses to quickly and accurately learn what customers think about their goods or services. Naive Bayes is an ML technique with many applications that is very adaptable and adaptive. It can be used for sentiment analysis, text classification, customer churn prediction, etc. Due to this, it is a standard option for many data science initiatives.
3.8 K-nearest neighbors The K-nearest neighbors (K-NN) algorithm uses neighborhood relationships among data points as a classification and regression technique to find groups and forecast outcomes. This well-known ML technique is used to find the dataset’s data point that most closely matches a given data point. By examining a batch of data points, K-NN finds the k-nearest issues with the correct labels. If the k-nearest points are all marked as positive, then the current moment is also celebrated as positive. These points serve as reference points for the current data point. The K-NN act as a filter and advise on categorizing the present spot. K-NN is a straightforward but effective technique that does not need much training or preprocessing of the data. It is simple to implement and does not need complex algebra or statistics [22]. The algorithm is typically less accurate than other more complex approaches like decision trees and neural networks because of its simplicity. KNN can categorize brand-new data points, spot trends in a dataset, find outliers, and forecast brand-new data points. K-NN is a well-liked option for various ML tasks, including image classification, anomaly detection, time series forecasting, and customized recommendation systems due to its versatility.
3.8.1 K-NN algorithm The K-NN algorithm is a classification system that groups objects according to their characteristics or qualities. It is a nonparametric approach, which implies that it does not rely on any presumptions regarding how the data are distributed. Instead, it examines the nearest data points and applies their labels to those points to classify the
70
G. Venkataramana Sagar et al.
new issue. The foundation of the K-NN method is the idea that related objects should be grouped. The K-NN technique begins by choosing the “K”’ data points nearest to the newly discovered data point that needs to be categorized. Afterward, a “weight” is assigned based on separation of each of those data points from the new issue. The algorithm uses the weighted average of the labels assigned to the unique point’s K-NN to forecast its title [23]. Applications for the K-NN algorithm include pattern recognition, image processing, and data mining. It is well-liked in many applications since it is a simple and quick method. The K-NN algorithm benefits from not making assumptions about the underlying distribution of the data and performs well with noisy data. The K-NN technique is frequently used in a wide range of applications and can be a valuable tool for identifying objects.
3.8.2 Distance metrics A classification and regression approach called K-NN is used to find the nearest data points in a dataset. The algorithm must create a distance metric or metric to measure the separation between two data points to make decisions based on its data points. A K-NN method can employ a wide range of distance metrics, including Euclidean, Manhattan, Minkowski, Chebyshev, Mahalanobis, Hamming, cosine similarity, and BrayCurtis. Each measure has unique qualities and is valuable for specific kinds of data. – The Euclidean distance metric is the most used distance metric for K-NN algorithms. The straight-line distance between two points is measured using a straightforward distance metric. This metric is frequently employed when the distribution is somewhat uniform and the data are continuous. – Rather than calculating the distance along a straight line, the Manhattan distance metric calculates the total of the absolute differences between two places. It is generally employed when the data is discrete and has a nonuniform distribution. – The Minkowski distance metric, which combines the Manhattan and Euclidean distance metrics, is another intricate distance metric. It is employed when the underlying distribution can be roughly classified as either uniform or nonuniform and the data is continuous. – The maximum absolute difference between two points is measured using the Chebyshev distance metric, a straightforward distance metric. It is frequently applied to highly variable data that could contain anomalies or other unreliable data pieces. – A more intricate distance metric that accounts for the correlation between data variables is the Mahalanobis distance metric. It is frequently applied to datasets with numerous dependent variables correlated with one another. – A straightforward distance metric known as the Hamming distance counts the number of bits that differ between two binary integers. It is generally used when there is no ordering for categorical data.
3 Classification and regression algorithms
–
–
71
The cosine similarity distance metric, which is more frequently employed for text processing applications, quantifies the angle between vectors. NLP activities like sentiment analysis and document categorization are where it is most commonly used. The ratio of the difference between two series of values is calculated using the Bray-Curtis distance metric. It is typically used when evaluating ecological distance metrics to compare the composition of various organisms in a dataset.
K-NN algorithms provide various distance measures with unique properties and applications. Accurate outcomes with the K-NN method depend on selecting the appropriate distance metric for the task.
3.8.3 Choosing the optimal K-value The best K-value must be chosen for an ML algorithm to be accurate and have a low prediction error. K determines how many adjacent data points are used to produce a prediction. Since more data points can be considered for each prediction with a higher K-value, predictions are typically more accurate. The algorithm may become too influenced by the constraints of the training data rather than the patterns of the data as a whole, if a higher K-value is used. This is known as overfitting. The elbow approach is a reliable way to find the ideal K-value. Finding the “elbow” point, or the point at which the error rate starts to plateau, entails plotting a graph of the mean error rate versus the K-value. This is an example of the excellent K-value, which offers the optimum compromise between precision and overfitting. The dimensionality problem should also be considered when choosing a K-value [24]. This is the phenomenon where more K-values (neighbors) are needed to represent each point, and more qualities (variables) appropriately are present in the data set. Choosing a K-value that considers the number of variables present in the data set is crucial because doing otherwise can result in overfitting. Finally, domain expertise should also be considered while selecting an outstanding K-value. For instance, a more considerable K-value might be chosen in a medical dataset to produce a more precise forecast. The elbow approach should be used to calculate the ideal K-value, considering the number of variables in the data collection and the domain’s needs. These elements can be considered to select a K-value that strikes a balance between overfitting and precision.
3.8.4 Applications and use cases Powerful ML algorithms like K-NN are frequently employed for supervised and unsupervised tasks. K-NN is a straightforward yet effective algorithm that may be used for various issues across numerous fields. In classification and regression problems, it is
72
G. Venkataramana Sagar et al.
a nonparametric approach that is frequently employed. The nearest neighbors of a particular data point are found using a straightforward technique, and the values of those neighbors are then used to determine the class or label for that data point. – The classification of data is one of the primary uses of K-NN. If the K-NN model is trained correctly, it can classify category and numerical data. When there are numerous classes and the data points form intricate clusters, K-NN classification is beneficial. The K-NN model can correctly categorize incoming data points by locating the closest training data points and determining their class labels. Regression problems can also be performed with K-NN. The K-NN model can be used to forecast numerical values (like a stock’s price or a house’s value) rather than categorizing data points. The K-NN method can produce precise regression predictions by locating the closest neighbors and considering their numerical values. – Outlier detection is a further typical application of K-NN. Data points that do not fit the rest of the dataset can be found using K-NN. It is possible to determine whether a data point is an outlier by examining its neighbors. Finding flaws in datasets or new patterns might both benefit from this. – K-NN can also make product, film, or book recommendations. The K-NN algorithm can make product recommendations by identifying users who are similar to a given user. This can make it easier to match the requests to the user’s tastes. Powerful ML algorithms like K-NN can identify outliers, solve regression and classification problems, and recommend products. It can process various data kinds and produce reliable answers, primarily when the data is organized into complicated clusters or contains many classes.
3.9 Ensemble methods A range of approaches called ensemble methods is used to integrate several prediction models to produce a single, more potent model. Usually, this prediction model is more precise than any other model. The group of models collaborates to provide superior predictions by taking advantage of the advantages of various ML methods. Additionally, ensemble approaches are known to lessen overfitting because they combine multiple models to produce a more reliable and generalizable model. For ensemble methods to function, a base model must first be trained before being combined with additional base models using a meta-learning process. Careful selection of the basic models is necessary to produce a powerful ensemble. Every model ought to be distinct and reliable in its own right. The random forest algorithm is an illustration of an ensemble method. It integrates different decision trees to produce a more reliable and precise prediction model.
3 Classification and regression algorithms
73
Due to its accuracy, the random forest has gained popularity as a method for various applications. Ensemble approaches come in many different varieties. Bagging, boosting, stacking, and blending are a few of these [25]. Bagging makes use of statistical averaging and bootstrapping, which both serve to lessen overfitting. Boosting examines the significance of incorrectly categorized data items and bolsters their predictions. Stacking integrates many models, including neural networks and decision trees, to create a more potent prediction model. Through blending, the predictions from each model are combined and weighted according to how accurate they were. A robust approach for increasing the accuracy of models is ensemble methods. They are getting increasingly popular because they provide a technique to build a more robust, specialized model out of a collection of weaker, varied models. They are used in various applications, such as medical ones, speech recognition, NLP, picture and text recognition, and so on.
3.9.1 Bagging The bagging algorithm creates ensembles of classifiers from a single base model. It randomly selects features from a given dataset or sampling the training set. It then uses the chosen features to train the base model multiple times, resulting in numerous trained models. The predictions from each of these models are then combined to give an overall better estimation or classification than any models can provide. This algorithm can be used for any type of ML task, such as classification, regression, or clustering, as long as it produces a set of labeled data. In particular, it is used for classification tasks because it can take multiple “weak” classifiers and combine them to create a “strong” overall classification result. Bagging can reduce the variance of a single classifier by combining the votes of multiple models. It is beneficial for complex and nonlinear datasets. By using many different models trained on different subsets of the training data, the variance of the predictions from each classifier is reduced. Also, bagging can reduce the time required to train a model since it only requires a smaller subset of the training data for each base model. Bagging is a powerful ensemble learning method that can improve the accuracy of an ML model while retaining low bias and managing its variance. It can make complex datasets more accurate, provided that its base model is already best-suited for the problem. It is an essential technique for improving the performance of an ML model and can be used for various ML tasks such as classification, regression, and clustering.
3.9.2 Boosting An ML meta-algorithm called boosting is created to increase the precision of other ML algorithms. Its foundation is the concept of creating a high-accuracy ensemble by inte-
74
G. Venkataramana Sagar et al.
grating several low-accuracy models and boosting works by training each model independently using various subsets of the same dataset. The algorithm introduces the models iteratively, with each new model concentrating on the flaws of the preceding models. Each model is initially given equal weight, but the consequences are later changed such that the most accurate models have a more significant impact on the ensemble’s decisions. This indicates that the models improve and become more accurate as the algorithm develops since each new model makes up for flaws in the preceding one. Boosting can also be used for regression and extreme learning machines, although it is frequently employed to increase accuracy in classification tasks like recognizing spam emails. It often works with other ML methods, including decision trees, neural networks, and SVMs. Boosting can be used to address a wide range of issues. It can be used, for instance, to determine which features in a multidimensional dataset are the most crucial [26]. It can also increase accuracy for applications like automatic text classification, object identification, and picture processing. The ensemble can create a robust system that can accurately classify or identify complicated patterns by integrating several weak models. Combining soft models into strong ensembles using the ML approach of boosting is powerful and effective. It is an essential tool in the ML toolbox and is frequently used to increase accuracy in challenging jobs.
3.9.3 Stacking A powerful ML technique called the stacking ensemble method combines the output from various models to provide a combined forecast that is more precise. With this technique, data scientists can generate super-predictive models from weak learners that can be mixed and stacked to produce a super-model that may be more accurate than any of its component models. The stacking ensemble approach, in contrast to conventional ML algorithms, uses “meta-learning” to improve the overall model based on the outcomes of each fundamental model. It begins by training several basic models, such as decision trees or simple regression, before calculating the weighted average of all the forecasts to arrive at the final prediction. This approach enables data scientists to identify minute details and variations in the data, resulting in greater accuracy. The main benefit of the stacking ensemble method is that it allows for creating of highly predictive models by combining the results of several learners. Since different learners can be used to find various patterns or features within the data, it is also a valuable tool for handling multiple sorts of data. Additionally, it combines many learning techniques to outperform single models regarding predicted accuracy. The complexity of the stacking ensemble approach is its main flaw. To train numerous learners, the data scientist must devote time to model selection, feature engineering, hyperparameter optimization, and other associated processes [27]. As a result, it takes longer and requires more effort than traditional ML techniques. Combining numerous weak learners, the stacking ensemble approach is a potent tool for creating accurate predictive models.
3 Classification and regression algorithms
75
The complexity of this method can make it difficult. Hence it is best employed by data scientists with experience.
3.9.4 Voting classifiers The results from numerous additional classifiers are combined using a voting classifier and an ML algorithm to produce a consensus forecast. The voting classifier’s prediction is made using the majority rule, where the class with the most votes is selected as the output label. Voting classifiers have drawn interest in the ML community because of their ease of use and potential to increase prediction accuracy over a single classifier, theoretically. By merging several weak classifiers, the classifiers will perform better on a given task than any single classifier. Voting classifiers can be applied to various studies, including classification, regression analysis, and anomaly detection. Voting classifiers have the benefit of requiring little work from the user. The majority of classifiers determine the forecast. Thus, the user can combine a variety of classifiers to get a reasonably accurate prediction. Voting classifiers can quickly adapt to new data. Therefore, they only require a little fine-tuning. Due to the variety of classifiers used in the combination, voting classifiers tend to be more robust to noise [28]. Better generalization performance is the consequence, which is advantageous for data sets with high levels of unpredictability. Voting classifiers have certain downsides, nevertheless, in addition to these advantages. Their main drawback is that they use more computer resources because multiple classifiers need to be trained and assessed using various data sets. The majority vote will likewise be weak if the classifier contains weak classifiers because they are only as good as their pieces. Additionally, although voting classifiers might boost a weak classifier’s performance, feature-based fusion is sometimes a preferable way to merge classifiers at a higher order level.
3.9.5 Applications and use cases An ML technique called the ensemble technique enables the use of numerous models in combination to produce more accurate predictions than the models used alone. Ensemble approaches aggregate the outcomes of various models for increased forecast accuracy. Applications for ensemble methods include sentiment analysis, picture classification, image recognition, recommender systems, and many more. Sentiment analysis is one such application for ensemble methods. To increase the precision of predictions in sentiment analysis, ensemble methods can combine the output of various models. The production of numerous sentiment analysis models, such as those based on NLP and ML algorithms, can be combined in a sentiment analysis system, for example. This makes it easier to comprehend the sentiment expressed in an input text document more precisely. Two more applications of ensemble methods are pic-
76
G. Venkataramana Sagar et al.
ture classification and image recognition [29]. Using image classification, models can be combined to identify whether an image contains a specific object or scenario. Like text recognition, image recognition algorithms can be integrated to identify and categorize items in photos accurately. Another practical application for ensemble algorithms is recommender systems. Recommender systems use ensemble models to aggregate the findings of several models and make user-specific product recommendations. As a result, the algorithm may produce more precise proposals. Applications for the ensemble methods include sentiment analysis, picture classification, image recognition, and recommender systems, among many more. Ensemble methods can produce extremely precise forecasts by combining the outcomes of various models.
3.10 Regression algorithms Regression algorithms are supervised learning techniques applied to supervised learning issues like forecasting a continuous outcome (like the price of a stock) or a discrete outcome (like the lifetime value of a customer). They employ mathematical models to discover relationships from raw data and create predictions. Decision trees, logistic regression, and linear regression are some popular regression techniques.
3.10.1 Linear regression A statistical technique for examining associations between two or more variables is linear regression. If all other factors have been considered, it can estimate the impact of one variable on another. Due to its universality and adaptability, it is a technique that is used most often in statistical analysis. Linear regression can take many different shapes depending on the type of data being analyzed and the nature of the relationship between the variables. Ordinary least squares (OLS) regression is the most basic type of linear regression, and it simply involves fitting a straight line to a scatter plot of the data points to represent the connection between two variables. This reveals the best-fitting line and the nature of the correlation between the two variables. Additionally, more sophisticated linear regression models, including multiple linear and polynomial regressions, can analyze data from more than two variables. These models use more complicated functions and can give more in-depth explanations of how the variables relate to one another. A potent tool for prediction is linear regression. On a given dataset, it can be used to predict pertinent answers by fitting a linear model to the data points. The development of risk-reduction or opportunityseizing tactics can then be done using this prediction. Finally, any relationship between variables, whether linear, nonlinear, or a combination of both, can be found using linear regression.
3 Classification and regression algorithms
77
Additionally, it can be used to find patterns, correlations, and trends in data that may take time to be apparent. In various fields, from economics to biology, linear regression is a flexible and necessary tool for data analysis [30]. It offers a handy method of examining the connections between factors and may aid researchers in discovering solutions to issues.
3.10.2 Polynomial regression Regression analysis that uses polynomials to represent the connection between a single dependent variable (y) and one or more independent variables (x) is known as polynomial regression. This approach allows us to fit models representing the data as linear, quadratic, cubic, or higher order polynomials. Polynomial regression might be helpful when data display a nonlinear pattern, such as when an object is speeding up, slowing down, or changing directions. It can also be used to examine situations when it is challenging to choose the correct sequence for the polynomial to employ. In reality, it can occasionally approximate a number that is not otherwise directly observable, such as the rate of acceleration in issues involving motion. This method allows us to fit a more accurate approximation of the connection than a linear regression would. Additionally, we can capture more specific information in our predictions using higher order polynomials, which could lead to more precise estimations for the dependent variable. Ensuring that our model is not overfitting the data is a crucial factor to consider while utilizing polynomial regression. When a model tries to account for too much data variance, it is said to be overfitting; this can produce inaccurate predictions and a poorer overall fit. Testing our model using cross-validation and leveraging techniques to prevent this is crucial. A potent method, polynomial regression is beneficial when working with data that exhibits a nonlinear pattern. We can capture more detail in our predictions and get a better fit by using higher order polynomials. Overfitting can be problematic. Thus, it is advisable to use cross-validation and leverage techniques to avoid it.
3.10.3 Support vector regression A supervised ML approach called support vector regression (SVR) is used for regression tasks. It is an SVM approach that transforms data into higher dimensional spaces from which it may then use linear regression to identify nonlinear patterns. The program divides the input data into two classes using hyperplanes and maps it to a higher dimensional space. The support vectors are employed to choose the best hyperplane to execute linear regression and divide the data into classes. The SVR algorithm can be applied to issues like projecting the price of a property or automobile, a stock
78
G. Venkataramana Sagar et al.
market trend, sales figures, etc. When data is not linear and linear regression cannot be used to solve the problem, this method is appropriate. SVR is a versatile model that works well to increase prediction accuracy. Additionally, it can be used to implement regularization and lower variance for more accurate outcomes quickly. SVR is less well-known than other ML methods, but it provides a better data representation that can aid in making more accurate predictions [31]. SVR is one of the most effective strategies for regression issues. In many situations, it can perform better than conventional ML methods and contribute to generating more accurate outcomes. Due to this, it is essential to anticipate stock prices and market trends in the financial industry.
3.10.4 Decision tree regression A supervised learning method called decision tree regression can be used to forecast continuous values. It operates by segmenting the input space into several parts (or segments) and predicting the average data response in each portion. The decision parameters for the specific problem determine how the partitions are divided and what values are assigned to the regions. Decision tree regression forecasts actual values rather than classes like classification tree functions. The concept behind learning a decision tree regression model to predict numerical values is to build a tree where each node represents a decision point or a rule. The model traces the tree from root to leaf to forecast the outcome. Decision tree regression is appropriate for a wide range of issues since it can handle nonlinear data and accept category and numerical characteristics.
3.10.5 Random forest regression A predictive model is fitted to a dataset using the random forest regression technique to evaluate the associations between variables and make predictions. An ensemble learning technique called a random forest combines numerous decision trees into a single model, where each tree is a highly predictive model customized to a different subset of the data. As it builds a model from labeled data (such as training data), the random forest illustrates a supervised learning technique. A new data point can then be added to the model without being labeled to forecast its class or value. Data science activities like classification, clustering, and regression may all be performed using random forest regression [32]. It can spot crucial details and connections between variables, spot outliers and missing data, and forecast new data points.
3 Classification and regression algorithms
79
3.10.6 Gradient boosting regression A powerful ensemble model is intended to be produced through the ML technique known as gradient boosting regression. It functions by creating new models in stages while using the mistakes of the prior model as training data. A robust model with high predictive ability is the end outcome, with each iteration marginally better than the previous one. It is an effective method that works well for many large data applications. Since it leverages complicated nonlinear properties of data that are frequently too complex for existing algorithms to comprehend, it often outperforms traditional statistical and ML techniques. It is typically less accurate and uses fewer processing resources.
3.10.7 Applications and use cases Continuous target variables are predicted using regression techniques from a given collection of characteristics. Logistic and linear regression are the two most widely used regression techniques. A constant target variable can be predicted using linear regression. It is employed to forecast an outcome in response to a predictor variable and to clarify the connection between two or more variables. Logistic regression is used to estimate the chance of an event occurring based on the values of a collection of predictor variables. Finding the class to which a specific observation belongs is the objective of this kind of classification problem. Other regression algorithms, such as polynomial, neural networks, and SVR, exist in addition to linear and logistic regression. These techniques can all be used to resolve many kinds of prediction issues [33]. Regression algorithms have numerous typical applications and use cases, such as predicting customer churn rates, server load, stock prices, marketing campaign performance, loan payback, and fraud detection. These algorithms have been widely applied in engineering, marketing, healthcare, and finance. Additionally, regression algorithms can be used to anticipate future sales, consumer behavior trends, and product demand.
3.11 Evaluation and model selection Any ML method must include the two essential steps of model selection and evaluation. It entails choosing a model that best matches the data and assessing its performance using quantitative criteria. Data science and ML algorithms are often applied to the data before using statistical methods to find the best match. Possible evaluation criteria are accuracy, sensitivity, specificity, precision, and recall. When choosing a model, it is also essential to consider how useful the model will be in practice, including how well it will generalize to new data and capture essential relationships in the
80
G. Venkataramana Sagar et al.
data. Following model selection, the model’s performance is assessed to determine whether it achieves the desired result. If not, adjustments may be made to make the model more appropriate for the given task.
3.11.1 Cross-validation Cross-validation sets aside a portion of the dataset while the model is trained using the remaining data. This technique is used to assess predictive models. The model is then evaluated using the held-out data. Every fold of the dataset is utilized as the test set once throughout this process, which is performed numerous times. Therefore, cross-validation enables you to gauge a model’s performance using hypothetical data. In actual use, the estimate of model performance improves as the number of crossvalidation folds increases, but often 5–10 folds are utilized. Cross-validation has several benefits, but its main advantages are that it is more accurate than a straightforward train-test split and more adaptable because it can be applied to a range of sample sizes. It is also a popular way of fine-tuning model parameters, such as choosing the ideal number of trees for a random forest.
3.11.2 Hyperparameter tuning The process of choosing the best combination of hyperparameters for an ML model to obtain the best performance is known as hyperparameter tuning. Hyperparameters are controls typically established before the learning process [34] and govern how the ML algorithm functions. The number of neurons and layers in a neural network, the number of trees in a random forest, or the learning rate in gradient descent are a few examples of hyperparameters. The performance of a model can be maximized if the ideal values for its hyperparameters have been identified.
3.11.3 Model evaluation metrics Metrics for model evaluation are methods for measuring how well an ML model performs against a collection of test or training data. These metrics are generally employed to evaluate many models and choose the one that best satisfies a user’s needs. Accuracy, precision, recall, F1 score, ROC (receiver operating characteristic)/AUC (area under the curve), and log loss are standard model evaluation metrics. They aid in determining the model’s suitability for a specific task. Different metrics should be employed depending on the type of ML work; for instance, accuracy is appropriate for classification tasks, whereas F1 score and log loss are more appropriate for ranking tasks.
3 Classification and regression algorithms
81
3.11.4 Overfitting and underfitting When a function is too closely fitted to a small number of data points, overfitting, a modeling error, occurs. As a result, the model can correctly predict outcomes using the training data set, but it cannot generalize and predict effects using new data. Underfitting occurs when a statistical model cannot fully represent the underlying structure of the data. It indicates that the model needs to fit the data adequately and sufficiently account for much of the variation. Due to the model’s inability to account for the complexity of the data, underfitting causes erroneous predictions and a significant bias [35].
3.11.5 Model selection strategies Model selection strategies are collections of methods for selecting the best model from a pool of candidate models. In addition to statistical criteria like cross-validation and the Akaike information criterion, the procedures can include heuristics like the Bayesian information criterion and stepwise regression and visual tools like parallel coordinates and scree plots. In the end, bias and variance are traded off when choosing a model. A model is said to be overfitting if its tendency is low but its conflict is high, whereas it may be underfitting if its bias is high but its variance is low. The choice of a model is a crucial stage in predictive modeling because it can significantly increase its accuracy.
3.12 Conclusion and future directions Developing and applying more effective model selection techniques are critical to the field’s future. This would entail figuring out how to choose the ideal model for a particular situation precisely. This can entail applying sophisticated algorithms or employing various metrics to gauge the model’s effectiveness. In the future, selecting models will also emphasize enhancing interpretability and creating models that may be used to describe decision-making processes. Evaluating and choosing models using interactive models and visualization tools will also become more typical. The use of alternative datasets, including other data sources and techniques for inferring them to suit the data best, will then be further investigated.
82
G. Venkataramana Sagar et al.
3.12.1 Summary of classification and regression algorithms ML tools, such as classification and regression algorithms, are used to examine data and create predictions. Classification algorithms are used to categorize data into distinct groups, such as assessing whether or not an email is spam. Regression algorithms are employed to predict continuous quantities, such as the stock price or the volume of website visits. Regression and classification algorithms are frequently combined to create AI systems that are better capable of making precise predictions. Typical classification techniques include decision trees, logistic regression, and K-NN. SVMs and polynomial regression are frequently used regression techniques [36].
3.12.2 Emerging trends and advancements Regression and classification algorithms have advanced significantly in recent years. These algorithms have improved in complexity, accuracy, and effectiveness. ANN and deep learning are two growing trends in classification and regression techniques used to increase accuracy and decrease overfitting. SVMs are also gaining popularity since they have demonstrated excellent capability to manage large-scale, complex data sets. Additionally, ensemble method innovations like boosting, bagging, and stacking have been shown to increase predictive power and accuracy. Understanding the link between the independent factors and dependent variables has also benefited from the development of new data pre-processing and validation methods [37]. We have successfully selected significant features and enhanced the performance of ML models by using feature selection approaches, such as recursive feature removal. Crossvalidation has been made simpler by the new technologies, such as grid search, and has allowed us to gauge how well different algorithms perform more precisely.
3.12.3 Challenges and future research areas 3.12.3.1 Challenges Overfitting, selection bias, data noise, and interpretability are a few of the problems that classification and regression algorithms must deal with. An ML algorithm overfits the training set when it fits the data too closely, producing models that do not generalize well to new data [38]. Selection bias might happen when the data used to build the model do not represent the actual population. When there is a sufficient number of valid training data or an excessive number of unimportant characteristics, data noise may improve the model’s accuracy and interpretability.
3 Classification and regression algorithms
83
3.12.3.2 Future research Classification and regression algorithms have a wide range of possible applications in the future. Among these are creating reliable algorithms that can deal with noisy and missing data, creating advanced ensemble techniques, and figuring out how to understand the outcomes of ML models better. Additionally, research might look into combining neural networks with existing ML algorithms and applying unsupervised ML approaches for classification and regression problems.
References [1]
Strecht, P., Cruz, L., Soares, C., Mendes-Moreira, J., & Abreu, R. (2015). A comparative study of classification and regression algorithms for modelling students’ academic performance. Portugal: International Educational Data Mining Society. [2] Loh, W. Y. (2011). Classification and regression trees. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, 1(1), 14–23. [3] Hoffmann, F., Bertram, T., Mikut, R., Reischl, M., & Nelles, O. (2019). Benchmarking in classification and regression. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, 9(5), e1318. [4] Loh, W. Y. (2014). Fifty years of classification and regression trees. International Statistical Review, 82(3), 329–348. [5] Pathak, S., Mishra, I., & Swetapadma, A. (2018, November). An assessment of decision tree based classification and regression algorithms. In 2018 3rd International Conference on Inventive Computation Technologies (ICICT) (pp. 92–95). IEEE. [6] Zhou, Z. H., & Li, M. (2007). Semisupervised regression with cotraining-style algorithms. IEEE Transactions on Knowledge and Data Engineering, 19(11), 1479–1493. [7] Liaw, A., & Wiener, M. (2002). Classification and regression by random forest. R News, 2(3), 18–22. [8] Loh, W. Y. (2008). Classification and regression tree methods. Encyclopedia of Statistics in Quality and Reliability, 1, 315–323. [9] Ponnusamy, M., Bedi, P., Suresh, T., Alagarsamy, A., Manikandan, R., & Yuvaraj, N. (2022). Design and analysis of text document clustering using salp swarm algorithm. The Journal of Supercomputing, 78(14), 16197–16213. [10] Rocha, M., Cortez, P., & Neves, J. (2007). Evolution of neural networks for classification and regression. Neurocomputing, 70(16–18), 2809–2816. [11] Ren, Y., Zhang, L., & Suganthan, P. N. (2016). Ensemble classification and regression-recent developments, applications and future directions. IEEE Computational Intelligence Magazine, 11(1), 41–53. [12] Jaiswal, J. K., & Samikannu, R. (2017, February). Application of random forest algorithm on feature subset selection and classification and regression. In 2017 World Congress on Computing and Communication Technologies (WCCCT) (pp. 65–68). IEEE. [13] Li, H., Sun, J., & Wu, J. (2010). Predicting business failure using classification and regression tree: An empirical comparison with popular classical statistical methods and top classification mining methods. Expert Systems with Applications, 37(8), 5895–5904. [14] Mohammed, V. A., Mohammed, M. A., Mohammed, M. A., Logeshwaran, J., & Jiwani, N. (2023, February). Machine learning-based evaluation of heart rate variability response in children with
84
[15] [16] [17] [18] [19]
[20] [21] [22]
[23]
[24] [25]
[26]
[27]
[28]
[29]
[30]
[31]
G. Venkataramana Sagar et al.
autism spectrum disorder. In 2023 Third International Conference on Artificial Intelligence and Smart Energy (ICAIS) (pp. 1022–1028). IEEE. Orabona, F., Crammer, K., & Cesa-Bianchi, N. (2015). A generalized online mirror descent with applications to classification and regression. Machine Learning, 99, 411–435. Chou, P. A. (1991). Optimal partitioning for classification and regression trees. IEEE Transactions on Pattern Analysis and Machine Intelligence, 13(04), 340–354. Deziel, M., Olawo, D., Truchon, L., & Golab, L. (2013, July). Analyzing the mental health of engineering students using classification and regression. In Educational Data Mining 2013. Timofeev, R. (2004). Classification and regression trees (CART) theory and applications. Berlin: Humboldt University, 54. Corrales, D. C., Ledezma, A., & Corrales, J. C. (2020). A case-based reasoning system for recommendation of data cleaning algorithms in classification and regression tasks. Applied Soft Computing, 90, 106180. Thomas, P., & Suhner, M. C. (2015). A new multilayer perceptron pruning algorithm for classification and regression applications. Neural Processing Letters, 42, 437–458. Yang, Y. H., Lin, Y. C., Su, Y. F., & Chen, H. H. (2007, July). Music emotion classification: A regression approach. In 2007 IEEE International Conference on Multimedia and Expo (pp. 208–211). IEEE. Rani, P., Verma, S., Yadav, S. P., Rai, B. K., Naruka, M. S., & Kumar, D. (2022). Simulation of the Lightweight Blockchain Technique Based on Privacy and Security for Healthcare Data for the Cloud System. International Journal of E-Health and Medical Communications (IJEHMC), 13(4), 1–15. Kannan, S., Yuvaraj, N., Idrees, B. A., Arulprakash, P., Ranganathan, V., Udayakumar, E., & Dhinakar, P. (2021). Analysis of convolutional recurrent neural network classifier for COVID-19 symptoms over computerised tomography images. International Journal of Computer Applications in Technology, 66(3–4), 427–432. Reddy, C. K., & Park, J. H. (2011). Multi-resolution boosting for classification and regression problems. Knowledge and Information Systems, 29, 435–456. Khan, R. A., Ahmad, N., & Minallah, N. (2013). Classification and regression analysis of the prognostic breast cancer using generation optimizing algorithms. International Journal of Computer Applications, 68(25), 42–47. Ghasemain, B., Asl, D. T., Pham, B. T., Avand, M., Nguyen, H. D., & Janizadeh, S. J. V. J. O. E. S. (2020). Shallow landslide susceptibility mapping: A comparison between classification and regression tree and reduced error pruning tree algorithms. Vietnam Journal of Earth Sciences, 42(3), 208–227. Balda, E. R., Behboodi, A., & Mathar, R. (2019). Perturbation analysis of learning algorithms: Generation of adversarial examples from classification to regression. IEEE Transactions on Signal Processing, 67(23), 6078–6091. Liu, Q., Wang, X., Huang, X., & Yin, X. (2020). Prediction model of rock mass class using classification and regression tree integrated AdaBoost algorithm based on TBM driving data. Tunnelling & Underground Space Technology, 106, 103595. Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). A novel architecture of intelligent decision model for efficient resource allocation in 5G broadband communication networks. ICTACT Journal On Soft Computing, 13(3), 2986–2994. Yadav, S. P., & Yadav, S. (2019). Mathematical implementation of fusion of medical images in continuous wavelet domain. Journal of Advanced Research in Dynamical and Control System, 10(10), 45–54. Priyam, A., Abhijeeta, G. R., Rathee, A., & Srivastava, S. (2013). Comparative analysis of decision tree classification algorithms. International Journal of Current Engineering and Technology, 3(2), 334–337.
3 Classification and regression algorithms
85
[32] Zhang, L., Huettmann, F., Zhang, X., Liu, S., Sun, P., Yu, Z., & Mi, C. (2019). The use of classification and regression algorithms using the random forests method with presence-only data to model species’ distribution. MethodsX, 6, 2281–2292. [33] Gunjan, V. K., Vijayalata, Y., Valli, S., Kumar, S., Mohamed, M. O., & Saravanan, V. (2022). Machine Learning and Cloud-Based Knowledge Graphs to Recognize Suicidal Mental Tendencies. Computational Intelligence and Neuroscience, 2022, 1–12 [34] Lou, Y., Caruana, R., & Gehrke, J. (2012, August). Intelligible models for classification and regression. In Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (pp. 150–158). [35] Yang, T., Gao, X., Sorooshian, S., & Li, X. (2016). Simulating California reservoir operation using the classification and regression‐tree algorithm combined with a shuffled cross‐validation scheme. Water Resources Research, 52(3), 1626–1651. [36] Singh, J., Deepika, J., Sathyendra Bhat, J., Kumararaja, V., Vikram, R., Jegathesh Amalraj, J. . . . Sakthivel, S. (2022). Energy-efficient clustering and routing algorithm using hybrid fuzzy with grey wolf optimization in wireless sensor networks. Security and Communication Networks, 2022, 1–13 [37] Grubinger, T., Zeileis, A., & Pfeiffer, K. P. (2014). evtree: Evolutionary learning of globally optimal classification and regression trees in R. Journal of Statistical Software, 61, 1–29. [38] Yusoff, M. H., Chrol-Cannon, J., & Jin, Y. (2016). Modeling neural plasticity in echo state networks for classification and regression. Information Sciences, 364, 184–196.
Ashwin M., Nazeer Shaik, Faiz Akram, Getachew Mamo Wegari
4 Clustering and association algorithm Abstract: This chapter provides an extensive examination of clustering and association algorithms, offering a comprehensive understanding of these essential unsupervised learning techniques in the field of machine learning. The chapter commences with an introductory section that underscores the significance of unsupervised learning, and presents a detailed overview of clustering and association algorithms. The various clustering algorithms covered in the following sections are k-means, hierarchical clustering, densitybased clustering (like DBSCAN), and Gaussian mixture models. Additionally, the chapter explores evaluation metrics for clustering, including both internal and external evaluation metrics, and offers advise on how to properly interpret the evaluation results. The chapter also examines dimensionality reduction methods, particularly principal component analysis and t-SNE, and clarifies how to use them in clustering scenarios. The Apriori algorithm, frequent itemset mining, association rule development, and methods for judging rule interestingness are all covered in the section on association rule learning. Along with explaining their individual uses, the chapter also goes into advanced clustering methods like density-based spatial clustering, mean-shift clustering, spectral clustering, and affinity propagation. The chapter also discusses association rule mining, ensemble clustering, and time series clustering in huge datasets. The chapter concludes with an overview of the core ideas and developments in unsupervised learning, along with a list of the field’s present research roadblocks and potential future paths. Keywords: Clustering, association, algorithm, machine learning, research
4.1 Introduction 4.1.1 Overview of clustering and association algorithms Data mining and ML employ the potent clustering and association algorithms. In order to find underlying structures or patterns in the data, clustering algorithms put comparable data points together according to their properties [1]. K-means, DBSCAN, and hierarchical clustering are popular techniques for clustering data. These algorithms analyze the proximity or similarity between data points to form clusters, enabling insights into the natural groupings present in the data, as in Figure 4.1. On the other hand, association algorithms focus on identifying relationships or associations between different items or variables in a dataset. Apriori is the most well-known association algorithm; it finds frequent item sets and provides association rules, based on support and confidence metrics. The co-occurrence of items or variables in the dataset is indicated by these rules. https://doi.org/10.1515/9783111323749-004
88
Ashwin M. et al.
Large Data set
Clustering 1
Clustering 2
Clustering N
Final Clustering result
Figure 4.1: Clustering and association.
Both clustering and association algorithms are crucial in exploratory data analysis, customer segmentation, market basket analysis, and recommendation systems [2]. They provide valuable insights and enable decision-making based on patterns, relationships, and groupings present in the data.
4.1.2 Importance of unsupervised learning in machine learning Unsupervised learning [3] plays a crucial role in machine learning and has significant importance for several reasons: 1. Discovering hidden patterns: Unsupervised learning algorithms enable the identification of hidden patterns and structures within data without the need for labeled examples. This can lead to valuable insights and discoveries, especially in cases where the underlying patterns are unknown or difficult to define. 2. Data exploration and preprocessing: Unsupervised learning techniques provide tools for data exploration, visualization, and preprocessing. Clustering algorithms, for example, help identify natural groupings in the data, enabling a better understanding of the data’s characteristics and potential relationships. 3. Anomaly detection: Unsupervised learning is useful in detecting anomalies or outliers in datasets. By learning the normal patterns within the data, unsupervised
4 Clustering and association algorithm
4.
5.
6.
89
algorithms can flag instances that deviate significantly from the expected behavior, helping in fraud detection, fault diagnosis, and cybersecurity. Dimensionality reduction: Numerous unsupervised learning algorithms, like principal component analysis (PCA) and t-SNE, help to decrease the dimensionality of high-dimensional data while maintaining its fundamental structure. This makes subsequent supervised learning tasks more efficient and makes feature selection and data visualization easier. Recommendation systems: Unsupervised learning plays a vital role in building recommendation systems. Collaborative filtering techniques, such as matrix factorization and clustering-based methods, can analyze user behavior and preferences to provide personalized recommendations without requiring explicit labels or ratings. Data generation and synthesis: The distribution of the original data can be used to create synthetic data using unsupervised learning methods. This is especially beneficial when there is a lack of labeled data or when enhancing training datasets for supervised learning tasks.
In summary, unsupervised learning techniques are essential in exploring and understanding complex datasets, detecting anomalies, reducing dimensionality, building recommendation systems, and generating synthetic data. They provide valuable insights and solutions in various domains, even when labeled data is limited or unavailable.
4.2 Clustering algorithms Clustering algorithms are a fundamental part of unsupervised learning in ML [4]. They group similar data points together based on their characteristics, allowing the discovery of inherent structures or patterns in the data. By analyzing the proximity or similarity between data points, clustering algorithms form clusters that represent natural groupings within the data. Hierarchical clustering, DBSCAN, and K-means are popular clustering techniques. They play a crucial role in exploring and understanding data, enabling insights and decision-making, based on patterns and relationships present in the data.
4.2.1 K-means clustering A popular approach for dividing a dataset into K different clusters is K-means clustering. The final centroids represent the cluster centers, and each data point is associated with the cluster that has the closest centroid. K-means is efficient and scalable for large datasets, but it has limitations. It assumes a spherical shape of clusters, is sensi-
90
Ashwin M. et al.
tive to the initial centroid positions, and may converge to local optima. However, it remains a popular algorithm for clustering tasks in various domains due to its simplicity and effectiveness [5].
4.2.2 Hierarchical clustering A clustering algorithm called hierarchical clustering seeks to build a hierarchical structure of clusters within a dataset [6]. It is adaptable and appropriate for exploratory analysis because it does not require a set number of clusters. The fundamental structure of the data is captured via hierarchical clustering, enabling the identification of clusters at various granularities. The dendrogram offers insights into the relationships between clusters and specific data points, making it particularly helpful for displaying and comprehending complex datasets. However, for huge datasets, it can be computationally expensive.
4.2.3 Density-based clustering (e.g., DBSCAN) Discovering clusters based on the density of data points in a dataset is the goal of density-based clustering algorithms like DBSCAN (Density-Based Spatial Clustering of Applications with Noise). DBSCAN, in contrast to other clustering methods, does not presume that clusters have a particular shape and can find clusters of any shape [7]. DBSCAN is capable of handling noise in the data and can identify clusters of varying density. It automatically determines the number of clusters based on the data characteristics. Nevertheless, the performance of DBSCAN can be influenced by the choice of its parameters, including ’“epsilon” and “minPts.”. Overall, DBSCAN is a valuable algorithm for discovering clusters in complex and irregularly shaped datasets.
4.2.4 Gaussian mixture models Probabilistic models, known as Gaussian mixture models (GMMs), are used to represent and cluster data that is believed to have been created from a combination of Gaussian distributions. GMMs are widely used in various ML applications, including clustering, density estimation, and generative modeling. GMMs are flexible and can model complex distributions by combining multiple Gaussian components. They can handle overlapping or elongated clusters in the data. GMMs are useful for clustering tasks where the underlying data distribution is not easily separable or when there is uncertainty in cluster assignment. However, GMMs require the assumption that the data points are generated from Gaussian distributions, which may not hold in all cases.
4 Clustering and association algorithm
91
4.2.5 Applications and use cases The use of clustering algorithms is widespread in many different industries. Several well-known application cases include: 1. Image and object recognition: A dataset’s comparable images or objects can be found and categorized using clustering methods. This has applications in image recognition, content-based image retrieval, and object detection. 2. Anomaly detection: Clustering can help identify unusual or anomalous patterns in data that deviate from the expected behavior. This is useful in fraud detection, network intrusion detection, and identifying faulty equipment in manufacturing processes. 3. Market basket analysis: Clustering algorithms are applied to analyze transactional data and identify associations or co-occurrences between products. This information is used in market basket analysis to generate product recommendations, optimize product placement, and improve cross-selling opportunities. 4. Document clustering and topic modeling: Clustering algorithms can group similar documents based on their content, facilitating document organization, information retrieval, and topic modeling. It helps identify themes or topics within large document collections. 5. Social network analysis: Based on their relationships or interactions, communities or groups of people are identified in social networks using clustering. Understanding network topologies, influence analysis, and focused marketing campaigns are made easier by doing so. 6. Genetic and bioinformatics analysis: Clustering algorithms are applied to analyze genetic data, DNA sequences, and gene expression profiles. They help identify patterns, gene clusters, and genetic relationships, contributing to advancements in genomics and personalized medicine. 7. Recommender systems: Clustering techniques play a role in building collaborative filtering-based recommender systems. They group similar users or items to generate personalized recommendations, based on user preferences or item similarities. These are only a few of the numerous uses that clustering algorithms have. Clustering is a useful technique in many disciplines because it can reveal patterns, groupings, and linkages within data, assisting with decision-making, pattern detection, and data exploration.
92
Ashwin M. et al.
4.3 Evaluation metrics for clustering Clustering algorithms’ efficiency and quality are evaluated using evaluation measures [8, 9]. Some commonly used metrics include:
4.3.1 Internal evaluation metrics (e.g., Silhouette score and Dunn index) Internal evaluation metrics for clustering assess the quality of clusters, based solely on data and without relying on external or ground truth information. Here are some commonly used internal evaluation metrics, as in Figure 4.2.
Internal Evaluation Metrics
Silhouette Score
Dunn Index
Davies-Bouldin Index
Calinski-Harabasz Index
Figure 4.2: Internal evaluation metrics.
1.
2. 3.
4.
Silhouette score: The compactness and separation of clusters are assessed using the Silhouette score. Better-defined clusters are indicated by a higher Silhouette score. Dunn index: The Dunn index evaluates the compactness and separation of clusters. Davies-Bouldin index: Based on the ratio of the average dissimilarity across clusters to the greatest intra-cluster dissimilarity, the Davies-Bouldin Index evaluates the clustering quality. Lower numbers represent more well-defined clusters. Calinski-Harabasz index: To increase inter-cluster separation and reduce intracluster variation, this index evaluates the ratio of between-cluster dispersion to within-cluster dispersion. Higher Calinski-Harabasz scores indicate better clustering quality.
These internal evaluation metrics provide insights into the cohesion and separation of clusters, helping to determine the effectiveness of clustering algorithms without requiring external reference information. It is important to consider multiple metrics and compare results across different algorithms or parameter settings for a comprehensive evaluation.
4 Clustering and association algorithm
93
4.3.2 External evaluation metrics (e.g., Rand index and F-measure) External evaluation metrics for clustering compare the clustering results against a ground truth or reference clustering, when available [10–12]. These metrics assess the agreement between the obtained clusters and the known or expected clusters. Here are some commonly used external evaluation metrics for clustering, as in Figure 4.3:
External Evaluation Metrics
Rand Index (RI)
Adjusted Rand Index (ARI)
F-measure
Jaccard Coefficient
Figure 4.3: External evaluation metrics.
1.
2. 3.
4.
Rand index (RI): The RI measures the similarity between two clusterings by counting the number of agreements (both in the same cluster or in different clusters) and disagreements (one in the same cluster, the other in a different cluster) between the two clusterings. The index ranges from 0 to 1, where 1 indicates a perfect match. Adjusted RI (ARI): The ARI accounts for the expected agreement between random clusterings, and modifies the RI for chance agreement. F-measure: To rate the accuracy of clustering results, the F-measure combines precision and recall. By taking into account both the completeness of each cluster and the purity of each class inside a cluster, it determines the harmonic mean of the two measures. Jaccard coefficient: Based on the intersection and union of data point assignments, the Jaccard coefficient gauges the similarity between two clusterings.
External evaluation metrics provide a quantitative assessment of clustering quality when ground truth information is available. They are useful for comparing different clustering algorithms and parameter settings, as well as for validating the results against known clusters [12].
94
Ashwin M. et al.
4.3.3 Interpretation of evaluation results Interpreting evaluation results for clustering involves understanding the performance of clustering algorithms and assessing the quality of the obtained clusters. Here are some key points to consider when interpreting clustering evaluation results: 1. Consistency across multiple metrics: It is essential to consider multiple evaluation metrics and look for consistent results across them. If different metrics consistently indicate good clustering quality, it increases confidence in the reliability of the results. 2. Comparison to baseline or random results: Comparing the obtained clustering results to a baseline or random clustering is important. If the evaluation scores significantly outperform the random or baseline results, it indicates that the algorithm is effective in clustering the data. 3. Domain knowledge and context: Understanding the domain and the specific characteristics of the dataset is crucial for interpretation. Evaluate if the obtained clusters align with domain-specific knowledge, expected patterns, or known ground truth. Expert judgment can provide valuable insights into the meaningfulness and usefulness of the clusters. 4. Visual exploration: Visualization techniques, such as scatter plots or dendrograms, can help visually inspect the clustering results. Visual examination allows for a better understanding of the cluster structure, potential outliers, and the overall separation and compactness of the clusters. 5. Limitations and trade-offs: It is important to consider the limitations and assumptions of the clustering algorithm. Some algorithms may work well for certain types of data or cluster structures but may struggle with others. Understanding the trade-offs between different evaluation metrics and considering the specific requirements of the application is crucial. Interpreting clustering evaluation results requires a comprehensive analysis, considering multiple factors such as evaluation scores, domain knowledge, visual exploration, and algorithm limitations. It is crucial to take a holistic approach and make informed judgments based on these considerations.
4.4 Association rule learning and evaluation metrics A data mining technique called association rule learning identifies intriguing correlations, patterns, and relationships between items in huge datasets [13]. It entails locating frequent itemsets, groups of items that commonly appear together. Association rules are produced from these frequently occurring itemsets, defining the co-occurrence associations between items [14]. The rules consist of an antecedent (the items on the left-
4 Clustering and association algorithm
95
hand side) and a consequent (the items on the right-hand side). These rules provide insights into the dependencies and associations between items, enabling applications such as market basket analysis, recommendation systems, and understanding customer behavior in retail, e-commerce, and various domains [15].
4.4.1 Apriori algorithm A common approach used for frequent itemset mining in association rule learning is the Apriori algorithm. It effectively finds common itemsets by using a breadth-first search approach.
4.4.2 Frequent itemset mining The process of identifying groups of items that frequently appear together in a dataset is known as frequent itemset mining. The support calculation for itemsets is the primary equation utilized in frequent itemset mining. To perform frequent itemset mining, the process typically involves the following steps: Step 1. Generating candidate itemsets Initially, individual items are considered as candidate itemsets. Then, based on the Apriori principle, larger candidate itemsets are generated by combining the frequent itemsets from the previous iteration. Step 2. Pruning Pruning is done by checking if all subsets of a candidate itemset are frequent. Step 3. Iteration The process iterates by generating larger candidate itemsets, counting their support, and pruning until no more frequent itemsets can be found. The frequent itemsets discovered through this process provide insights into the cooccurrence relationships between items, which can be used to generate association rules and understand patterns in the data.
4.4.3 Association rule generation Association rule generation involves deriving rules that express relationships between items based on the frequent itemsets discovered through frequent itemset min-
96
Ashwin M. et al.
ing. The primary equation used in association rule generation is the confidence calculation. Here is the equation for confidence: ConfidenceðAntecedent ! ConsequentÞ = ðSupportðAntecedent ∪ ConsequentÞÞ = ðSupportðAntecedentÞÞ By setting a minimum confidence threshold, only association rules with confidence values above the threshold are considered interesting and significant. The process of association rule generation typically involves the following steps: Step 1. Selection of frequent itemsets First, frequent itemsets are identified through frequent itemset mining using a minimum support threshold. These frequent itemsets serve as the basis for rule generation. Step 2. Rule generation By taking into account various combinations of items as antecedents and consequents, association rules are constructed from frequent itemsets. Each common item set has the potential to generate several rules. Step 3. Confidence calculation The equation shown above is used to determine each association rule’s confidence level. In this step, the relationship between the antecedent and the consequent is quantified. Step 4. Pruning Association rules are trimmed if they do not match the minimum confidence criterion since they are viewed as being less significant or reliable. The generated association rules provide insights into the dependencies and co-occurrence relationships between items. They help in understanding the behavior and preferences of customers, market basket analysis, cross-selling, and various decision-making processes. In summary, association rule generation involves calculating the confidence of association rules using the support of itemsets. By setting a minimum confidence threshold, meaningful and significant rules can be extracted from the frequent itemsets, providing valuable insights into item associations and patterns in the data.
4.4.4 Applications and use cases Numerous disciplines can benefit from association rule mining, which has a wide range of applications. Here are some typical use cases and applications for association rule mining.
4 Clustering and association algorithm
97
4.4.5 Market basket analysis Association rule mining is extensively used in retail and e-commerce for market basket analysis. It helps retailers identify frequently co-occurring items in customer transactions, enabling them to understand purchasing patterns, recommend related products, optimize store layouts, and implement effective cross-selling strategies.
4.4.6 Customer behavior analysis Association rules aid in analyzing customer behavior and preferences. By identifying associations between products, services, or actions taken by customers, businesses can personalize marketing campaigns, improve customer segmentation, and offer targeted recommendations, thereby enhancing customer satisfaction and loyalty.
4.4.7 Web usage mining Association rule mining is applied in web usage mining to understand patterns and associations in user browsing behavior. It helps website owners optimize website layouts, recommend relevant content, and personalize user experiences based on the navigation patterns and preferences of website visitors.
4.4.8 Healthcare and medical analysis Association rule mining is used in healthcare to analyze patient records, medical datasets, and clinical studies. It helps identify relationships between symptoms, diagnoses, treatments, and patient outcomes. These insights assist in disease diagnosis, treatment planning, drug discovery, and healthcare resource optimization.
4.4.9 Fraud detection Association rule mining is employed in fraud detection systems to identify patterns of fraudulent behavior. By analyzing transactional data and identifying associations between specific transactions or behaviors, it helps detect anomalies, suspicious activities, and potential fraud cases across various industries such as finance, insurance, and telecommunications.
98
Ashwin M. et al.
4.4.10 Recommender systems Association rules are utilized in recommender systems to generate personalized recommendations for users. By identifying associations between users’ preferences and past behaviors, it helps suggest relevant products, movies, music, or content, enhancing user satisfaction and engagement.
4.4.11 Supply chain management Association rule mining aids in optimizing supply chain management by identifying relationships and associations between products, suppliers, and customer demand. It helps in inventory management, demand forecasting, supplier selection, and supply chain optimization.
4.4.12 Bioinformatics and genomics Association rule mining is applied in bioinformatics to analyze genetic data, protein sequences, and biological networks. It helps identify associations between genetic markers, disease susceptibility, gene expression patterns, and drug responses, leading to advancements in personalized medicine and genomics research. These are just a few examples of the numerous applications of association rule mining. The technique is versatile and can be applied to various domains where discovering associations and patterns in large datasets is valuable for decision-making, optimization, and understanding complex systems.
4.4.13 Evaluation metrics When assessing the utility and impact of association rules in association rule mining, several measures are commonly used [16, 17]. Here are three widely utilized measures: Support: Support quantifies how frequently or frequently an itemset appears in the dataset. Support is calculated as the percentage of transactions in the dataset that contain both A and B for an association rule A → B. It is calculated as given below and aids in determining how frequently the rule occurs: Support ðA ! BÞ = ðNumber of transactions containing A and BÞ = ðTotal number of transactionsÞ
4 Clustering and association algorithm
99
These additional measures provide different perspectives on the value of the association rules, capturing various aspects such as implication strength, deviation from independence, similarity, and asymmetry between the antecedent and consequent. Depending on the specific requirements and objectives of the analysis, these measures can be utilized to evaluate and select the most interesting and relevant association rules.
4.5 Advanced clustering techniques Beyond the conventional approaches like K-means, hierarchical, and density-based clustering, advanced clustering techniques relate to more complex and specialized algorithms [18]. Here are a few instances of sophisticated clustering methods:
4.5.1 Density-based spatial clustering of applications with noise (DBSCAN) DBSCAN groups data points based on their density. It identifies dense regions separated by sparser areas and can handle arbitrary-shaped clusters while detecting outliers as noise.
4.5.2 Mean-shift clustering Mean shift is a nonparametric algorithm that identifies cluster centers by iteratively shifting the data points toward regions of maximum density. It is particularly useful for clustering in image segmentation and object tracking.
4.5.3 Spectral clustering Graph theory and a similarity matrix’s eigenvectors are used in spectral clustering to produce grouping. It clusters data points based on their connection patterns and treats them as nodes in a graph. Nonlinearly separable data can be handled through spectral clustering, which is especially useful for handling data with complicated structures.
100
Ashwin M. et al.
4.5.4 Affinity propagation Affinity propagation determines clusters by passing messages between data points to find exemplars that best represent the clusters. It can discover the number of clusters automatically and is suitable for applications like gene expression analysis and document clustering.
4.5.5 Gaussian mixture models (GMM) GMM assumes that the data points are generated from a mixture of Gaussian distributions. It models clusters as Gaussian distributions with different means and covariances. GMM can estimate the likelihood that a data point will belong to each cluster and is effective for data with overlapping clusters.
4.5.6 Self-organizing maps (SOM) It forms a grid of neurons that organize themselves to represent different clusters. SOM is useful for visualization and clustering of high-dimensional data.
4.5.7 BIRCH (balanced iterative reducing and clustering using hierarchies) BIRCH is an algorithm that performs hierarchical clustering while addressing scalability issues. It constructs a tree-like structure to represent clusters, allowing for efficient clustering of large datasets.
4.5.8 Applications and use cases Advanced clustering techniques have diverse applications across various domains. Here are some key use cases: 1. Image segmentation: Spectral clustering and Mean shift are used for segmenting images into meaningful regions, enabling object recognition and computer vision tasks. 2. Anomaly detection: Advanced clustering methods such as DBSCAN and GMM help detect anomalies in data, allowing for fraud detection, network intrusion detection, and outlier identification. 3. Customer segmentation: Advanced clustering algorithms aid in segmenting customers based on behavior, preferences, and demographics, enabling targeted marketing, personalized recommendations, and customer retention strategies.
4 Clustering and association algorithm
4. 5.
6.
7.
101
Bioinformatics: Advanced clustering techniques are applied in gene expression analysis, protein structure clustering, and identifying genetic patterns associated with diseases. Social network analysis: Clustering algorithms help identify communities, influential users, and patterns of interactions in social networks, supporting recommendation systems and targeted advertising. Document clustering: Advanced clustering techniques are used in text mining and natural language processing to cluster documents, enabling topic modeling, sentiment analysis, and information retrieval. Recommender systems: Advanced clustering algorithms contribute to collaborative filtering-based recommender systems, providing personalized recommendations for products, movies, music, and more.
These applications demonstrate the versatility of advanced clustering techniques in extracting insights, discovering patterns, and facilitating decision-making across domains such as healthcare, finance, marketing, and data analysis.
4.6 Time series clustering According to their patterns, trends, or similarities, similar time series data are grouped together into clusters through the process of time series clustering. It involves applying clustering algorithms specifically designed to handle time-dependent data [19, 20]. Here are some key aspects and techniques related to time series clustering: Distance measures: Time series clustering often relies on distance measures that capture the similarity or dissimilarity between two time series. Examples include Euclidean distance, dynamic time warping (DTW), and correlation-based measures. Shape-based clustering: Shape-based clustering methods focus on identifying similar patterns or shapes in time series data. This approach involves aligning and comparing the shapes of time series, using techniques like DTW-based clustering or shapeletbased clustering. Feature-based clustering: Feature-based clustering involves extracting meaningful features or representations from time series and using them as input to traditional clustering algorithms. Features can include statistical measures, frequency domain attributes, wavelet coefficients, or autoregressive model parameters. Hybrid approaches: Hybrid approaches combine different clustering techniques or incorporate additional domain-specific knowledge to improve clustering results. For example, a combination of shape-based and feature-based clustering methods can be employed to capture both global patterns and local variations in time series.
102
Ashwin M. et al.
4.6.1 Applications and use cases Applications of time series clustering include: Financial analysis: Clustering stock price or financial market data to identify groups of similar time series for portfolio management or risk analysis Energy consumption: Clustering energy consumption patterns to detect anomalies, identify energy usage profiles, and optimize energy management strategies Healthcare: Clustering physiological signals or patient monitoring data to identify patterns for disease diagnosis, patient segmentation, or anomaly detection Sensor networks: Clustering sensor data to identify common patterns or behaviors for environmental monitoring, smart grids, or Internet of Things (IoT) applications Time series clustering helps in discovering meaningful patterns and structures within time-dependent data, enabling better understanding, prediction, and decision-making in various domains.
4.7 Ensemble clustering Ensemble clustering, also known as cluster ensemble or consensus clustering, is a technique that combines multiple clustering results to produce a single robust and improved clustering solution [21]. To provide more precise and reliable clustering results, it makes use of the diversity and complementary information present in various clustering algorithms or parameter settings. Here are some key aspects and techniques related to ensemble clustering: Diversity generation: Ensemble clustering involves generating diverse clustering solutions by applying different clustering algorithms, using various parameter settings, or employing different initialization methods. The goal is to obtain a set of diverse base clusterings. Ensemble clustering combines the basis clusterings into a single consensus clustering by using integration methods. Majority voting, weighted voting, clustering co-occurrence, and clustering agreement measures are common integration strategies. Consensus function: A consensus function determines the similarity or agreement between pairs of data points based on their clustering assignments across the base clusterings. It helps build a consensus similarity matrix that represents the overall clustering consensus.
4 Clustering and association algorithm
103
Cluster stability analysis: Ensemble clustering often involves assessing the stability of the clustering solutions to ensure the reliability of the ensemble result. Stability measures such as cluster-wise stability, pairwise similarity, or stability indexes are used to evaluate the stability of clusters across different ensemble members. Ensemble pruning and selection: Ensemble clustering may involve pruning or selecting a subset of base clusterings based on their quality or diversity measures to improve the final clustering result and reduce computational complexity.
4.7.1 Evaluation of ensemble clustering The evaluation of ensemble clustering aims to assess the quality and effectiveness of the clustering results obtained from ensemble techniques. Here are some common evaluation measures used for ensemble clustering:
4.7.2 Similarity measures – –
Jaccard coefficient: Measures the similarity between the clustering result and a reference clustering solution Fowlkes-Mallows index: Calculates the geometric mean of the pairwise precision and recall between the clustering result and a reference clustering
4.7.3 Stability measures – –
Cluster-wise stability: Assesses the stability of clusters by measuring the agreement among different clustering runs or ensemble members Pairwise similarity: Evaluates the stability of pairwise similarity between data points across different ensemble members
4.7.4 Consensus measures – –
C-index: Measures the consensus or agreement among the base clusterings by evaluating the clustering co-occurrence across ensemble members Normalized Mutual Information (NMI): Quantifies the agreement between the clustering result and a reference clustering by comparing their mutual information
104
Ashwin M. et al.
4.7.5 External evaluation measures – –
Rand index: Compares the similarity of the clustering result with a reference clustering by considering true positive and true negative pairwise agreements F-measure: Combines precision and recall to evaluate the clustering result against a reference clustering
4.7.6 Internal evaluation measures –
Silhouette score: Based on the average intra-cluster distance and inter-cluster distance, evaluates the separation and compactness of clusters
4.7.7 Applications and use cases Applications of ensemble clustering include: Image and object recognition: Ensemble clustering is employed in image and object recognition tasks to combine multiple feature extraction and clustering algorithms for improved accuracy and robustness. Bioinformatics: Ensemble clustering is used in genomic data analysis to identify gene expression patterns, discover co-expressed gene modules, and classify gene expression profiles. Text mining: Ensemble clustering helps in text document clustering to group similar documents for information retrieval, topic modeling, and document organization. Social network analysis: Ensemble clustering is applied to analyze social networks and identify community structures, detect overlapping communities, and analyze network evolution. Ensemble clustering techniques aim to enhance clustering results by aggregating multiple clustering solutions, reducing the sensitivity to algorithmic choices and initialization, and improving the robustness and stability of the final clustering output.
4.8 Clustering and association algorithms in large datasets Clustering and association algorithms face unique challenges when dealing with large datasets. Here are some considerations and techniques for handling clustering and association tasks in large datasets [22]:
4 Clustering and association algorithm
105
Sampling and subsampling: Large datasets can be computationally expensive to process. Sampling techniques, such as random sampling or stratified sampling, can be employed to create smaller representative subsets of the data for analysis. Subsampling techniques, such as mini-batch clustering, divide the data into smaller batches and process them iteratively. Distributed computing: The calculation can be split over several nodes or machines using distributed computing frameworks like Apache Hadoop or Apache Spark. Due to this, processing may be done in parallel and computing resources can be used effectively. Incremental and online algorithms: Instead of processing the entire dataset at once, incremental and online algorithms process the data in small chunks or as new data arrives. This approach enables real-time analysis and reduces memory requirements. Dimensionality reduction: Clustering and association algorithms can face difficulties when dealing with high-dimensional datasets. PCA and t-SNE are dimensionality reduction techniques that can be used to minimize the number of features and streamline the data representation while retaining the crucial information. Scalable algorithms: Specific clustering and association algorithms are designed to handle large datasets efficiently. For example, k-means can utilize approximate algorithms like K-means++, and scalable versions of association rule mining algorithms, like FP-Growth, can be employed. Parallel processing: Exploiting parallel processing techniques, such as parallelization of distance calculations or parallelizing different stages of the algorithms, can significantly speed up the computation process. Preprocessing and data cleaning: Large datasets often contain noisy or irrelevant data. Proper data preprocessing, including data cleaning, noise removal, and outlier detection, can improve the quality and efficiency of clustering and association algorithms. Incremental update and maintenance: In scenarios where the dataset is continuously updated, algorithms should be capable of incremental updates and maintenance to accommodate new data without reprocessing the entire dataset. Parallel and distributed processing: To transfer the computational load across numerous nodes or machines, use parallel and distributed computing frameworks like Apache Spark or Apache Flink. Large dataset processing can be accelerated greatly as a result. Indexing and approximation: Employ indexing techniques like k-d trees or R-trees to speed up nearest neighbor searches or distance calculations in clustering algorithms. Approximation algorithms can also be used to trade off accuracy for faster computation in large-scale scenarios.
106
Ashwin M. et al.
Stream processing: If dealing with streaming data, employ stream processing techniques that allow for real-time clustering or association analysis. Algorithms like CluStream or DStream can handle continuous data streams and provide up-to-date insights. Parallel feature extraction: If dimensionality reduction techniques are applied, consider parallelizing the feature extraction process to accelerate the computation. This can be achieved through techniques like map-reduce or parallelized matrix operations. Sampling techniques: In addition to random sampling, advanced sampling techniques like stratified sampling, reservoir sampling, or importance sampling can be employed to maintain data representativeness while reducing the computational burden. Approximate and streaming association mining: Utilize approximation algorithms, such as Count-Min Sketch or Bloom Filters, to perform approximate association mining in large datasets. Streaming association mining algorithms like StreamingFP or Space-Saving can handle high-speed data streams. Scalable data storage: Efficiently store and retrieve large datasets using scalable data storage systems like Apache Hadoop Distributed File System (HDFS), NoSQL databases, or columnar databases. This enables seamless access to data during the clustering and association process. Dimensionality adaptation: Think about strategies for adaptive dimensionality, which dynamically change the number of dimensions based on the properties of the data. This increases the effectiveness of grouping and association algorithms and lessens the effects of the curse of dimensionality. Applying these techniques and considering the specific characteristics of large datasets can help address the computational, scalability, and efficiency challenges associated with clustering and association algorithms, enabling their effective application in real-world scenarios.
4.9 Conclusion and future directions Clustering and association algorithms play a crucial role in uncovering patterns, relationships, and insights from data. With the advancement of technology and the proliferation of large and complex datasets, the field of clustering and association mining continues to evolve. In summary, clustering methods such as K-means, hierarchical clustering, and density-based clustering allow the identification of groups or clusters within data, whereas association rule learning, such as the Apriori algorithm, identifies significant links between objects or variables. These techniques have found applications in various domains, including finance, healthcare, marketing, and social network analysis.
4 Clustering and association algorithm
107
Moving forward, future directions in clustering and association mining include: Scalability: Developing algorithms and techniques that can efficiently handle ever increasing volumes of data, including big data and streaming data Parallel and Distributed processing: Enhancing parallel and distributed computing frameworks to handle larger datasets and leverage the power of distributed systems Handling high-dimensional data: Addressing the challenges posed by high-dimensional datasets through advanced dimensionality reduction techniques and specialized algorithms Incorporating domain knowledge: Integrating domain-specific knowledge and constraints into clustering and association algorithms to improve their accuracy and interpretability Advanced evaluation metrics: Designing novel evaluation metrics that capture the quality, stability, and interpretability of clustering and association results, considering specific application requirements Interpretable and Explainable results: Focusing on developing techniques that provide clear explanations and interpretations of clustering and association patterns to enhance trust and usability Integration with machine learning: Exploring the integration of clustering and association techniques with other machine learning methods, such as deep learning or reinforcement learning, to enable more comprehensive data analysis Privacy and security: Addressing privacy and security concerns in clustering and association mining, ensuring the protection of sensitive information and complying with regulations The future of clustering and association mining lies in advancing scalability, efficiency, interpretability, and integration with other techniques. With ongoing research and innovation, these methods will continue to provide valuable insights and support decision-making across a wide range of applications and industries.
References [1] [2] [3]
Ahmad, A., & Khan, S. S. (2019). Survey of state-of-the-art mixed data clustering algorithms. IEEE Access, 7, 31883–31902. Gurudath, S. (2020). Market basket analysis & recommendation system using association rules. Master of Science in Big Data Management and Analytics, Griffith College, Dublin. Zhang, T., & Mao, S. (2020). Machine learning for end-to-end congestion control. IEEE Communications Magazine, 58(6), 52–57.
108
[4]
[5] [6] [7] [8]
[9]
[10]
[11]
[12] [13] [14]
[15] [16]
[17] [18] [19] [20]
[21] [22]
Ashwin M. et al.
Babenko, V., Panchyshyn, A., Zomchak, L., Nehrey, M., Artym-Drohomyretska, Z., & Lahotskyi, T. (2021). Classical machine learning methods in economics research: Macro and micro level example. WSEAS Transactions on Business and Economics, 18, 209–217. Sinaga, K. P., & Yang, M. S. (2020). Unsupervised K-means clustering algorithm. IEEE Access, 8, 80716–80727. Govender, P., & Sivakumar, V. (2020). Application of k-means and hierarchical clustering techniques for analysis of air pollution: A review (1980–2019). Atmospheric Pollution Research, 11(1), 40–56. Chen, Y., Zhou, L., Bouguila, N., Wang, C., Chen, Y., & Du, J. (2021). BLOCK-DBSCAN: Fast clustering for large scale data. Pattern Recognition, 109, 107624. Silvia Priscila, S., Sathish Kumar, C., Manikandan, R., Yuvaraj, N., & Ramkumar, M. (2022, February). Interactive artificial neural network model for UX design. In International Conference on Computing, Communication, Electrical and Biomedical Systems (pp. 277–284). Cham: Springer International Publishing. Seem, A., Chauhan, A. K., Khan, R., & Yadav, S. P. (2020). Distributed artificial intelligence for document retrieval. In: Satya Prakash Yadav, Dharmendra Prasad Mahato & Nguyen Thi Dieu Linh (Eds.). Distributed Artificial Intelligence (pp. 59–68). Boca Raton: CRC Press. Zebari, R., Abdulazeez, A., Zeebaree, D., Zebari, D., & Saeed, J. (2020). A comprehensive review of dimensionality reduction techniques for feature selection and feature extraction. Journal of Applied Science and Technology Trends, 1(2), 56–70. Kiarashinejad, Y., Abdollahramezani, S., & Adibi, A. (2020). Deep learning approach based on dimensionality reduction for designing electromagnetic nanostructures. npj Computational Materials, 6(1), 12. Al-Turjman, F., Yadav, S. P., Kumar, M., Yadav, V., & Stephan, T. (Eds.). (2022). Transforming management with AI, Big-Data, and IoT. India: Springer. Nithya, C., & Saravanan, V. (2018). A study of machine learning techniques in data mining. International Scientific Refereed Research Journal, 1, 31–38. Praghash, K., Yuvaraj, N., Peter, G., Stonier, A. A., & Priya, R. D. (2023, May). Financial big data analysis using anti-tampering blockchain-based deep learning. In Hybrid Intelligent Systems: 22nd International Conference on Hybrid Intelligent Systems (HIS 2022), December 13–15, 2022 (pp. 1031–1040). Cham: Springer Nature Switzerland. Das, S., Kong, X., & Tsapakis, I. (2021). Hit and run crash analysis using association rules mining. Journal of Transportation Safety and Security, 13(2), 123–142. Vougas, K., Sakellaropoulos, T., Kotsinas, A., Foukas, G. R. P., Ntargaras, A., Koinis, F., & Gorgoulis, V. G. (2019). Machine learning and data mining frameworks for predicting drug response in cancer: An overview and a novel in silico screening process based on association rule mining. Pharmacology and Therapeutics, 203, 107395. Telikani, A., Gandomi, A. H., & Shahbahrami, A. (2020). A survey of evolutionary computation for association rule mining. Information Sciences, 524, 318–352. Alashwal, H., El Halaby, M., Crouse, J. J., Abdalla, A., & Moustafa, A. A. (2019). The application of unsupervised clustering methods to Alzheimer’s disease. Frontiers in Computational Neuroscience, 13, 31. Maharaj, E. A., D’Urso, P., & Caiado, J. (2019). Time series clustering and classification. New York: CRC Press. Bañales, S., Dormido, R., & Duro, N. (2021). Smart meters time series clustering for demand response applications in the context of high penetration of renewable energy resources. Energies, 14(12), 3458. Huang, D., Wang, C. D., Wu, J. S., Lai, J. H., & Kwoh, C. K. (2019). Ultra-scalable spectral clustering and ensemble clustering. IEEE Transactions on Knowledge and Data Engineering, 32(6), 1212–1226. Wu, W. T., Li, Y. J., Feng, A. Z., Li, L., Huang, T., Xu, A. D., & Lyu, J. (2021). Data mining in clinical big data: The frequently used databases, steps, and methodological models. Military Medical Research, 8, 1–12.
Nazeer Shaik, S. Chandra Sekaran, Amit Mahajan, Balkeshwar Singh
5 Reinforcement learning Abstract: The field of reinforcement learning (RL) is introduced in this chapter, which also looks at several RL techniques. The main goal of RL is to provide algorithms that let agents discover the best policies through interactions with their surroundings while maximizing cumulative rewards. In the first part of the chapter, Markov decision processes (MDPs), which provide a mathematical foundation for modeling RL problems, are discussed. We look at value iteration and policy iteration as iterative approaches to addressing MDPs. To help you find the ideal action-value function, we present Q-Learning, an off-policy model-free RL algorithm. Deep Q-networks (DQNs), which combine Q-learning with deep neural networks, are also addressed in order to handle high-dimensional state spaces. Policy gradient methods are presented as an alternative approach that directly optimizes policy parameters using gradient ascent. Proximal policy optimization (PPO), a leading policy gradient algorithm, is discussed for its ability to balance stability and policy performance. The chapter concludes by emphasizing the significance of RL methods in training agents to make sequential decisions in complex environments across various domains. Keywords: Scalability, communication, coordination, nonstationarity, learning dynamics, generalization
5.1 Introduction The core of the subject of reinforcement learning (RL), a subsection of machine learning, is the investigation of agents that interact with their environment and discover how-to-do behaviors that maximize a cumulative reward signal. This chapter will examine numerous RL techniques that let agents discover the best strategies through making mistakes. RL methods provide powerful tools for training agents to make sequential decisions in complex environments. From classical methods like value iteration and policy iteration to more recent advances like deep Q-networks and policy gradient algorithms, RL continues to evolve and find applications in various domains, including robotics, game playing, and autonomous systems [1, 2].
https://doi.org/10.1515/9783111323749-005
110
Nazeer Shaik, S. Chandra Sekaran, Amit Mahajan, Balkeshwar Singh
5.2 Markov decision processes (MDPs) Sequential decision-making issues in RL are modeled mathematically using Markov decision processes (MDPs). Different elements define the environment and its dynamics in MDPs. States (S): States represent the configurations or situations an agent can encounter during the decision-making process. The set S encompasses all possible states, and each state s ∈ S captures relevant information about the environment. Actions (A): Actions refer to the choices an agent can make in a given state. The set A represents the available actions, and an action a ∈ A represents a specific decision the agent can take. Transition probabilities (T): Transition probabilities depict the likelihood of transitioning from one state to another based on the chosen action. Denoted as T(s, a, s’), when the agent acts, it calculates the likelihood of the transition from state s to state s’. Rewards (R): The agent receives instant input from rewards, which influences how it learns. R(s, a, s’) denotes the reward received when moving from state s to state s’ following the completion of action a. Discount factor (γ): The discount factor, denoted by γ (0 ≤ γ ≤ 1), determines whether future rewards are relevant in relation to immediate earnings. With a number closer to 1, it gives future rewards greater weight while balancing short-term and long-term incentives.
5.3 Value iteration By calculating the ideal state-value function, value iteration is an iterative approach used to solve MDPs. The algorithm updates the value estimates for each state iteratively until convergence [3]. Here is a detailed explanation of the value iteration algorithm: Initialization: Set a random beginning value for all states for the value function V(s) from s ∈ S.
5 Reinforcement learning
111
Iterative Update: Follow these steps repeatedly until convergence: First, for each state s∈ S: Compute the action-value function for each action a ∈ A: Qðs, aÞ =
X
s′ Tðs, a, s′Þ ½Rðs, a, s′Þ + γVðs′Þ
The value function can be updated by selecting the action that maximizes the actionvalue function: Convergence Check: VðsÞ = max½Qðs, aÞ for all a 2 A In order to attain the maximum number of iterations, keep repeating step 2 until the change in the value estimations becomes minimal. The value iteration algorithm utilizes the Bellman equation, which states that the value of a state is equal to the maximum expected return achievable from that state: V✶ ðsÞ = max½Q✶ ðs, aÞ for all a 2 A Here, the ideal state-value function is represented by V✶(s), while the ideal actionvalue function is represented by Q✶(s, a). The algorithm modifies the value function after each iteration using the most recent estimates of the action-value function. When calculating the action-value function, the transition probabilities T(s, a, s’) and immediate rewards R(s, a, s’) linked to the state-action pairs are taken into consideration. The discount factor is used to rank future benefits [4].
5.4 Policy iteration By repeatedly completing the stages of policy evaluation and policy improvement, the iterative technique known as “policy iteration” is used to solve MDPsMDPs. By revising the policy until convergence, it seeks to identify the best course of action [5]. Here is a thorough description of policy iteration: Initialization: Initialize the policy π arbitrarily for all states s ∈ S. Policy evaluation: Conduct a policy evaluation to calculate the value function Vπ(s) for the present policy π. Continue iterating until the value estimations converge:
112
Nazeer Shaik, S. Chandra Sekaran, Amit Mahajan, Balkeshwar Singh
a. For each state s ∈ S: Calculate the state’s value using the policy evaluation equation: X X VπðsÞ = a πðajsÞ s′ Tðs, a, s′Þ ½Rðs, a, s′Þ + γVπðs′Þ Policy improvement: Based on the current value function estimates Vπ(s), update the policy π by selecting the action that maximizes the expected return: a. For each state s ∈ S: Compute the action-value function for each action a ∈ A: X Qπðs, aÞ = s′ Tðs, a, s′Þ ½Rðs, a, s′Þ + γVπðs′Þ Update the policy by selecting the action that maximizes the action-value function: πðsÞ = argmax½Qπðs, aÞ for all a 2 A Convergence check: Until the policy changes or the allotted number of iterations is reached, repeat steps 2 and 3 as necessary. During policy iteration, the value function V(s) is estimated by taking into account the transition probabilities T(s, a, s’) and immediate rewards R(s, a, s’) connected to the state-action pairings. The significance of future awards is determined by the discount factor [6]. Policy iteration enables us to locate the best policy that maximizes the predicted cumulative rewards in the MDP by repeatedly undertaking policy evaluation and policy improvement [7–9].
5.5 Deep reinforcement learning Playing video games, operating robots, and optimizing complex systems are just a few of the hard challenges that deep reinforcement learning (DRL) has been able to solve with amazing effectiveness [10]. Here are some key components and techniques used in DRL: Deep neural networks (DNNs): DRL employs DNNs as function approximators to approximate the value function or policy. DNNs can learn hierarchical representations from raw sensory inputs, enabling efficient and expressive representation of complex state-action mappings [11].
5 Reinforcement learning
113
Model-based methods: There are model-based approaches that develop models of the dynamics of the environment, even though the majority of DRL algorithms are model-free. These models can be used for planning, data augmentation, or improving sample efficiency in learning [12]. Asynchronous advantage actor-critic (A3C): The distributed actor-critic algorithm is known as A3C. It employs multiple agents that interact with their own copies of the environment asynchronously. The agents update a shared neural network asynchronously, allowing for efficient exploration and faster convergence. Proximal value optimization (PPO): PPO is an approach for on-policy optimization that updates the policy with a substitute objective function. It constrains the policy update to a fixed region around the previous policy to ensure stability and prevents large policy changes. Rainbow: Rainbow is a combination of various DRL techniques to achieve state-of-the-art performance in value-based RL. It incorporates enhancements such as prioritized experience replay, distributional Q-learning, n-step returns, dueling networks, and NoisyNet exploration to improve sample efficiency and stability. DRL from human demonstrations (DRLfD): DRLfD combines imitation learning with RL. It leverages expert demonstrations to provide initial guidance to the RL agent, allowing it to learn more efficiently from limited interactions with the environment. DRL has demonstrated significant advancements in various domains and has the potential to address complex real-world problems. However, training deep RL agents often requires substantial computational resources, careful hyperparameter tuning, and addressing challenges such as sample efficiency, exploration-exploitation tradeoffs, and generalization to new environments. Ongoing research aims to tackle these challenges and further advance the capabilities of DRL algorithms.
5.6 Exploration strategies The agent is guided by exploration methods to investigate the world and find new actions and states, which is a critical part of RL. They try to achieve a balance between exploitation (making choices based on the information of the moment) and exploration (trying out new things) [13]. This section provides a process flow to elaborate on exploration strategies:
114
Nazeer Shaik, S. Chandra Sekaran, Amit Mahajan, Balkeshwar Singh
Initialization: Start by initializing the agent’s policy and value estimates based on prior knowledge or random initialization. Select action: When faced with a state, the agent needs to select an action. This decision-making process is guided by the exploration strategy [14]. Exploration strategy selection: Depending on the particular needs and characteristics of the situation, select an exploratory strategy. Some typical exploratory methods include: a. Epsilon-greedy: The epsilon-greedy approach is a straightforward and popular method. The majority of the time, it includes choosing the action with the highest estimated value (exploitation), but with a slim chance of doing so (epsilon) it entails choosing a random action (exploration). b. Boltzmann exploration: Boltzmann exploration, also known as softmax exploration, selects actions probabilistically based on their estimated values. A stochastic exploration technique is possible since the likelihood of choosing an action is inversely correlated with its exponential magnitude. c. Upper confidence bound (UCB): UCB is an exploration strategy that balances exploration and exploitation by assigning a bonus to actions based on their exploration potential. The bonus term considers both the estimated value of the action and the uncertainty associated with it. d. Thompson sampling: A distribution over the value estimates for each action is maintained using the probability-based exploration technique known as sampling. By taking a value from the distribution for each action and sampling it, it selects the action with the greatest sampled value. e. Noise injection: In some cases, random noise can be injected into the action selection process to encourage exploration. This can involve adding Gaussian noise to the action values or applying random perturbations to the policy parameters. Action selection: Based on the chosen exploration strategy, select an action to take in the current state. This action will determine the agent’s interaction with the environment.
5 Reinforcement learning
115
Environment interaction: Execute the chosen action in the environment, then see the state transition and reward that result. Update policy and value estimates: Update the agent’s policy and value estimations in accordance with the new state and reward after doing so using the RL method of your choice (e.g., Q-learning, and policy gradient). Usually, the agent’s learning algorithm and the observed transition data are used to inform this update. Repeat steps 2–6: Updating the agent’s policy and value estimations while continuing to choose actions must be done repeatedly until the learning process converges or a predetermined stopping criterion is satisfied. For the agent to investigate and learn from the environment, exploration tactics are essential. By intelligently balancing exploration and exploitation, these strategies enable the agent to discover optimal actions and states, leading to effective decisionmaking and learning in RL tasks.
5.7 Continuous action spaces In RL, continuous action spaces refer to scenarios where the agent can select actions from a continuous range of possible values. Unlike discrete action spaces, which have a finite number of options, continuous action spaces require different approaches to handle the infinite set of possible actions [14]. This section provides an elaboration on continuous action spaces: Action representation: In continuous action spaces, the actions are typically represented as continuous variables. For simplicity, let us consider a single-dimensional continuous action space. Policy representation: Parameterized functions, such as neural networks, are frequently used to map states to action values while learning a policy in continuous action spaces. The parameters that define the probability distribution across the continuous action space are produced by the policy function π(s; θ) from the state s as an input [15]. Actor-critic methods: The parameters of both networks are changed by employing methods like the deterministic policy gradient or advantage-based techniques [16, 17].
116
Nazeer Shaik, S. Chandra Sekaran, Amit Mahajan, Balkeshwar Singh
Exploration strategies: In continuous action spaces, exploration is particularly challenging due to the infinite set of possible actions. Exploration strategies like adding noise to the policy parameters (e.g., Ornstein-Uhlenbeck process) or using action space exploration techniques (e.g., parameterizing the policy with a random variable) can be employed to encourage exploration and prevent the agent from being stuck in local optima. These are some of the common approaches used to handle continuous action spaces in RL. By using techniques such as deterministic policy gradients, function approximation, and actor-critic architectures, agents can effectively learn policies in continuous action spaces.
5.8 Model-based reinforcement learning The model-based reinforcement learning (MBRL), which combines model learning with RL, is a strategy for solving difficult issues. For planning and decision-making in MBRL, the agent learns a model of the dynamics of the environment [18–25]. MBRL is explained in more detail in this section: Model learning: In MBRL, the agent learns a model of the environment dynamics, which captures how the states and rewards evolve over time. Model-based planning: Once the model is learned, the agent can use it for planning. Planning involves using the model to simulate future trajectories and optimize the expected cumulative reward. Action sequences that optimize the anticipated return can be created using planning techniques like Monte Carlo tree search or model predictive control. Value function and policy learning: The agent in MBRL can change its value function or policy using the learnt model. The value function can be updated using techniques like value iteration or Q-learning; the policy can be directly optimized using policy-based approaches employing gradientbased algorithms. Model-based control: Model-based control is the method used in MBRL to choose actions in the real world based on the learnt model. The agent runs simulations of potential action sequences using the model, analyzing the expected returns to make judgments.
5 Reinforcement learning
117
Model-based exploration: In MBRL, the learned model can also be used for exploration. By simulating possible actions and their outcomes using the model, the agent can explore regions of the state-action space that it has not yet encountered. Various exploration strategies discussed earlier can be employed in conjunction with model-based methods to encourage exploration. MBRL uses a variety of equations and algorithms, depending on the particular method and problem area. MBRL has benefits including sample effectiveness and the capacity for planning and environmental reasoning. However, it also requires accurate model learning and planning algorithms that can handle the complexities of the environment dynamics. For effective and efficient RL, current research is focused on enhancing model learning methods, creating reliable planning algorithms, and combining model-based and model-free approaches.
5.9 Multiagent reinforcement learning The field of RL known as multiagent reinforcement learning” (MARL) examines interactions between multiple agents in a shared environment where they might learn to make cooperative or hostile decisions. MARL extends the single-agent setting to handle the challenges arising from the presence of other intelligent agents. In MARL, agents can learn to coordinate, communicate, compete, or cooperate to achieve their individual or collective objectives. There are several types of MARL, each with its own characteristics and objectives. Let us delve into some of the main types of MARL: Cooperative MARL: Cooperative MARL focuses on situations where agents aim to maximize a shared global reward. The agents work together, sharing information and coordinating their actions to achieve a common goal. Since one agent’s activities affect the outcomes and rewards of other agents, learning effective coordination procedures can be challenging. The agents may employ centralized training with decentralized execution, where a central learner aids in guiding the agents during training but they act independently during execution. Competitive MARL: In competitive MARL, agents are pitted against each other, and their objectives are adversarial or conflicting. Each agent seeks to maximize its own reward or performance metric, often at the expense of others. Examples include competitive games like chess or multiplayer video games. The agents need to learn strategies to outsmart their opponents, anticipating their actions and finding optimal decision-making policies in the face of competition.
118
Nazeer Shaik, S. Chandra Sekaran, Amit Mahajan, Balkeshwar Singh
Mixed cooperative-competitive MARL: In mixed cooperative-competitive MARL, agents simultaneously engage in both cooperative and competitive interactions. They may form teams to cooperate towards a common goal while also competing with each other for individual rewards. This type of MARL is prevalent in scenarios with complex dynamics where both cooperation and competition are necessary for success. Agents must learn to balance cooperation and competition strategically to optimize their overall performance. Communication-based MARL: Communication-based MARL involves agents that can exchange information or messages with each other during their interactions. Communication enables agents to share knowledge, coordinate actions, and cooperate more effectively. It allows for higher levels of coordination and can help overcome challenges posed by complex environments. Agents can learn to communicate through explicit message passing or through learned communication protocols. Hierarchical MARL: Hierarchical MARL involves learning policies at different levels of abstraction. Agents learn both low-level policies that directly map states to actions and high-level policies that determine goals or subtasks. High-level policies direct the agent’s decisionmaking, and low-level policies manage the specifics of action selection. Hierarchical approaches can simplify learning and decision-making in complex environments, enabling agents to tackle larger problems effectively. Decentralized MARL: Decentralized MARL uses local observations to allow each agent to independently learn its own policy. The global state and comprehensive details of other agents’ policies are not available to agents. They make decisions based on their limited local observations, often resulting in partial observability. Decentralized MARL requires agents to learn effective individual policies while considering the interactions with other agents in the environment. Centralized MARL: In contrast to decentralized MARL, centralized MARL involves agents with access to a centralized controller that provides global information about the environment and other agents. The central controller can gather and process information from all agents and provide guidance to optimize the overall performance. This type of MARL can enhance coordination and decision-making, but it also raises challenges related to communication overhead and scalability. These types of MARL represent different problem settings and objectives in multiagent environments. Researchers employ a variety of algorithms and techniques to address the complexities of interactions among agents, including value-based meth-
5 Reinforcement learning
119
ods, policy gradients, opponent modelling, game theory, and more. The choice of MARL type and algorithm depends on the specific problem, the nature of interactions.
5.10 Applications and use cases MARL has a wide range of applications across various domains. Here are some notable applications and use cases where MARL techniques have shown promise: Robotics: MARL is extensively used in robotics for multirobot coordination and collaboration. It enables teams of robots to work together on tasks such as search and rescue, swarm robotics, cooperative transportation, and multirobot exploration. MARL allows robots to communicate, share information, and coordinate their actions to achieve common objectives efficiently. Autonomous vehicles: MARL is valuable for developing intelligent systems for autonomous vehicles. Agents representing vehicles can learn to navigate complex traffic scenarios, optimize traffic flow, and handle coordination issues like merging, lane changing, and intersection management. MARL can improve traffic efficiency, reduce congestion, and enhance overall road safety. Multiagent games: MARL is extensively applied to competitive and cooperative games, both in video games and real-world sports. It enables game agents to learn optimal strategies, adaptive behavior, and decision-making in dynamic and uncertain environments. MARL has been used in games like poker, chess, Dota 2, and autonomous drone racing, achieving impressive performance against human players. Social networks: MARL is employed in modeling and analyzing social networks to understand the dynamics of social interactions. It can help in predicting user behavior and influence maximization, recommendation systems, and online advertising. MARL enables agents to learn and adapt to the evolving preferences and behaviors of individuals in the network. Energy management: MARL is beneficial for optimizing energy consumption and resource allocation in smart grids and energy systems. Multiple agents, such as power generators, consumers, and storage units, can learn to coordinate and balance energy supply and demand efficiently. MARL can help reduce energy costs, improve renewable energy integration, and enhance the overall stability and reliability of the grid.
120
Nazeer Shaik, S. Chandra Sekaran, Amit Mahajan, Balkeshwar Singh
Multi-agent healthcare systems: MARL finds applications in healthcare systems where multiple agents, such as medical devices, sensors, and health care providers, interact to provide personalized and efficient patient care. MARL can optimize treatment plans, resource allocation, and patient scheduling while considering constraints and individual patient needs. Supply chain management: MARL is employed in optimizing supply chain operations involving multiple entities, such as suppliers, manufacturers, distributors, and retailers. Agents can learn to collaborate, coordinate inventory management, and make decisions to minimize costs, reduce lead times, and improve overall supply chain performance. Multirobot simulations: MARL is used to simulate multirobot systems and test different coordination and control strategies. Simulations help in evaluating system performance, scalability, and robustness before deploying the algorithms on real robots. MARL allows researchers to investigate various scenarios, such as disaster response, environmental monitoring, and industrial automation. These are just a few examples of the broad range of applications where MARL techniques are being explored and deployed. MARL offers the potential to address complex problems that involve interactions among multiple intelligent agents, leading to improved efficiency, coordination, and decision-making in various domains.
5.11 Challenges and future directions There are several challenges that researchers are actively addressing to further advance the field. Here are some of the key challenges and potential future directions in MARL: Scalability: Traditional MARL algorithms have trouble handling large numbers of agents since the state-action space is expanding exponentially and interactions are becoming more complicated. The creation of scalable algorithms with effective learning and coordination capabilities in contexts with numerous agents is a future direction. Communication and coordination: Effective communication and coordination among agents are crucial for successful MARL. Designing communication protocols, ensuring information sharing, and learning to coordinate actions in a decentralized manner are ongoing research areas. Future directions involve developing more sophisticated communication mechanisms,
5 Reinforcement learning
121
including learning communication strategies from data and incorporating natural language processing for human-agent collaboration. Nonstationarity and learning dynamics: In environments, where the dynamics of the agents or the environment change over time, learning in MARL becomes challenging. Agents need to adapt to nonstationarity and learn in real-time. Future directions involve developing algorithms that can handle nonstationarity, model changes, and concept drift. Techniques like online learning, meta-learning, and continual learning can be explored to improve the adaptability of MARL algorithms. Exploration and exploitation: Balancing exploration and exploitation is critical in MARL. Agents need to explore the environment to discover new strategies while leveraging existing knowledge to exploit learned policies. Future directions involve developing advanced exploration strategies, including curiosity-driven exploration, meta-learning-based exploration, and adaptive exploration techniques that can effectively explore the vast state-action space in multiagent environments. Robustness and adversarial environments: MARL in adversarial environments, where agents actively compete against each other, raises challenges of robustness and learning in the presence of adversaries. Agents need to learn to anticipate opponent strategies, handle deception, and adapt to adversarial behavior. Future directions involve exploring techniques from game theory, opponent modeling, and adversarial training to enhance the robustness of MARL algorithms. Transfer and generalization: MARL algorithms often struggle with transferring learned knowledge from one environment to another or generalizing across different tasks. Future directions involve developing transfer learning techniques that enable agents to leverage knowledge learned in one setting and apply it to similar new settings. Generalization techniques such as meta-learning, hierarchical RL, and domain adaptation can also enhance the ability of agents to generalize across different scenarios. Ethical and fair decision-making: MARL raises ethical considerations when agents interact with humans or make decisions that impact society. Ensuring fairness, avoiding biases, and addressing issues of value misalignment are important challenges. Future directions involve incorporating ethical principles into MARL algorithms, designing reward structures that promote fairness, and developing techniques for value-aligned and interpretable decisionmaking. Real-world applications and safety: Deploying MARL in real-world applications, such as autonomous systems or healthcare, requires addressing safety and reliability concerns. Future directions involve de-
122
Nazeer Shaik, S. Chandra Sekaran, Amit Mahajan, Balkeshwar Singh
veloping methods for safe exploration, robustness to uncertain environments, and ensuring the compliance of learned policies with safety constraints and regulations. These challenges and future directions highlight the ongoing research efforts in MARL to overcome limitations and improve the applicability and effectiveness of multiagent systems. By addressing these challenges, MARL has the potential to revolutionize various domains, enabling intelligent collaboration, decision-making, and coordination among multiple agents in complex and dynamic environments.
References [1]
Foerster, J., Assael, Y. M., de Freitas, N., & Whiteson, S. (2016). Learning to communicate with deep multi-agent reinforcement learning. Advances in Neural Information Processing Systems, 26, 2137–2145. [2] Lowe, R., Wu, Y., Tamar, A., Harb, J., Abbeel, P., & Mordatch, I. (2017). Multi-agent actor-critic for mixed cooperative-competitive environments. Advances in Neural Information Processing Systems, 30, 6379–6390. [3] Sunehag, P., Lever, G., Gruslys, A., Vinyals, O., Spiessens, A., Mohanty, S., & Whiteson, S. (2018). Value-decomposition networks for cooperative multi-agent learning. Proceedings of the AAAI Conference on Artificial Intelligence, 32(1), 1–14. [4] Rashid, T., Samvelyan, M., & Nowé, A. (2018). Qmix: Monotonic value function factorisation for deep multi-agent reinforcement learning. In Thirty-Second AAAI Conference on Artificial Intelligence. [5] Yu, H., Gupta, A., Xu, Z., Levine, S., & Abbeel, P. (2019). MOCA: Model-based offline contextual bandit for continuous actions. In International Conference on Learning Representations. [6] Omidshafiei, S., Pazis, J., Amato, C., How, J. P., & Vian, J. (2019). AlphaStar: Mastering the real-time strategy game StarCraft II. Proceedings of the 38th International Conference on Machine Learning, PMLR 139, 2021, China, pp.1–15 [7] Foerster, J. N., Farquhar, G., Afouras, T., Nardelli, N., & Whiteson, S. (2019). Counterfactual multiagent policy gradients. In International Conference on Machine Learning (pp. 2046–2054). [8] Zheng, Y., Tang, H., Chen, H., Xu, Z., Hu, W., & Huang, Z. (2020). Coma: Communication efficient multiagent reinforcement learning with actor-critic methods. Proceedings of the AAAI Conference on Artificial Intelligence, 34(04), 3962–3969. [9] Iqbal, S., & Sha, F. (2019). Actor-attention-critic for multi-agent reinforcement learning. In International Conference on Machine Learning (pp. 2985–2994). [10] Son, J., Kim, J., Choi, H., & Kim, S. (2019). Qtran: Learning to factorize with transformation for cooperative multi-agent reinforcement learning. Proceedings of the AAAI Conference on Artificial Intelligence, 33(01), 4282–4289. [11] Rashid, T., Samvelyan, M., Van hasselt, H., Silver, D., & Nowé, A. (2020). Off-policy multi-agent decPOMDP planning with deep value iteration. In International Conference on Machine Learning (pp. 7921–7931). [12] Matsuo, Y., LeCun, Y., Sahani, M., Precup, D., Silver, D., Sugiyama, M., & Morimoto, J. (2022). Deep learning, reinforcement learning, and world models. Neural Networks, 152, 267–275. [13] Brunke, L., Greeff, M., Hall, A. W., Yuan, Z., Zhou, S., Panerati, J., & Schoellig, A. P. (2022). Safe learning in robotics: From learning-based control to safe reinforcement learning. Annual Review of Control, Robotics, and Autonomous Systems, 5, 411–444.
5 Reinforcement learning
123
[14] Ladosz, P., Weng, L., Kim, M., & Oh, H. (2022). Exploration in deep reinforcement learning: A survey. Information Fusion, 85, 1–22. [15] Fawzi, A., Balog, M., Huang, A., Hubert, T., Romera-Paredes, B., Barekatain, M., & Kohli, P. (2022). Discovering faster matrix multiplication algorithms with reinforcement learning. Nature, 610(7930), 47–53. [16] Degrave, J., Felici, F., Buchli, J., Neunert, M., Tracey, B., Carpanese, F., & Riedmiller, M. (2022). Magnetic control of tokamak plasmas through deep reinforcement learning. Nature, 602(7897), 414–419. [17] Wurman, P. R., Barrett, S., Kawamoto, K., MacGlashan, J., Subramanian, K., Walsh, T. J., & Kitano, H. (2022). Outracing champion Gran Turismo drivers with deep reinforcement learning. Nature, 602(7896), 223–228. [18] Wu, D., Lei, Y., He, M., Zhang, C., & Ji, L. (2022). Deep reinforcement learning-based path control and optimization for unmanned ships. Wireless Communications and Mobile Computing, 2022, 1–8. [19] Le, H., Wang, Y., Gotmare, A. D., Savarese, S., & Hoi, S. C. H. (2022). Coderl: Mastering code generation through pretrained models and deep reinforcement learning. Advances in Neural Information Processing Systems, 35, 21314–21328. [20] Seno, T., & Imai, M. (2022). d3rlpy: An offline deep reinforcement learning library. The Journal of Machine Learning Research, 23(1), 14205–14224. [21] Yuvaraj, N., Karthikeyan, T., & Praghash, K. (2021). An improved task allocation scheme in serverless computing using gray wolf Optimization (GWO) based reinforcement learning (RIL) approach. Wireless Personal Communications, 117, 2403–2421. [22] Silvia Priscila, S., Sathish Kumar, C., Manikandan, R., Yuvaraj, N., & Ramkumar, M. (2022, February). Interactive artificial neural network model for UX design. In International Conference on Computing, Communication, Electrical and Biomedical Systems (pp. 277–284). Cham: Springer International Publishing. [23] Awasthi, S., Naruka, M. S., Yadav, S. P., & De Albuquerque, V. H. C. (Eds.). (2023). AI and IoT-based intelligent health care & sanitation. U.A.E: Bentham Science Publishers. [24] Nayak, R. K., Tripathy, R., Saravanan, V., Das, P., & Anguraj, D. K. (2020, March). A novel strategy for prediction of cellular cholesterol signature motif from G protein-coupled receptors based on rough set and FCM algorithm. In 2020 Fourth International Conference on Computing Methodologies and Communication (ICCMC) (pp. 285–289). IEEE. [25] Sivapriya, T. R., & Saravanan, V. (2011). Automatic brain MRI mining using support vector machine and decision tree. CiiT International Journal of Artificial Intelligent Systems and Machine Learning, 3(2), 109–116.
Syed Ibad Ali, Padmavathi Thasappan Kalaivaani, Ambigaipriya S., MD Rafeeq
6 Evaluation of AI model performance Abstract: Artificial intelligence (AI) models are becoming more complex as a result of recent advances in AI technology. As a result, there are numerous approaches to assess the effectiveness of an AI model. In order to evaluate a model’s performance, both its accuracy and how quickly it processes fresh data are measured. Examining the model’s robustness to data that is not in the training set and its capacity for sophisticated problem solving are also included in the evaluation process. Because it identifies opportunities for improvement and facilitates better decision-making, interpretability is a crucial component of evaluating an AI model’s performance. Finally, a range of measures and methods are used in performance evaluation of AI models to assess their efficacy. Keywords: Artificial, intelligence, model, performance, complex, problems
6.1 Introduction The effectiveness of artificial intelligence (AI) models can be evaluated using a variety of metrics, such as accuracy, recall, precision, specificity, and F1-score. How accurately a model can classify an input serves as a gauge of its accuracy. Recall gauges how well a model can identify all pertinent instances of a particular class [1]. Precision evaluates the model’s capacity to provide pertinent results for a specific query. The model’s accuracy in identifying only pertinent results in response to a given query is measured by specificity. The F1-score, which combines recall and precision, serves as a measure of the model’s overall performance. To create an AI model that performs well, evaluation criteria should be selected based on the task at hand, and evaluated frequently.
6.1.1 Importance of evaluating AI model performance AI model performance evaluation is an important step in developing a successful AI system. This evaluation helps determine the accuracy and reliability of an AI system, as well as its potential drawbacks. Moreover, it is essential in helping AI developers to identify areas where the model needs improvement. Evaluation is also necessary for the selection of appropriate AI models for specific tasks [2]. Finally, evaluating AI model performance is critical in increasing the trustworthiness of AI systems, especially in scenarios where accuracy and reliability are critical. https://doi.org/10.1515/9783111323749-006
126
Syed Ibad Ali et al.
6.1.2 Role of evaluation metrics in assessing AI model effectiveness Evaluation metrics are key to assessing the effectiveness of an AI model. Metrics provide a comprehensive set of data points on which to measure progress and performance. Different evaluation metrics measure different aspects of an AI model’s performance. Metrics, including accuracy, precision, recall, AUC (area under the curve), and F1-score, are often employed. These metrics include details on the model’s ability to accurately identify objects, its capacity to make predictions, and its sensitivity to specificity. Knowing how well an AI model is performing from these metrics allows data teams to make informed decisions on how to improve the model. By utilizing evaluation metrics, teams can identify potential areas of improvement and perform modifications to the model that would increase the accuracy of its predictions [3]. The right evaluation metric can provide valuable insight into how an AI model is performing and help to ensure that the model is meeting its accuracy goals.
6.2 Confusion matrices A table called a confusion matrix is typically used to show how well a classification model (sometimes referred to as a “classifier”) performs on a set of test data for which the true values are known. It allows one to gauge an algorithm’s effectiveness. In the matrix, each row represents an occurrence of a predicted class, while each column represents an occurrence of an actual class, or vice versa. The name alludes to how easy it is to tell if the system is combining two classes (often mislabeling one as another) [4].
6.2.1 Definition and components of a confusion matrix A table called a confusion matrix is frequently used to show how well a classification model – also known as a “classifier” – performs on a set of test data for which the true values are known. It makes it possible to see how an algorithm performs. The confusion matrix is composed of four cells, each of which counts the number of true positive (TP), true negative (TN), false positive (FP), and false negative (FN) values. – TPs are the occurrences in which the model correctly predicted the positive class. – TNs are those instances that occur when the model correctly predicts the negative class. – FPs are situations where the model anticipated the positive class wrongly. – FNs are cases where the model predicted the negative class wrongly.
6 Evaluation of AI model performance
127
The confusion matrix provides for a more thorough examination than is possible when only looking at the accuracy score, making it a useful tool for evaluating the effectiveness of a classification algorithm. It enables the user to comprehend the model’s mistakes and how to enhance performance.
6.2.2 Sensitivity and specificity Two important metrics are utilized to assess a diagnostic test’s effectiveness: sensitivity and specificity. The percentage of true positive tests (those accurately recognized as having the target disease) out of all individuals with the condition is referred to as sensitivity [5]. The percentage of true negative tests (those accurately recognized as not having the target condition) out of all those without the condition is what is referred to as specificity. Most people with the condition will be accurately identified by a test with high sensitivity, and most people without the condition will be correctly identified by a test with high specificity.
6.2.3 Calculation from the confusion matrix A table known as a confusion matrix is typically used to show how well a classification model (also known as a “classifier”) performs on a set of test data for which the true values are known. It makes it possible to see how well an algorithm performs. The cases in each row of the matrix, as opposed to the examples, correspond to the actual classes, whereas the columns of the matrix, or vice versa, belong to the anticipated classes [6]. The name confusion matrix was created because it makes it simple to see when the system confuses two classes (i.e., when one is regularly referred to as another). The confusion matrix can be used to calculate a variety of metrics, such as precision, recall, accuracy, and the F1-score. – Accuracy: This is determined by dividing the percentage of correct forecasts by all other guesses. – Precision: Divide the total number of true positives by the total number of TPs and FPs to get at this. – Bear in mind that this corresponds to the sum of the TPs and FNs. – F1-score: A precision and recall harmonic mean is represented by this score. The F1-score is considered to be the most helpful indicator when comparing models.
128
Syed Ibad Ali et al.
6.2.4 Interpreting results and assessing model performance Interpreting the results of a predictive model and assessing its performance is a process that involves understanding the concepts of accuracy, precision, and recall in order to make an informed decision on whether a given predictive model is effective or not [7]. Precision indicates how often the model can predict the outcome accurately; recall is how many of the positive examples were properly identified; and accuracy measures how closely the predictions match the actual values. By taking into account all of these factors, one can evaluate the performance of a predictive model and make an informed decision on whether to use it or not.
6.3 Kappa statistics Kappa statistics is a metric used to measure the agreement between two raters, or observers, by taking into account chance agreement. It is expressed as a number from 0 to 1, with 1 signifying complete agreement and 0 signifying random chance agreement. The most frequent application of kappa is in inter-rater reliability testing, where it is used to gauge how consistently two or more raters make decisions on a given subject [8]. It is also used for nonbinary, multilevel ratings. For example, if there are 3 or more possible outcomes/ratings, kappa can be used to measure the reliability and agreement between raters.
6.3.1 Introduction to Cohen’s kappa coefficient A measure of how well two parties agree on a subject is known as the Cohen’s kappa coefficient, also known as the Cohen’s kappa statistic or Cohen’s kappa score. It is frequently employed when attempting to determine the level of agreement between two or more raters or observers on the same items in a collection. The kappa coefficient takes into account the agreement that can be expected by random chance, thus helping to eliminate any potential bias [9]. The result ranges from −1 (total disagreement) to 1 (total agreement). A score of 0 indicates that the raters are in agreement only to the extent expected by chance.
6.3.2 Measuring inter-rater agreement beyond chance Measuring inter-rater agreement beyond chance is the process of quantifying agreement between two or more raters on a set of criteria. The most common method for assessing such agreement is calculating the kappa statistic. This is done by taking into
6 Evaluation of AI model performance
129
account both the number of agreements and disagreements between the raters while also accounting for the possibility of agreement or disagreement by chance. Higher values of the kappa statistic, which ranges from 0 to 1, imply greater agreement than would be predicted by chance. Other methods, such as intra-class correlation coefficients, can also provide measures of inter-rater agreement.
6.3.3 Calculation and interpretation of kappa statistics Kappa statistics is a measure of agreement between two raters or assessments that rate or measure the same event or phenomenon. It is a way of measuring accuracy when a binary (yes/no, true/false) assessment is made by two different people. It is commonly used to evaluate things like medical diagnoses, lab tests, or surveys. It can also be used to compare different gold standards or quality checks [10]. Kappa statistics can be used to investigate the degree of agreement between two raters or assessments that have measured the same event or phenomenon. It measures the ratio of observed agreement (percent agreement) to a chance agreement (percent agreement). The calculation of the kappa statistic is relatively straightforward. First, we calculate the number of agreements and disagreements between two raters or assessments. Then, based on the presumption that the two raters are making their decisions independent of one another, we contrast it with a chance agreement. A kappa statistic between 0 and 1 is produced as a result, with a higher number being better. Typically, a set of qualitative categories are used to interpret kappa statistics. A kappa statistic of 0.41 to 0.60 is generally considered “slightly above chance,” whereas a score of 0.61 to 0.80 is “substantial” agreement, and a score greater than 0.81 is considered “almost perfect.”
6.3.4 Understanding kappa values and their significance Kappa values are a measure of agreement between two or more people who are rating the same item or event. They measure how well two judges agree with each other on a particular rating, with higher values representing greater consistency. Numerous situations, including employee evaluations, consumer satisfaction surveys, legal proceedings, and medical diagnoses, make use of kappa values [11]. The value is also often used to measure inter-rater reliability.
130
Syed Ibad Ali et al.
6.4 Receiver operating characteristic (ROC) curve The performance of a binary classifier is shown graphically by a receiver operating characteristic (ROC) curve. This graph illustrates the TP rate (TPR) and FP rate (FPR), with varying probabilities, for a binary classifier system when the discrimination threshold is changed. It can be utilized to assess the potency of multiple classifiers. The optimal one can be chosen for a particular issue [12]. To evaluate the accuracy of a classification, utilize the area under the ROC curve (AUC).
6.4.1 Introduction to ROC curves and area under the curve (AUC) The performance of a binary classifier – a classifier with two output classes – is graphically represented by ROC curves. A classifier’s performance under various parameter values is shown using a ROC curve. On the y-axis and the x-axis, respectively, the TPR and FPR are plotted for each potential parametric value [13]. The AUC on ROC curves measures how well a classifier can distinguish two groups. It is a single value that includes every ROC curve data point and provides a crystal clear picture of how well the classifier works. Since it can discriminate between the two classes more clearly, a classifier with a higher AUC is more accurate. While an AUC of 0.5 implies a classifier that delivers results that are no better than random, an AUC of 1.0 shows a perfect classifier that produces results with no overlap between the two groups.
6.4.2 Plotting the ROC curve and interpreting results Using ROC curves is a popular method for determining how well categorization models perform. The ROC curve is a graph that contrasts, on the x-axis, the true positive rate for a specific model with the FPR. The ROC curve is used to gauge how well a classifier can differentiate between real and fake positives (positive and negative classes). A typical statistic to evaluate a classifier’s prediction accuracy is the AUC. To plot a ROC curve, the concept is simple. To start off, you must have a set of data (or a model) that has been evaluated over a range of different thresholds (or decision boundaries). Determine the ratio of true positives to FPs for each level. Then, plot these places on a chart and draw a line that connects them. The graph that follows is the ROC curve [14]. Interpreting the results of a ROC curve is somewhat subjective, depending on the context, but it is nonetheless useful for gauging the performance of a model. The expected outcome for a random classifier is that a model with an AUC score closer to 0.5 has better predictive power than a model with an AUC value closer to 1, on average. Although two models may have comparable AUC values, this does not always imply
6 Evaluation of AI model performance
131
that they will perform similarly; other metrics like precision and recall can be utilized to better understand the model.
6.4.3 Evaluating the performance of binary classifiers An essential step in choosing the best classifier for a particular job is assessing a binary classifier’s performance. It entails judging how well the classifier can put data points into one of two categories. A binary classifier’s performance can be assessed using its accuracy, precision, recall, and F1-score. – Accuracy is the proportion of data points among all the data points that were correctly categorized. This metric demonstrates the effectiveness of the classifier. – Precision is the percentage of data points that were correctly classified relative to those that were actually given a positive label. – The percentage of correctly identified data points among all positive data points is measured by recall. – To evaluate the overall effectiveness of a binary classifier, the F1-score, a combined measure of accuracy and recall, is used. The harmonic mean of recall and precision can be used to calculate it. Overall, the performance of a binary classifier can be evaluated using these measures in order to choose the appropriate classifier for a given task.
6.5 Performance parameters Measurements used to assess an AI model’s performance are called performance parameters. These variables assess how well a machine learning or AI model performs the task for which it was designed [15]. A few of the numerous variables that determine an AI model’s performance include the type of data utilized, the complexity of the task, the amount of training it receives, and the techniques employed. Accuracy, precision, recall, AUC, and F1-score are common measures used to assess the performance of AI models.
6.5.1 Calculating accuracy and error rate from the confusion matrix When assessing a classification algorithm’s effectiveness, the confusion matrix is a highly helpful tool. It is used to evaluate a model’s performance in terms of recall, precision, the prevalence of FP and FN outcomes, forecast accuracy, and recall. The
132
Syed Ibad Ali et al.
confusion matrix, which includes information on the actual classifications and predictions made by the model, can be used to calculate the accuracy of the model as well as the error rate [16]. To determine accuracy, divide the total number of examples that were correctly categorized (represented by the diagonal of the confusion matrix) by the overall number of instances. The proportion or percentage of things that were correctly categorized is shown below. Accuracy is subtracted from 1 to get error rate. The percentage of incorrect classifications is shown here.
6.5.2 Balanced accuracy and its significance For binary classification problems, balanced accuracy is the most popular evaluation statistic. When the definitions of a “true positive” and “true negative” are unclear, this statistic, which is more reliable than standard accuracy, can be used to evaluate the performance of a model [17]. By taking into consideration both FPs and FNs, a model’s performance can be evaluated more accurately. Balanced accuracy can also help to prevent a model from learning an imbalanced dataset and instead, ensure an even level of performance across different types of predictions. This is especially important in situations involving important decisions where missed predictions can lead to significant negative implications.
6.5.3 Matthews correlation coefficient (MCC) A binary classifier’s effectiveness is measured by the Matthews correlation coefficient (MCC). Higher values indicate higher classification accuracy. It is a binary classification task evaluation metric with a range of −1 to +1. The MCC is used to determine how well the anticipated classifications from the model and the observed classifications coincide. It is a sign that the model can correctly categorize samples from both classes. A score of 0 implies no better than random prediction, whereas a perfect score of +1 shows faultless prediction [18].
6.6 Cross-validation A model validation technique called cross-validation evaluates how well the findings of a statistical analysis transfer to a different dataset. It is primarily utilized in situations where comparing machine learning algorithms is the major objective. The original sample is partitioned into k equal-sized subsamples at random for k-fold cross-validation. K1 subsamples are utilized as training data out of a total of k subsamples, while the remaining k subsamples are used as validation data to test the model. Following that, the
6 Evaluation of AI model performance
133
cross-validation process is carried out k times, using the validation data from each of the k subsamples just once [19]. By averaging the k outcomes from the folds, an estimation can then be made. This approach has the advantage of repeated random subsampling in that all observations are used for training and validation, and each observation is used for validation exactly once.
6.6.1 Introduction to cross-validation techniques In order to assess supervised learning models and avoid overfitting, a technique called cross-validation is performed. This involves testing the model against a dataset different from the training data. It works well for selecting the best model parameters in support vector machines (SVMs), such as the kernel parameter and regularization level. Cross-validation, which divides the training dataset into numerous folds, evaluates the performance of the model fitted with each fold. The dataset is initially divided into k equal sections before doing k-fold cross-validation on k separate trials [20]. Each experiment uses one of the components for testing while using the other components for training. The average performance throughout the k experiments serves as the basis for the estimation of model performance.
6.6.2 Benefits of cross-validation in assessing model generalization In order to evaluate a model’s performance on untested data, cross-validation is an essential machine learning technique. It is useful to evaluate a model’s generalizability to new data rather than just the data it was trained on. It can help to reduce bias in model estimates and improve accuracy. Advantages of cross-validation include: – Reduced optimizing hyperparameters: The ability to test different hyperparameters on subsets of the data helps to reduce the risk of overfitting the data. – Improved model accuracy: By using multiple subset tests, the model can be tested on more than one dataset and this can help to increase accuracy. – Greater insight into data behavior: Cross-validation can provide insight into how well a model generalizes to unseen data by looking at how it performs on different datasets. It can also help to identify any potential bias that should be removed from the model. – Improved predictive coding: By testing a model on different subsets of the data, it can better understand the data structure and detect the unusual patterns that may exist. This can lead to improved predictive coding.
134
Syed Ibad Ali et al.
Overall, cross-validation is a powerful tool that helps to improve machine learning models and generalizes the models to new data. It provides better insight into the data structure and can help to reduce the risk of over-fitting.
6.6.3 Implementing cross-validation and evaluating performance across multiple folds Cross-validation is a method for assessing how well a specific machine learning model is performing. Finding the model that will perform the best on unforeseen data is beneficial. It is used to assess how a model will generalize to independent data sets. It divides a dataset into multiple partitions, called folds. Each fold is used as a testing dataset while all other folds are used as training datasets [21]. The model is then trained on each fold and tested using the appropriate testing set. An average score is then calculated after evaluating the model’s performance across all folds. Crossvalidation is a helpful technique for evaluating a machine learning model’s overall performance since it takes the dataset’s unpredictability into account. Every single piece of data in the dataset is used for both training and testing; so it can be used to fairly evaluate the performance of the model.
6.7 Model selection and hyperparameter tuning Model selection is the process of selecting the most suitable model or algorithm for a supervised machine learning problem out of a set of potential models. It involves selecting models (or algorithms) based on specific criteria and comparing them using evaluation metrics. Model selection involves considering factors such as the problem being solved (classification, regression, etc.), type of data (discrete, continuous, etc.) and the goals of a model (accuracy, computational time, etc.) [22]. Enhancing a model’s hyperparameters through a process of trial and error is known as “hyperparameter tuning”. Finding the set of hyperparameters that performs best on a validation set is the aim. Any machine learning model can profit from hyperparameter tuning, but sophisticated models with lots of hyperparameters, like SVMs and DNNs, gain the most.
6.7.1 Importance of model selection and hyperparameter tuning Model selection and hyperparameter tuning are two essential parts of any machine learning task. Model selection involves choosing a suitable learning algorithm, and
6 Evaluation of AI model performance
135
hyperparameter tuning is modifying the algorithm’s performance by aiming to optimize its hyperparameters [23]. There are a few key reasons why model selection and hyperparameter tuning are important: – It help optimize your models’ performance. Through experimentation with different algorithms, it allows you to make the best possible choice, based on your data and task. It can also ensure that an appropriate amount of bias is selected, as too much can lead to underfitting and too little can lead to overfitting. – It can save huge amounts of time. By properly tuning your hyperparameters, you can avoid wasting time training a model that does not perform well. – It can ensure that you are using the most suitable algorithm for the task. This is important because using the wrong algorithm can cause models to perform badly. Model selection and hyperparameter tuning are vital processes for the optimization of the machine learning process. As such, it should be an integral part of any machine learning project.
6.7.2 Cross-validation for model selection By partitioning the data, training the model on a piece of the data, and then assessing the model’s performance on a different subset of the data, the machine learning approach of cross-validation is used to evaluate a model. Since it divides the data into “folds” and utilizes each data partition to estimate the performance of the model, this technique is also used to identify the model with the highest accuracy [24]. It is crucial for model selection since it not only aids in choosing the best model but also in identifying the perfect set of hyperparameters, which enhances the model’s generalizability and accuracy. Cross-validation aids in determining a model’s potential performance when used with “unseen” data.
6.7.3 Grid search and random search for hyperparameter optimization Grid search is a method for optimizing hyperparameters that thoroughly explores a manually chosen portion of the hyperparameter space. It is usually used in small data sets where its exhaustive search is feasible. It works by training individual models for each combination of hyperparameters in the grid. The hyperparameter set with the best performance is then chosen. Random search is a technique for hyperparameter optimization that samples many different combinations of hyperparameters from a given distribution. It does not attempt to evaluate all possible combinations of hyper-
136
Syed Ibad Ali et al.
parameters, but instead selects a few at random to focus on. As new hyperparameter sets are sampled, models are trained and evaluated, and the best (or best-performing) hyperparameter set is selected [25]. In short, grid search provides an exhaustive search of the hyperparameter space to find the best combination of parameters, while random search provides a more stochastic approach, sampling a random combination of parameters to find the best combination.
6.7.4 Evaluating performance on validation sets A machine learning model’s performance on a validation dataset is assessed using the “Evaluating performance on validation sets” approach. A collection of data that has been divided and preprocessed for the purposes of developing and testing a machine learning model is known as a validation set. On a validation set, a model’s performance is evaluated using accuracy, precision, recall, and other performance measures. After comparing the model’s performance to those of other models, the best model can be selected for additional testing and deployment. The goal of evaluating performance on validation sets is to identify the model that will have the best performance in practical applications [26]. This helps ensure the model is able to generalize well to new data, and that it has the highest likelihood of performing well in production.
6.8 Bias and fairness evaluation Bias and fairness evaluation are the processes of examining and measuring the fairness of models as well as the data used to train or to generate them. The objective is to increase fairness and accuracy while reducing the possibility of bias in decision results. In the context of artificial intelligence and machine learning, this analysis can identify and eliminate any potential biases and prejudice in software algorithms and the data used to construct them [27]. Bias and fairness evaluation also provide guidance on responsible data usage practices to reduce bias and reduce potential risk. The evaluation process typically includes exploring data sources used to create software algorithms, evaluating model performance, and identifying mitigating strategies to reduce bias.
6.8.1 Addressing bias and fairness issues in model evaluation Addressing bias and fairness issues in model evaluation is an important step in building an effective and reliable model. It is essential to evaluate the model for potential bias and fairness as it can lead to incorrect conclusions and lead to costly errors in
6 Evaluation of AI model performance
137
the future. In order to evaluate a model for bias and fairness, it is important to analyze how it treats different segments of the population. The evaluation should consider statistical evidence such as correlations, measures of socio-economic disparities, and other external factors that may influence predictions. Additionally, it is important to review the model’s design, data, and results to identify potential bias or unfairness [28]. Finally, the model should be tested in a variety of scenarios and circumstances in order to ensure its accuracy. Addressing these issues will help ensure the model is reliable and will lead to better decisions and outcomes.
6.8.2 Fairness metrics Fairness or equity of decision-making processes, policies, and procedures is measured and evaluated using fairness metrics. They can be used to evaluate how machine learning models affect various demographic groups, law enforcement agencies, organizations, and nations. Fairness metrics have been developed to measure inclusiveness and equity in various domains, covering both quantitative and qualitative attributes such as race, gender, age, nationality, or educational level [29]. The purpose of fairness metrics is to identify the extent of fairness in a system and to provide insight into the potential of an overall system to be biased. They can also be used to measure the degree to which a decision making process is fair and equitable, or to discover unintended consequences of a policy or machine learning model. In general, fairness metrics quantify and compare equitable and equitable outcomes between different groups of people affected by a system, policy, or process.
6.8.3 Assessing model fairness using evaluation techniques A model fairness evaluation is a collection of techniques for determining how fair a machine learning model is. These techniques are used to understand and prevent discrimination in areas such as Credit Decisioning, Lending, Healthcare, Insurance, and Employment. The goal is to understand how the model is predicting outcomes or decisions for different segments of the population, identify potential bias or discrimination, and proactively address any issues. Essentially, it is about making sure the model does not make decisions that are biased toward any particular group, and instead treats all individuals equitably and fairly. A/B testing, accuracy, recall, precision, F1, and lift scores are a few examples of fairness metrics that can be used to test data for fairness. Other techniques include using visualization tools to identify the data features that are influencing the model’s decisions and interpreting the model’s outputs to make sure they do not negatively affect any population [30]. The ultimate objective is to guarantee that a model is equitable to all individuals, regardless of their
138
Syed Ibad Ali et al.
color, gender, or other demographic characteristics. Building moral and accountable machine learning models requires doing fairness evaluations.
6.9 Interpretability and explainability The ability of AI systems to provide explanations of their model output and reasoning is referred to as both interpretability and explainability. Interpretability is the ability to understand the reason behind the conclusion the AI system has arrived at. Explainability is the ability to explain the logic that was used to arrive at that conclusion. Both are important in helping people trust and use AI models. Explainability is especially important when the system is making decisions that have legal or ethical implications, as it enables users to determine whether the AI system is being fair and accurate [31]. Interpretability allows developers and data scientists to debug AI models and ensure that models are performing as expected.
6.9.1 Interpreting model predictions and decisions Interpreting model predictions and decisions is the process of understanding why a model makes certain predictions or decisions. It involves deciphering the logic that the model is using to come up with its predictions or decisions. This requires looking at the data available to the model, the features the model is using for analysis, and understanding how the model’s algorithms are working to classify or estimate the data. It can also involve identifying potential pitfalls or areas of bias in the model. As machine learning models become more popular and widely used, understanding how they make decisions or predictions is becoming increasingly important for both researchers and practitioners who are using them [32]. The accuracy of a model’s predictions can be improved by understanding why it makes certain decisions, and it can also assist practitioners in selecting the best models to use in certain contexts.
6.9.2 Evaluation of interpretability methods Methods used to increase the human interpretability of machine learning models are referred to as interpretability methods. These strategies include creating individual conditional expectation (ICE) curves and partial dependence plots (PDPs), calculating feature importance scores to determine the model’s most crucial characteristics, and employing feature selection techniques to limit the amount of input features. Evaluating interpretability methods can include assessing the utility of the methods for model developers and model users. For model developers, the effectiveness
6 Evaluation of AI model performance
139
and accuracy of the interpretability methods in helping to explain the model and designing strategies to improve the model should be evaluated. For model users, the degree to which interpretability methods enable them to understand and use the model to make decisions should be assessed [33]. Additionally, the user experience associated with using these interpretability methods should be evaluated in terms of user interface design, user friendliness, and the ease with which users can interpret the outputs.
6.9.3 Assessing the trade-off between model performance and interpretability In machine learning, weighing the trade-offs between model performance and interpretability is a crucial step. This trade off refers to the fact that more complicated models may be more difficult to comprehend and interpret while being more accurate at predicting outcomes, whereas simpler models may be more interpretable but less at doing so. It is crucial to take into account the model’s performance and interpretability when choosing the type to use. In some cases, such as an application involving medical decisions, interpretability may be more important than performance. On the other hand, in cases where high accuracy is required, such as autonomous driving or fraud detection, performance may be more important than interpretability. In general, both features should be assessed when selecting a model and the one that best meets the needs of the application should be chosen.
6.10 Challenges and best practices Challenges: – Explaining AI decisions: As AI systems become more advanced, it becomes harder to explain how AI makes its decisions. This lack of transparency can lead to unanticipated risks or even bias in decision-making. – Security and privacy: AI systems can be vulnerable to malicious attacks due to their reliance on large amounts of data, which could expose personal information. Additionally, AI can be used for malicious activities like hacking, data manipulation, or identity theft. – Data quality and Integrity: AI programs require access to quality data to provide accurate results. Poor-quality data that is too limited, inconsistent, or outdated can lead to problems for AI systems.
140
Syed Ibad Ali et al.
Best Practices: – Explainability: Developers must ensure AI systems are explainable and easily understood by users. AI decision-making should be traceable, so users can track why specific decisions were made. – Security and privacy: Develop robust security measures to protect AI systems and associated data. This includes creating safeguards against malicious actors, such as ensuring users have proper authentication and authorization to access systems. – Data quality and Integrity: The reliability and integrity of the data have a big impact on how accurate AI systems are. Ensure that data is dependable, consistent, and frequently updated to maximize the accuracy of outcomes.
6.10.1 Common challenges in AI model evaluation Common challenges in AI model evaluation include bias, lack of interpretability in the model, and inadequate data. – Bias: AI models often incorporate bias from the data that they are trained on, which can lead to poor results or unintended outcomes. AI models tend to mirror the biases in the data that is used to create them; so data must be carefully selected and monitored for bias. – Lack of interpretability in the model: AI models are often quite complex, and it can be difficult to fully understand how and why they are producing the results they are. This makes it difficult to evaluate the model’s accuracy and reliability. – Inadequate data: AI models require large amounts of data in order to learn and produce accurate results. If the amount of data that is available is insufficient, then the model’s accuracy and reliability will likely be compromised.
6.10.2 Overfitting and data leakage When a machine learning model is extremely complex, overfitting happens. As a result, it learns the best solution for the training dataset but struggles to generalize effectively to new datasets [34]. It is caused by having too many features or too complex models such as deep learning networks. Data leakage, on the other hand, is the unintentional inclusion of information in the training dataset that results in an overly optimistic model performance. When training data contains specifics that must be used to evaluate the model’s performance rather than making predictions, this happens. As a result, the model develops the ability to categorize unknown data points using knowledge that was unavailable during inference, leading to an optimistic performance. To avoid data leakage, it is
6 Evaluation of AI model performance
141
important to make sure that training data does not contain data that is not available during inference.
6.10.3 Selection bias and imbalance in evaluation datasets A form of bias known as selection bias happens when a researcher decides to only look at a portion of the population under consideration. This means that the sample data they use could be disproportionately biased toward certain individuals or factors, resulting in an inaccurate depiction of the overall population [35]. Imbalance in evaluation datasets can occur when the evaluation dataset does not accurately reflect the proportion of different class labels or categories in the original dataset. For example, a large dataset may have 80% of samples from one class label, while the evaluation dataset only contains 30% of those samples. This can lead to a biased evaluation as the test model might not be able to accurately represent the full range of data and experiences that it should be able to.
6.10.4 Best practices for robust model evaluation Any machine learning modeling project must include a thorough model evaluation. A robust model evaluation involves a careful assessment of the model’s performance against different types of data sets, different evaluation metrics, and under various conditions [36]. Best practices for robust model evaluation include: – Test on held-out data: Whenever possible, validate models on a set of held-out data. This will give a more accurate assessment of how the model will perform in production. – Use multiple metrics: Utilizing numerous evaluation metrics can help you gain a deeper knowledge of the model’s performance – using F1-score, accuracy, precision, recall, etc. – Compare to baselines: Comparing the model’s performance against various baselines, such as random guessing or a straightforward model like logistic regression, might be useful. – Use cross-validation and hyperparameter tuning: Make that the model is not under- or overfitting the data. Optimize the model using hyperparameter tweaking, and evaluate the model’s generalizability using cross-validation. – Check assumptions: Ensure that the assumptions that you are making about the data are valid. This includes checking for data leakage or bias in the data.
142
Syed Ibad Ali et al.
References [1]
[2]
[3]
[4]
[5]
[6]
[7]
[8]
[9] [10]
[11]
[12]
[13] [14]
[15] [16]
Allen, B., Dreyer, K., Stibolt Jr, R., Agarwal, S., Coombs, L., Treml, C., . . . & Wald, C. (2021). Evaluation and real-world performance monitoring of artificial intelligence models in clinical practice: try it, buy it, check it. Journal of the American College of Radiology, 18(11), 1489–1496. Tabbussum, R., & Dar, A. Q. (2021). Performance evaluation of artificial intelligence paradigms – artificial neural networks, fuzzy logic, and adaptive neuro-fuzzy inference system for flood prediction. Environmental Science and Pollution Research, 28(20), 25265–25282. Hicks, S. A., Strümke, I., Thambawita, V., Hammou, M., Riegler, M. A., Halvorsen, P., & Parasa, S. (2022). On evaluation metrics for medical applications of artificial intelligence. Scientific Reports, 12(1), 5979. Li, L., Qin, L., Xu, Z., Yin, Y., Wang, X., Kong, B., . . . & Xia, J. (2020). Using artificial intelligence to detect COVID-19 and community-acquired pneumonia based on pulmonary CT: evaluation of the diagnostic accuracy. Radiology, 296(2), E65–E71. Dumitrascu, O., Dumitrascu, M., & Dobrotǎ, D. (2020). Performance evaluation for a sustainable supply chain management system in the automotive industry using artificial intelligence. Processes, 8(11), 1384. Ghahramani, M., Qiao, Y., Zhou, M. C., O’Hagan, A., & Sweeney, J. (2020). AI-based modeling and data-driven evaluation for smart manufacturing processes. IEEE/CAA Journal of Automatica Sinica, 7(4), 1026–1037. Khanagar, S. B., Al-Ehaideb, A., Maganur, P. C., Vishwanathaiah, S., Patil, S., Baeshen, H. A., . . . & Bhandi, S. (2021). Developments, application, and performance of artificial intelligence in dentistry–A systematic review. Journal of Dental Sciences, 16(1), 508–522. Wu, E., Wu, K., Daneshjou, R., Ouyang, D., Ho, D. E., & Zou, J. (2021). How medical AI devices are evaluated: limitations and recommendations from an analysis of FDA approvals. Nature Medicine, 27(4), 582–584. Amin, M. M., Cambria, E., & Schuller, B. W. (2023). Will affective computing emerge from foundation models and general ai? A first evaluation on chatgpt. arXiv preprint arXiv:2303.03186. Forghani, R., Savadjiev, P., Chatterjee, A., Muthukrishnan, N., Reinhold, C., & Forghani, B. (2019). Radiomics and artificial intelligence for biomarker and prediction model development in oncology. Computational and Structural Biotechnology Journal, 17, 995. Tobler, P., Cyriac, J., Kovacs, B. K., Hofmann, V., Sexauer, R., Paciolla, F., . . . & Hirschmann, A. (2021). AI-based detection and classification of distal radius fractures using low-effort data labeling: evaluation of applicability and effect of training set size. European Radiology, 31, 6816–6824. AI-Atroshi, C., Rene Beulah, J., Singamaneni, K. K., Pretty Diana Cyril, C., Neelakandan, S., & Velmurugan, S. (2022). Automated speech based evaluation of mild cognitive impairment and Alzheimer’s disease detection using with deep belief network model. International Journal of Healthcare Management, 2022, 1–11. Gunning, D., & Aha, D. (2019). DARPA’s explainable artificial intelligence (XAI) program. AI Magazine, 40(2), 44–58. Zou, Y., Zheng, C., Alzahrani, A. M., Ahmad, W., Ahmad, A., Mohamed, A. M., . . . & Elattar, S. (2022). Evaluation of artificial intelligence methods to estimate the compressive strength of geopolymers. Gels, 8(5), 271. Yadav, S. P., & Yadav, S. (2020). Image fusion using hybrid methods in multimodality medical images. Medical & Biological Engineering & Computing, 58, 669–687. Irfan, D., Tang, X., Narayan, V., Mall, P. K., Srivastava, S., & Saravanan, V. (2022). Prediction of Quality Food Sale in Mart Using the AI-Based TOR Method. Journal of Food Quality, 2022, 1–13
6 Evaluation of AI model performance
[17]
[18]
[19] [20]
[21]
[22] [23]
[24]
[25]
[26]
[27] [28]
[29]
[30]
[31]
[32]
143
Yuvaraj, N., Srihari, K., Dhiman, G., Somasundaram, K., Sharma, A., Rajeskannan, S. M. G. S. M. A., . . . & Masud, M. (2021). Nature-inspired-based approach for automated cyberbullying classification on multimedia social networking. Mathematical Problems in Engineering, 2021, 1–12. Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). Energy‐efficient resource allocation model for device‐to‐device communication in 5G wireless personal area networks. International Journal of Communication Systems, 2023, 1–14. Yadav, S. P., Agrawal, K. K., Bhati, B. S., Al-Turjman, F., & Mostarda, L. (2022). Blockchain-based cryptocurrency regulation: An overview. Computational Economics, 59(4), 1659–1675. Rajalakshmi, M., Saravanan, V., Arunprasad, V., Romero, C. T., Khalaf, O. I., & Karthik, C. (2022). Machine Learning for Modeling and Control of Industrial Clarifier Process. Intelligent Automation & Soft Computing, 32(1), 339–359. Manikandan, V., & Yuvaraj, N. (2019, May). Advanced expert system using particle swarm optimization based adaptive network based fuzzy inference system to diagnose the physical constitution of human body. In Emerging Technologies in Computer Engineering: Microservices in Big Data Analytics: Second International Conference, ICETCE 2019, Jaipur, India, February 1–2, 2019, Revised Selected Papers, 985, 349. Springer. Sangeetha, R., Logeshwaran, J., Rocher, J., & Lloret, J. (2023). An Improved Agro Deep Learning Model for Detection of Panama Wilts Disease in Banana Leaves. AgriEngineering, 5(2), 660–679. Yadav, S. P., Zaidi, S., Mishra, A., & Yadav, V. (2022). Survey on machine learning in speech emotion recognition and vision systems using a recurrent neural network (RNN). Archives of Computational Methods in Engineering, 29(3), 1753–1770. Saravanan, V., & Neeraja, A. (2013, January). Security issues in computer networks and stegnography. In 2013 7th International Conference on Intelligent Systems and Control (ISCO) (pp. 363–366). IEEE. Gobinathan, B., Mukunthan, M. A., Surendran, S., Somasundaram, K., Moeed, S. A., Niranjan, P., . . . & Sundramurthy, V. P. (2021). A novel method to solve real time security issues in software industry using advanced cryptographic techniques. Scientific Programming, 2021, 1–9. Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). A novel architecture of intelligent decision model for efficient resource allocation in 5G broadband communication networks. ICTACT Journal On Soft Computing, 13(3), 2986–2994. Yadav, S. P., Mahato, D. P., & Linh, N. T. D. (Eds.). (2020). Distributed artificial intelligence: A modern approach. Boca Raton: CRC Press. Singh, J., Deepika, J., Sathyendra Bhat, J., Kumararaja, V., Vikram, R., Jegathesh Amalraj, J., . . . & Sakthivel, S. (2022). Energy-efficient clustering and routing algorithm using hybrid fuzzy with grey wolf optimization in wireless sensor networks. Security and Communication Networks, 2022, 1–13 Natarajan, Y., Kannan, S., & Mohanty, S. N. (2021). Survey of various statistical numerical and machine learning ontological models on infectious disease ontology. Data Analytics in Bioinformatics: A Machine Learning Perspective, 13(1), 431–442. Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). L-RUBI: An efficient load-based resource utilization algorithm for bi-partite scatternet in wireless personal area networks. International Journal of Communication Systems, 2023, 1–15. Wan, Y., Tong, Y., Liu, Y., Huang, Y., Yao, G., Chen, D. Q., & Liu, B. (2022). Evaluation of the combination of artificial intelligence and radiologist assessments to interpret malignant architectural distortion on mammography. Frontiers in Oncology, 12, 1739. Hosamo, H. H., Nielsen, H. K., Kraniotis, D., Svennevig, P. R., & Svidt, K. (2023). Digital Twin framework for automated fault source detection and prediction for comfort performance evaluation of existing non-residential Norwegian buildings. Energy and Buildings, 281, 112732.
144
Syed Ibad Ali et al.
[33] Lee, D. H., Woo, S. E., Jung, M. W., & Heo, T. Y. (2022). Evaluation of odor prediction model performance and variable importance according to various missing imputation methods. Applied Sciences, 12(6), 2826. [34] Markus, A. F., Kors, J. A., & Rijnbeek, P. R. (2021). The role of explainability in creating trustworthy artificial intelligence for health care: A comprehensive survey of the terminology, design choices, and evaluation strategies. Journal of Biomedical Informatics, 113, 103655. [35] Orhan, K., Bayrakdar, I. S., Ezhov, M., Kravtsov, A., & Özyürek, T. A. H. A. (2020). Evaluation of artificial intelligence for detecting periapical pathosis on cone‐beam computed tomography scans. International Endodontic Journal, 53(5), 680–689. [36] Schwartz, R., Dodge, J., Smith, N. A., & Etzioni, O. (2020). Green ai. Communications of the ACM, 63(12), 54–63.
K. Arthi, Veeramalai Sankaradass, Nikhat Parveen, Muralidharan J.
7 Methods of cross-validation and bootstrapping Abstract: A statistical method called cross-validation is used to determine a machine learning model’s parameters and gauge its correctness. It is a type of resampling method where the dataset is arbitrarily split into training and test sets. Using different subsets of the same data, the model is tested and trained. Comparing the projected values from the testing data with the actual values of the same data allows one to evaluate the model’s performance. Keywords: Cross-validation, technique, accuracy, dataset, random
7.1 Introduction A statistical resampling method called “bootstrapping” is used to evaluate the effectiveness of a machine learning model. This technique replaces the need for k-fold cross-validation and can quickly provide measures of accuracy on unseen data. Bootstrapping randomly generates new datasets out of the original dataset by using sampling with replacement. This technique is an efficient way of obtaining new datasets with a higher degree of variance than the original dataset [1]. A machine learning model’s performance can be evaluated more robustly and precisely with the use of bootstrapping.
7.1.1 Importance of cross-validation and bootstrapping in model evaluation Both bootstrapping and cross-validation are significant techniques for assessing a model’s performance. Cross-validation is a model evaluation method that separates the dataset into subsets for training and validation in order to evaluate how well the model performs on fresh data. It is employed to evaluate how well a model generalizes from its training set of data. Additionally, when the model is trained and tested on the same data, bias and variance issues arise. Cross-validation aids in removing these issues. Bootstrapping is a model evaluation technique, which involves resampling the dataset with replacement to generate multiple datasets [2]. After that, the model is trained and tested using each dataset. By testing the model on numerous datasets rather than just one, it is possible to quantify its performance correctly. Bootstrapping also helps reduce bias and variance issues that can occur when a model is https://doi.org/10.1515/9783111323749-007
146
K. Arthi et al.
evaluated on a single dataset. Both cross-validation and bootstrapping are important methods for measuring a model’s performance and for providing an accurate assessment of its ability to generalize to unseen data. This helps ensure that only the best performing models are chosen for deployment.
7.2 Cross-validation techniques Cross-validation is a technique for evaluating the correctness of a machine learning algorithm that divides the data set into a training set and a test set to develop a model and validate the model. It helps to spot overfitting and select the best model by applying a variety of procedures, such as k-fold cross-validation or stratified cross-validation. The model’s performance can be estimated and its hyperparameters can be adjusted with the use of cross-validation. It can be used as a statistic to compare several models and decide which model is better for deployment [3].
7.2.1 k-Fold cross-validation The k-fold cross-validation approach is used to evaluate a machine learning model in order to assess how effectively the model generalizes to new data. The process involves randomly dividing the original data set into k-folds (or subsets) of equal size, training the model on k−1 of these folds, and testing it on the final fold. Once with each of the k folds acting as the testing set and once as the training set, the method is then repeated k times. A more accurate evaluation of how well the model works with unseen data is then obtained by averaging the results of each experiment. Using this technique, we can reduce the amount of data needed to test the model while improving our ability to anticipate how well the model performs on unobserved data [4].
7.2.1.1 Explanation of k-fold cross-validation k-Fold cross-validation is one technique for evaluating a machine learning model’s accuracy. With this method, a dataset is split into k subgroups, and k-rounds of training and testing are performed on each subset. Each round uses a different subset of the data for training and the remaining data for testing. Every round includes an evaluation of the model’s accuracy. The mean accuracy of the k trials is then used to estimate the model’s accuracy. k-Fold cross-validation is an effective method for assessing a model’s accuracy because it not only guards against overfitting but also has the capacity to produce more accurate estimations of the model accuracies.
7 Methods of cross-validation and bootstrapping
147
7.2.1.2 Steps involved in k-fold cross-validation The k-fold cross-validation method is used to assess the accuracy and fairness of a machine learning algorithm. k-Fold cross-validation randomly selects k subsets (or “folds”) from a dataset [5]. Then, each subset is used as a training set and a test set once each. After this technique has been applied k times, the final result is obtained by averaging the outcomes of all k trials. 1. Data division into k subsets (folds): The first step is to divide the data into k equal-sized, nonoverlapping sections (or “folds”). 2. Train and test the model k times: The model is then trained and tested k times, each on a distinct fold. 3. Average the model results: To obtain the final model score, the outcomes from all k-folds are averaged.
7.2.1.3 Benefits and limitations of k-fold cross-validation Multiple test iterations are conducted using the k-fold cross-validation technique to verify the accuracy of a model. The training set for each iteration is a portion of the data, and the testing set is the remainder. The average testing accuracy of different models can be calculated by doing a loop through all the data. Benefits include: – A reduction in overfitting’s negative effects. k-Fold cross-validation gives an objective assessment of the model’s performance because each data point is used for testing. – It provides an estimated population error rate: By performing multiple iterations, the average test accuracy can be estimated, thus providing a good indication of the model’s performance on unseen data. Limitations: – Inefficient use of data: Since each iteration requires a portion of the data to be used for training and a portion for testing, it is not possible to use all the data simultaneously. – It may suffer from selection bias: Selecting a randomly chosen subset of data points from the pool may not be representative of the population. Thus, k-fold cross-validation may lead to inaccurate results.
7.2.2 Stratified cross-validation Stratified cross-validation is a method of model validation, which helps ensure that all the data in a given set of samples are equally represented in the different validation
148
K. Arthi et al.
sets. It is typically used when there is a significant imbalance in the data, such as when classes are unevenly distributed. Stratified cross-validation is a type of cross-validation that attempts to address this problem by dividing the data into subsets according to the class labels [6]. Then, it partitions the subsets into testing and validation sets. As a result, each sample is guaranteed to be represented in both the validation and test sets. This strategy aids in ensuring that a model is reflective of the entire dataset, rather than just the classes with the highest percentages.
7.2.2.1 Definition and purpose of stratified cross-validation For evaluating and contrasting several models or approaches, machine learning employs a sort of cross-validation called stratified cross-validation. By dividing the data into various folds, it becomes easier to assess how well machine learning models are performing. With stratified cross-validation, each class is represented equally in both the training and test sets for each fold [7]. As a result, the model will be tested and trained using data sets that are representative of the entire population. Additionally, when working with tiny datasets, it can aid in preventing bias or distorted outcomes.
7.2.2.2 Application in handling imbalanced datasets In handling imbalanced datasets, application is a set of tools and techniques that can be used to deal with the problem of imbalanced datasets. These tools and techniques can help balance the data by either adding more minority class examples or by removing majority class examples. Some of these techniques include downsampling, clustering, and oversampling. When a dataset is downsampled, instances of the majority class are randomly removed while remaining examples of the minority class are retained. Clustering algorithms can be used to assign similar examples into clusters and remove outliers [8]. Oversampling involves creating additional instances of the minority class from existing examples. The application of these tools and techniques can help reduce the bias in machine learning models trained on imbalanced datasets.
7.2.3 Leave-one-out cross-validation (LOOCV) Leave-one-out cross-validation (LOOCV) is a cross-validation technique used to assess the accuracy of predictive models. The procedure comprises splitting the data into training and test sets, with one sample in each test set. After training with the training set, the data is compared to the single sample of the test set. Since this is repeated for each sample in the data collection, all samples can be used in both the training and
7 Methods of cross-validation and bootstrapping
149
testing sets [9]. The accuracy of the model is then assessed by averaging the outcomes of each iteration. Particularly when the data set is small, this method can offer estimates of model accuracy that are more accurate than those provided by conventional cross-validation techniques.
7.2.3.1 Explanation of LOOCV LOOCV is a validation technique that involves removing one data point from the dataset and using the remaining details to develop a predictive model. After that, the model is assessed using the same criteria as the remaining data on the deleted data point. Every last data point in the set is subjected to this process once more, and the results are averaged to provide an overall assessment of model performance [10]. When determining the precision and dependability of predictive models, LOOCV is a popular method since it is effective at decreasing bias and volatility in model evaluation.
7.2.3.2 LOOCV and its implications Cross-validation that uses one data point for testing while using the remaining data points for training is known as leave-one-out cross-validation (LOOCV). When N is the total number of data points, this operation is performed N times. Since the complete dataset must be examined several times, this is one of the most demanding types of validation and is frequently quite time-consuming. The implications of using LOOCV are that the accuracy of the model is maximized while minimizing the bias. Since all observations are used for both test and training, the risk of overfitting or underfitting is significantly reduced. This method yields more reliable predictions than other cross-validation techniques [11].
7.2.4 Nested cross-validation Nested cross-validation is a technique used to evaluate the model selection and parameter tuning of a predictive model. It involves conducting an inner loop of cross-validation to tune the model and an outer loop of cross-validation to test the effectiveness of the model selection and/or parameter tuning. This technique is used to minimize overfitting of the model and to ensure an unbiased estimate of the model’s generalization error [12]. This technique is more computationally expensive than traditional test/train splitting, but can give higher accuracy estimated out-of-sample performance.
150
K. Arthi et al.
7.2.4.1 Introduction to nested cross-validation A method for evaluating a model’s generalization performance is called nested crossvalidation, also referred to as double cross-validation. When inner and outer crossvalidation cycles are necessary to calculate a model’s generalization error, this technique is used. The hyperparameters of a machine learning model are optimized using the results of a conventional k-fold cross-validation that is carried out in the inner loop of this method. The outer loop then runs a second k-fold cross-validation once the inner loop is finished to calculate the generalization error of the model using the optimized hyperparameters.
7.2.4.2 Utilizing nested cross-validation for hyperparameter tuning For hyperparameter adjustment in machine learning, nested cross-validation is a strategy. Using different combinations of hyperparameter values for each iteration, a dataset is repeatedly separated into training and testing sets. The average of each iteration’s results is used to determine the ideal set of hyperparameters, which is then chosen [13]. This approach works well because it lowers the likelihood of overfitting and guarantees that the model is resilient to unobserved data. It also offers a method that can be used universally to determine the ideal set of hyperparameters for every given dataset.
7.2.5 Time series cross-validation When working with time series data, a cross-validation approach known as time series cross-validation is employed. The process is splitting the data into two or more separate, nonoverlapping subgroups, training a model on one of those subsets, and testing it on the other. The model can learn temporal information through this type of cross-validation, which also evaluates the model’s capacity for generalization across time. It is especially helpful when working with lengthy time series, as creating more subsets for traditional cross-validation methods like k-fold cross-validation can take a lot of time [14]. Additionally, time series cross-validation can be used to assess the impact of using a different window size for training or testing, as well as the impact of using various training and testing datasets.
7.2.5.1 Handling time-dependent data using time series cross-validation Time series cross-validation is a technique used to evaluate machine learning models on time-dependent data for further analysis. It works by dividing the time series into
7 Methods of cross-validation and bootstrapping
151
multiple training and test sets. The model is trained on the training sets and evaluated against the test sets. This method effectively prevents the model from overfitting to any single point in time, which might not represent other points in time. It also enables the model to generalize knowledge about the time series and be better prepared for future forecasts [15]. To evaluate the generalization of forecasting models, calibration metrics, and feature relevance, time series cross-validation can be used.
7.2.5.2 Rolling window and expanding window approaches Rolling window approach is a type of time series analysis used to estimate a value or calculate metrics based on a limited-size window. A rolling window comprises a set of consecutive time periods and the most recent data points are used to estimate a value or calculate a metric. This approach captures recent inputs and creates dynamic time frames for analysis. In contrast, the expanding window approach works with the entire data set instead of just the recent time period data [16]. This is a useful technique when the data set is composed mostly of historical observations and you are trying to map out logic or detect trends. The expanding window approach looks at increasing amounts of data, instead of just the most recent, to create a wider frame of reference. Additionally, this technique can capture changes in the underlying trend and identify shift points in the data.
7.3 Bootstrapping Using replacement sampling from the original sample, the approach known as “bootstrapping” can be used to estimate the sampling distribution of an estimator. This method is most frequently used to calculate standard errors and confidence intervals for population parameters. This method can do away with the requirement for difficult mathematical calculations that are required to derive the sampling distribution of an estimate. The procedure involves taking several samples from the original dataset with replacement and creating the sampling distribution by applying the estimator of interest to each sample. This distribution can provide insight into the variability of the estimator [17]. Bootstrapping can be used for a variety of different population parameters such as means, medians, variances, correlations, and regression coefficients. Additionally, it can also be used to test hypotheses, compare differences between groups, and construct confidence intervals.
152
K. Arthi et al.
7.3.1 Understanding bootstrapping as a resampling method A resampling technique called “bootstrapping” is used to calculate the sampling distribution of an estimator. By replacing random observations from the original set, it is a technique for creating new samples from an existing sample. We generate new sets of data through resampling and replacement, which are subsequently used to compute the relevant statistics. To calculate the average or mean of the estimated statistic and to get a sense of the variation that might exist across different samples, this process is performed numerous times. The estimation of variances, confidence intervals, and other statistical features that are challenging to determine analytically can be done via bootstrapping. It can also be used to test hypotheses and make predictions in cases where the assumptions of traditional statistics are violated [18]. Bootstrapping has become a popular tool for many statistical procedures due to its relative ease of implementation in many different analysis packages.
7.3.2 Bootstrap sampling process and creation of bootstrap samples Bootstrap sampling is a method used in statistics to create bootstrap samples which are resampled data from a larger set. The parameters of the population are estimated using a tiny sample. In bootstrapping, data are resampled using nonparametric techniques to get estimates based on the information that is now available. By replacing a sample from the original data set, the procedure operates. In a single bootstrap sample, a single observation can thus be drawn several times [19]. This process is repeated multiple times, creating a new data set or bootstrap sample for each iteration. Multiple bootstrap samples are used to create a population distribution of the parameters of interest from the original data set. These bootstrap samples can then be used to provide estimates for the statistics that are typically computed on the original data set, such as mean, median, variance, and others. Bootstrap sampling is a useful statistical tool since it is a strong and reliable way to get estimates from a small sample size.
7.3.3 Application of bootstrapping in model evaluation With replacement, a number of random samples are picked from a larger population as part of the model evaluation approach known as “bootstrapping.” By analyzing the variability of a predictive model’s performance across resampled datasets, it is possible to evaluate a predictive model’s stability and reliability. Bootstrapping can be used to assess data, estimate uncertainty, and validate models. It is also used to calculate the standard deviation of a statistical model’s parameters [20]. Bootstrapping is often used in conjunction with cross-validation to provide good estimates on model
7 Methods of cross-validation and bootstrapping
153
performance. Bootstrapping is a helpful method for evaluating models since it reveals the cause of any performance inconsistencies. When used appropriately, it can reveal important details about the model’s structure and can be applied to find any potential flaws or solutions to make the model better.
7.3.4 Bootstrap aggregating (bagging) and its benefits By combining the predictions of several weak models, such as decision trees, bootstrap aggregating, also known as bagging, is a machine learning approach used to increase the generalization accuracy of a given model. Rather than use the same set of data to fit each tree, the technique uses a bootstrapping technique where randomly chosen samples are used for training each tree. The benefit of bootstrap aggregating over a single decision tree model is that it reduces the variance and increases the predictive accuracy by taking the average of multiple trees. Bagging also helps reduce the risk of overfitting by creating an ensemble of random forests that ultimately creates a more stable prediction. Additionally, it can decrease the time needed to build a model, as individual trees can be built in parallel.
7.3.5 Out-of-bag (OOB) estimation in bagging A bagging model, a type of ensemble learning methodology, can be evaluated for accuracy using out-of-bag (OOB) estimates. OOB estimate is founded on a theory known as bootstrap aggregation, or “bagging” for short. The goal of bagging is to build numerous iterations of a single model, each using a different set of training data, and then aggregate the output of each model to produce a more reliable model. By examining the OOB samples, OOB estimation measures the correctness of each individual model [21]. The OOB samples are those portions of the training dataset that were not used to train the model. The correctness of OOB samples for each model is monitored. The average of the accuracy of each individual model makes up the bagging model’s overall accuracy. Any bagging model, including random forests, support vector machines, and neural networks, can be evaluated for accuracy using OOB estimation.
7.4 Comparison of cross-validation and bootstrapping Predictive models are trained and evaluated using a variety of techniques, including cross-validation and bootstrapping. The process of producing a cross-validation set from a given data set is an example of using cross-validation as a model evaluation
154
K. Arthi et al.
technique to see how well a prediction model’s findings generalize to an additional data set. Through this procedure, a portion of the data is put aside for testing and used to judge the model’s capability for making predictions. The data set is divided into k subsamples, and each sample is examined separately during this procedure known as “k-fold” cross-validation. The model’s capability for prediction is then evaluated in light of the result. However, bootstrapping is a type of model selection method that can be used to identify the model that performs the best on a given collection of data [22]. Bootstrapping employs the complete set of data, in contrast to cross-validation, which evaluates each model using only a portion of the data. In other words, the remaining data are utilized for the model’s validation while a subset of the data is randomly selected for the model’s training. The optimum model is then determined by how well it performs on the sampled data. Cross-validation is used to evaluate how well a predictive model performs on an independent data set, in contrast to bootstrapping, which is used to select the best model for a specific data set.
7.4.1 Contrasting cross-validation and bootstrapping The statistical analysis technique of cross-validation is employed to assess the accuracy of a given algorithm for a certain dataset. It divides the dataset into two or more subsets and evaluates the model’s performance on each set of data. In order to finetune a model’s hyperparameters, cross-validation is generally used to assess a model’s capacity for generalization. Bootstrapping is a statistical technique that resamples a dataset in order to estimate the properties of a population. It is commonly used to perform statistical inference on a dataset without making any assumptions about its underlying distribution. In order to determine a statistic’s precision or an estimated parameter’s accuracy, bootstrapping is also performed. Bootstrapping and cross-validation are two distinct methods used for various objectives. A model’s prediction performance is assessed using cross-validation and a population’s characteristics are estimated using bootstrapping.
7.4.2 Differences in underlying principles and applications The main difference between the two terms is in their underlying principles. The term analytics has traditionally referred to the use of data to analyze trends and to inform decision-making. It encompasses the use of methods such as statistics, data mining, machine learning, and predictive analysis to understand, describe, and predict user behavior. It is typically used to support decision making in areas such as product marketing and development.
7 Methods of cross-validation and bootstrapping
155
The goal of artificial intelligence (AI), on the other hand, is to make computers intelligent by developing systems that can learn and solve problems. It uses techniques such as neural networks, natural language processing, and Bayesian networks to create systems that are able to reason and react in complex environments. AI is focused on replicating certain cognitive abilities, enabling the development of more intelligent computer systems, which can engage in tasks that would require higher level cognitive skills from a human. AI is seen in applications such as self-driving cars, robotics, and facial recognition [23]. The analytics is focused on the use of data to analyze trends and inform decisionmaking, while AI utilizes techniques to create intelligent computer systems to solve complex problems.
7.4.3 Choosing between cross-validation and bootstrapping based on data characteristics Cross-validation and bootstrapping are both methods of assessing the accuracy of predictive models. The primary difference lies in the way in which sample data is divided and used for the purpose of estimation. Cross-validation separates the data into n equal datasets known as folds, as opposed to bootstrapping, which divides the data into several smaller datasets using random sampling with replacement. The type of data and the features of the problem determine which approach should be used. Cross-validation is preferable when the training dataset is large enough and there is enough data to divide into multiple folds for testing. Bootstrapping is more appropriate for smaller datasets, where the data points are not as homogeneous as in larger datasets. Bootstrapping also offers more flexibility in how samples are selected, allowing for the inclusion of outliers or rare values. Both cross-validation and bootstrapping are effective techniques for assessing the accuracy of predictive models. The selection of one technique over the other should be based on the characteristics of the data and the nature of the problem.
7.5 Advanced techniques and variations Advanced techniques and variations are methods of problem-solving that can be used to further develop and refine the solution to a given problem. These techniques and variations can be used to enhance the problem-solving process, making it easier or more effective, and ultimately improving the overall solution [24]. Complex techniques and their adaptations include heuristics, linear programming, dynamic programming, genetic algorithms, simulated annealing, and others. Each of these methods has a distinct set of steps and algorithms for use in the solution process, as well as a unique set of benefits and drawbacks.
156
K. Arthi et al.
7.5.1 Repeated cross-validation The output of a machine learning model can be verified via repeated cross-validation (RCV). The training set and test set are created by randomly splitting the dataset. The model is then trained on the training set using the specified algorithm. The test set is then used to evaluate the trained model’s performance after splitting the dataset at random, training, and evaluating the model, and repeating the process many times with different random splitting values. This enables us to accurately assess the performance of the model by taking into account the variability due to different splits. RCV is commonly used to tune the hyperparameters of a model, such as regularization in logistic regression or learning rate in neural networks.
7.5.1.1 Explanation and benefits of repeated cross-validation RCV is a statistical method for validly assessing predictive models. Iteratively, a given dataset is divided into training and test sets at random in this method. The predictive model is then tested on the test dataset after being tested on the training dataset. Then, the procedure is repeated a predetermined number of times with a different division of the test and training data each time. The performance of the model is then calculated by averaging the results of all iterations. Since it accounts for the variability in the training and test data used to evaluate the model, this method’s key advantage is that it enables estimates of the model’s performance to be more precise. Additionally, by using this method, any faults with the modeling procedure itself, such as overfitting difficulties, can be found.
7.5.2 Stratified group k-fold cross-validation The basic k-fold technique is extended to account for the relative class distribution of the data, and this is known as stratified group k-fold cross-validation. The data is divided into k groups for this type of validation, and each group is stratified to make sure that the distribution of the various classes within each group is the same as it was in the original data set. As a result, the validation findings are more accurate representations of the complete data set. When performing stratified group k-fold crossvalidation, for instance, each group should have 30% class A and 70% class B for a classification problem where the data set is composed of 30% class A and 70% class B.
7.5.2.1 Handling grouped data with stratified group k-fold cross-validation Stratified group k-fold cross-validation is a technique used to evaluate a model’s predictive performance on a dataset. This technique splits a dataset into a number of
7 Methods of cross-validation and bootstrapping
157
folds each containing an equal amount of data from all groups (or strata) in the dataset. In this technique, the data remains grouped together while being randomly partitioned into the folds. This is useful when the dataset contains related data that should not be separated, such as when working with grouped data containing different age ranges or different ethnicities [25]. The stratified approach ensures that each fold contains balanced amounts of different groups from the dataset, in order to allow model evaluation to be as accurate as possible. After all the folds have been generated, each fold can be used in a separate training and validation process, with the average performance across all folds being used as the final evaluation result.
7.5.3 Monte Carlo cross-validation The Monte Carlo cross-validation (MCCV) technique uses a dataset that has been randomly divided into training and test sets, with the training set being used to fit a model and the test set being used to assess the model’s accuracy. This type of crossvalidation is referred to as a “Monte Carlo” method because it randomly samples from the data before partitioning it. This allows the modeler to use all available data when constructing the model, without needing to split it beforehand. Then, repeatedly, subsets of the data are randomly sampled and a model is trained on one subset and evaluated on another. The average performance over all the iterations is then used as the final accuracy of the model.
7.5.3.1 Utilizing random sampling with Monte Carlo cross-validation MCCV and random sampling are used to assess the correctness of a model. Using this technique, the dataset is randomly split into training and testing sets. The testing set is used to evaluate the model once it has been created using the training set. The training and testing sets are then divided into several subsets using MCCV [26]. Each of these subsets is then used in combination with the rest to evaluate the model. In this manner, the model’s performance is tested multiple times on different combinations of data, providing a more accurate assessment of its performance. Random sampling with MCCV is a powerful tool for assessing model accuracy, as it allows for an efficient assessment of both overfitting and underfitting. However, it is important to note that the results from MCCV should be taken with a grain of salt and should always be used alongside other assessment methods when evaluating the accuracy of a model.
158
K. Arthi et al.
7.5.4 Time series bootstrapping Time series bootstrapping is a method of resampling that uses the past data points of a time series to simulate new data points. This method is used by data scientists to explore the variance and predict future trends in the data. Bootstrap samples are generated from the series of past data points by selecting random data points with replacement and are then used in various statistical tests [27]. This method is especially useful when the data set is too small to provide accurate or reliable results. By generating a bootstrap sample from the original data, the variance and potential trends can be explored, and realistic predictions can be made.
7.5.4.1 Extending bootstrapping to time series data A method known as extending bootstrapping to time series data makes use of resampling to reveal patterns in the behavior of a time series while taking into consideration its complexity. By selecting replacement samples at random from a data set, the resampling approach known as “bootstrapping” can estimate the distribution of a statistic. It is frequently used to replicate the uncertainty surrounding a statistic, allowing us to draw conclusions about the population’s unknowable values. When applied to time series data, it creates simulated time series that reflect the underlying distribution of the actual time series to be tested [28]. It can be used to assess the reliability of an estimate, identify potential outliers, provide insights into the structure of the time series, and allow comparison between multiple time series. By simulating future values of a time series, it can also be used to predict future values and analyze their stability.
7.5.4.2 Block bootstrap and stationary bootstrap methods Block bootstrap and the stationary bootstrap are two methods of sampling data that can be used to create a larger set of data for analysis and prediction. Block bootstrap is a method of sampling data that breaks the original dataset into smaller blocks called bootstrap replicates. This allows for the creation of a larger sample from the original dataset. The advantage of this methodology is that the original dataset tends to be preserved in the blocks, since each block will contain some of the original $\mu$. variability. The stationary bootstrap is a sampling technique that preserves the structure of the original data. This is done by sampling random points in the data, and assigning them to a new sample set. When the underlying generation method of the data is sufficiently varied and applicable to a wide range of data, it is valuable. It permits the generation of numerous distinct samples from a single initial dataset while maintaining the original $mu$ variability.
7 Methods of cross-validation and bootstrapping
159
7.6 Practical considerations and best practices When there are several parameters that need to be modified, a type of model validation called cross-validation is used to estimate the fit of a model. It is a method that is frequently used to evaluate a prediction model’s effectiveness [29]. Due to this, it is crucial to follow the following guidelines and best practices while employing crossvalidation: 1. Picking the right validation technique: Cross-validation should be selected based on the data size, the number of parameters to be adjusted, and the type of learning algorithm used. 2. Bias in the validation: Since a single validation set is used to validate each model, there is risk of introducing bias due to the data being repeated multiple times in the validation set. To reduce this risk, different validation techniques like k-fold cross-validation should be used. 3. Using the proper metrics: Accurate findings need careful consideration of the measures used to assess the model’s performance. Depending on the nature of the issue and the data at hand, a measure should be selected. 4. Determining the objectives and thresholds for optimization: Cross-validation should be used to refine the model’s parameters. To find the best model, precise objectives and thresholds should be decided upon beforehand. 5. Using the best validation strategy: Nested cross-validation, as opposed to k-fold cross-validation, is favored for larger datasets since it allows for more accurate results. Furthermore, bias can be avoided by employing an equally distributed validation approach. The accuracy of a predictive model can be increased by using cross validation, which is a crucial component of model validation. We may ensure that the model is properly evaluated and that the best outcomes are produced by taking into account the practical concerns and best practices listed above.
7.6.1 Handling small datasets with limited samples Managing small datasets with few samples can be difficult. It is usually more challenging to work with small datasets with few samples than with larger datasets with many samples. This is because the smaller sample size does not provide as much information as the larger sample size, making it hard to accurately analyze or make valid predictions based on the limited data [30]. To make the best use of the limited information, techniques such as feature engineering, data augmentation, and crossvalidation can be used to generate more meaningful results. Additionally, sampling techniques such as bootstrapping can be used to create more robust datasets. Finally,
160
K. Arthi et al.
machine learning algorithms that are designed to work well with small datasets should be chosen in order to maximize the predictive power of the limited data.
7.6.2 Impact of cross-validation and bootstrapping on computational resources Bootstrapping and cross-validation are two statistical methods that can be used to assess possible prediction models. With a “leave-one-out” methodology, cross-validation assesses a model’s predictive performance. This means that each data point in the dataset is removed in turn, and the model’s effectiveness is assessed according to how well it predicts the removed data point. In bootstrapping, a subset of the data points is randomly selected from the data set to create an intermediate data set. This process is repeated multiple times until there is a reliable estimate of the accuracy of the model. Both cross-validation and bootstrapping can have a significant impact on computational resources because they require the model to be retrained and reevaluated each time. This means that the algorithm must continuously process data and work through multiple iterations of validation and evaluation. This can be a very timeconsuming process and can result in considerable resource expenditure. However, both techniques are important for ensuring that models are trained properly and accurately serve their purpose.
7.6.3 Choosing appropriate evaluation metrics for each method Selecting the appropriate evaluation metrics is a vital stage in any machine learning process. Depending on the problem, different evaluation measures might be selected. For example, if the task at hand is a classification task, some common metrics that can be used include accuracy, precision, recall, F1 score, and area under the curve. When a regression task is involved, the preferred metrics are frequently the mean absolute error, mean squared error, and R-squared values [31]. Additionally, depending on the difficulty of the task, certain metrics may be required, including sensitivity, specificity, confusion matrix, true positive, false positive, and false negative. The chosen metrics must accurately reflect the quality of the output of the system being built, regardless of how the challenge may affect them.
7.6.4 Proper reporting and interpretation of results The proper reporting and interpretation of results is a critical step in any research project. Proper reporting includes providing an accurate and complete description of the study, presenting the results in an appropriate format and fashion, and discussing
7 Methods of cross-validation and bootstrapping
161
the implications of the results. Proper interpretation requires an understanding of the underlying assumptions that led to the results and potential outcomes and outcomes not measured by the results. It also requires an analysis of the results in the context of the overall research question and other related research. Proper interpretation of results should not be biased and should take into account potential confounding factors that could influence the results. Additionally, any potential bias or conflicts of interests should be declared and discussed in the analysis of the results.
7.7 Limitations and future directions Limitations of cross-validation: – Why a single model must be fitted several times in cross-validation, which can be computationally expensive. Additionally, it is constrained by the size of the training data set; if it is too small, the folds that result may not accurately represent the entire population. Limitations of bootstrapping: – Bootstrapping is limited to the size of the existing data set. It is also subject to strong assumptions, such as independence and identical distribution, which can lead to inaccuracies when the assumptions are violated. Future directions of cross-validation and bootstrapping: In the future, further research could be done to develop methods to reduce the computational complexity of the applications. Pursuing more efficient ways to establish the best hyperparameter values and avoiding data leakage as well as accommodating the possible presence of data drifts, would be further improvements. In order to solve the bias-variance tradeoff and increase the precision of the model findings, relevant measures may also be used. Bootstrapping might be improved further in order to more accurately account for current data inconsistencies and resolve between-sample discrepancies.
7.7.1 Recognizing limitations and potential pitfalls of cross-validation and bootstrapping Cross-validation is a technique for assessing a model or procedure using various data subsets. It is an iterative process, where each iteration provides a validation measure for the model. Bootstrapping is an iterative method of creating samples with replacement from a population of observations. Although cross-validation and bootstrapping can be used to provide robust estimates for a model’s performance, they both have
162
K. Arthi et al.
some limitations and pitfalls. Cross-validation can be sensitive to the choice of sample splits, so extreme care must be taken to avoid sample splitting bias. If the process of splitting the data set is not carefully planned and executed, it is possible to end up with an easy-to-fit model, which will produce unrealistic results when applied on unseen data [32]. The bootstrapping approach can potentially be hampered by overestimating a method’s capacity for generalization to fresh data. Using various resampling methods, such as subsampling or leave-one-out cross-validation, can help to alleviate this. Additionally, some characteristics of the data used to produce the sample may have an impact on how the data are manipulated, and the manipulated data may not accurately represent the population from which the sample was collected. Despite the fact that cross-validation and bootstrapping can be helpful in assessing a model’s performance, it is crucial to be aware of their drawbacks and danger zones. When using these techniques, careful thought should be applied to guarantee appropriate findings.
7.7.2 Emerging research and advancements in resampling techniques The emergence of resampling techniques has been a major part of greater advancements in predictive modeling, machine learning, and research in general. Resampling techniques provide data scientists with the ability to effectively measure the performance of a given model, by evaluating the accuracy of the model on data which has been repeatedly sampled. Through these techniques, researchers are able to find out if a given model is too complex or too simple. This is done by changing data and feature selection and resampling to create new training sets. Additionally, these techniques can help debug models for unexplained results, giving researchers greater confidence in their models. Furthermore, by providing new ways of testing the performance of a model, resampling techniques can help researchers find areas of opportunity and potential improvements. Ultimately, these techniques can help accelerate the development and deployment of more reliable algorithms in data science research.
7.7.3 Future directions for improving model evaluation methods The goal of model evaluation methods is to ensure models are properly trained and evaluated on data that closely follows the real-world environment they will be deployed in. As machine learning models become more pervasive, the need for more advanced evaluation methods that can analyze more complex datasets and capture subtle characteristics of datasets is growing. In the future, these model evaluation methods may incorporate more techniques that use adversarial or generative modeling to better mimic data from the real world. Additionally, tools that are better able to understand and detect outliers, noisy data, and patterns in datasets may improve
7 Methods of cross-validation and bootstrapping
163
model evaluation methods. Additionally, researchers may develop meta-learning methods, in which models are trained on simulated data to gain insight into how well it can handle a variety of scenarios, data types, and environments. Furthermore, the ability to compare model performances in a hierarchical way and to visualize these performances using dimensionality reduction techniques will be a useful way of understanding model performance. The machine learning models must be evaluated in ways that account for the risk associated with deployment. Methods such as decision thresholds and counterfactual analysis allow modelers to better understand model performance and make informed decisions about when and how to deploy a given model. The future of model evaluation methods is expected to combine sophisticated simulation tools, powerful visualizations, and risk-based evaluation to provide insights into how models interact with real-world data.
References [1]
Brodeur, Z. P., Herman, J. D., & Steinschneider, S. (2020). Bootstrap aggregation and cross‐validation methods to reduce overfitting in reservoir control policy search. Water Resources Research, 56(8), e2020WR027184. [2] Rafało, M. (2022). Cross validation methods: Analysis based on diagnostics of thyroid cancer metastasis. ICT Express, 8(2), 183–188. [3] Ramesh, G., Logeshwaran, J., & Kumar, A. P. (2023). The smart network management automation algorithm for administration of reliable 5G communication networks. Wireless Communications and Mobile Computing, 2023, 1–12. [4] Yadav, S. P., & Yadav, S. (2020). Image fusion using hybrid methods in multimodality medical images. Medical & Biological Engineering & Computing, 58, 669–687. [5] Vrigazova, B. (2021). The proportion for splitting data into training and test set for the bootstrap in classification problems. Business Systems Research: International Journal of the Society for Advancing Innovation and Research in Economy, 12(1), 228–242. [6] Vrigazova, B., & Ivanov, I. (2020). Tenfold bootstrap procedure for support vector machines. Computer Science, 21, 241–257. [7] de Rooij, M., & Weeda, W. (2020). Cross-validation: A method every psychologist should know. Advances in Methods and Practices in Psychological Science, 3(2), 248–263. [8] Uyun, S., & Sulistyowati, E. (2020). Feature selection for multiple water quality status: Integrated bootstrapping and SMOTE approach in imbalance classes. International Journal of Electrical and Computer Engineering, 10(4), 4331. [9] Irfan, D., Tang, X., Narayan, V., Mall, P. K., Srivastava, S., & Saravanan, V. (2022). Prediction of quality food sale in mart using the AI-Based TOR method. Journal of Food Quality, 2022, 1–13. [10] Yadav, S. P., Zaidi, S., Mishra, A., & Yadav, V. (2022). Survey on machine learning in speech emotion recognition and vision systems using a recurrent neural network (RNN). Archives of Computational Methods in Engineering, 29(3), 1753–1770. [11] Yuvaraj, N., & Vivekanandan, P. (2013, February). An efficient SVM based tumor classification with symmetry non-negative matrix factorization using gene expression data. In 2013 International Conference on Information Communication and Embedded Systems (Icices) (pp. 761–768). IEEE.
164
[12]
[13]
[14] [15] [16]
[17]
[18] [19]
[20]
[21]
[22]
[23]
[24]
[25] [26]
[27]
[28]
[29]
K. Arthi et al.
Fernandez-Felix, B. M., García-Esquinas, E., Muriel, A., Royuela, A., & Zamora, J. (2021). Bootstrap internal validation command for predictive logistic regression models. The Stata Journal, 21(2), 498–509. Sivaram, M., Yuvaraj, D., Mohammed, A. S., Manikandan, V., Porkodi, V., & Yuvaraj, N. (2019). Improved enhanced DBTMA with contention-aware admission control to improve the network performance in MANETS. CMC-Computers Materials & Continua, 60(2), 435–454. Sangeetha, R., Logeshwaran, J., Rocher, J., & Lloret, J. (2023). An Improved Agro Deep Learning Model for Detection of Panama Wilts Disease in Banana Leaves. AgriEngineering, 5(2), 660–679. Vrigazova, B. P. (2020). Detection of malignant and benign breast cancer using the AnovaBootstrap-SVM. Journal of Data and Information Science, 5(2), 62–75. Larracy, R., Phinyomark, A., & Scheme, E. (2021, November). Machine learning model validation for early stage studies with small sample sizes. In 2021 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC) (pp. 2314–2319). IEEE. Takada, T., Nijman, S., Denaxas, S., Snell, K. I., Uijl, A., Nguyen, T. L., . . . & Debray, T. P. (2021). Internal-external cross-validation helped to evaluate the generalizability of prediction models in large clustered datasets. Journal of Clinical Epidemiology, 137, 83–91. Cho, G., Jung, K., & Hwang, H. (2019). Out-of-bag prediction error: A cross validation index for generalized structured component analysis. Multivariate Behavioral Research, 54(4), 505–513. Zhang, Y., Qiao, S., Ji, S., Han, N., Liu, D., & Zhou, J. (2019). Identification of DNA–protein binding sites by bootstrap multiple convolutional neural networks on sequence information. Engineering Applications of Artificial Intelligence, 79, 58–66. Harrou, F., Saidi, A., & Sun, Y. (2019). Wind power prediction using bootstrap aggregating trees approach to enabling sustainable wind power integration in a smart grid. Energy Conversion and Management, 201, 112077. Saravanan, V., & Sumathi, A. (2012). Handoff mobiles with low latency in heterogeneous networks for seamless mobility: A survey and future directions. European Journal of Scientific Research, 81(3), 417–424. Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). A novel architecture of intelligent decision model for efficient resource allocation in 5G broadband communication networks. ICTACT Journal On Soft Computing, 13(3), 2986–2994. Yadav, S. P., & Yadav, S. (2019). Mathematical implementation of fusion of medical images in continuous wavelet domain. Journal of Advanced Research in Dynamical and Control System, 10(10), 45–54. Singh, J., Deepika, J., Sathyendra Bhat, J., Kumararaja, V., Vikram, R., Jegathesh Amalraj, J., . . . & Sakthivel, S. (2022). Energy-efficient clustering and routing algorithm using hybrid fuzzy with grey wolf optimization in wireless sensor networks. Security and Communication Networks, 2022, 1–10. Kim, B., Xu, C., & Barber, R. (2020). Predictive inference is free with the jackknife+-after-bootstrap. Advances in Neural Information Processing Systems, 33, 4138–4149. Khan, S., Saravanan, V. N. G. C., Lakshmi, T. J., Deb, N., & Othman, N. A. (2022). Privacy protection of healthcare data over social networks using machine learning algorithms. Computational Intelligence and Neuroscience, 2022, 1–9. Natarajan, Y., Kannan, S., & Mohanty, S. N. (2021). Survey of various statistical numerical and machine learning ontological models on infectious disease ontology. Data Analytics in Bioinformatics: A Machine Learning Perspective, 13(1), 431–442. Yuvaraj, N., Srihari, K., Dhiman, G., Somasundaram, K., Sharma, A., Rajeskannan, S. M. G. S. M. A., . . . & Masud, M. (2021). Nature-inspired-based approach for automated cyberbullying classification on multimedia social networking. Mathematical Problems in Engineering, 2021, 1–12. Yousefa, W. A. (2020). A leisurely look at versions and variants of the cross validation estimator. stat, 1050, 9.
7 Methods of cross-validation and bootstrapping
165
[30] Ho, S. Y., Phua, K., Wong, L., & Goh, W. W. B. (2020). Extensions of the external validation for checking learned model interpretability and generalizability. Patterns, 1(8), 100129. [31] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). Energy‐efficient resource allocation model for device‐to‐device communication in 5G wireless personal area networks. International Journal of Communication Systems, 2023, 1–14. [32] Rani, P., Verma, S., Yadav, S. P., Rai, B. K., Naruka, M. S., & Kumar, D. (2022). Simulation of the lightweight blockchain technique based on privacy and security for healthcare data for the cloud system. International Journal of E-Health and Medical Communications (IJEHMC), 13(4), 1–15.
S. Ramasamy, Kantharaju H. C., N. Bindu Madhavi, Haripriya M. P.
8 Meta-learning through ensemble approach: bagging, boosting, and random forest strategies Abstract: Meta-learning through ensemble approaches is an intriguing subfield of machine learning research. With this method, a more comprehensive learning model is created by combining many machine learning methods, including neural networks and support vector machines. By using an ensemble of models, meta-learning techniques are able to produce more robust results than individual algorithms alone. In addition, ensemble techniques are advantageous because they can easily be expanded to accommodate additional data sources or algorithms. This approach can also embed more knowledge from the data into a more powerful meta-model, which allows the system to generalize better and discover patterns more accurately. In short, metalearning through ensemble approaches is an effective and useful technique for tackling challenging problems in machine learning. Keywords: Meta-learning, ensemble approach, machine learning, neural networks, support vector machine
8.1 Introduction to meta-learning Meta-learning is a branch of machine learning where the goal is to improve learning algorithms by introducing the model of learning to learn. It refers to the process of learning how to learn. Meta-learning algorithms are typically used to learn new skills quickly from a few examples, or even to modify existing skills to adapt to new tasks [1]. Examples of applications of meta-learning include robots learning to perform various tasks, autocompletion of programming tasks, medical diagnosis, and self-driving cars. Meta-learning has both research and practical applications, allowing us to more quickly create machine learning models for new tasks.
8.1.1 Definition and importance of meta-learning in machine learning Machine learning research in the field of “meta-learning” focuses on creating algorithms that can successfully learn from their past experiences and knowledge. This contrasts with conventional machine learning algorithms, which only use the training data provided. By leveraging shared inductive biases observed in comparable tasks, https://doi.org/10.1515/9783111323749-008
168
S. Ramasamy et al.
meta-learning algorithms can swiftly and effectively learn how to adapt to problems that have never been seen before. Meta-learning’s objective is to develop models that are easily generalizable and adaptable to data from other areas [2]. Meta-learning is an important tool for data scientists and researchers as it increases the efficiency and effectiveness of the whole machine learning process. Instead of having to spend time training a model from scratch each time a task needs to be accomplished, the meta-learning model can be quickly configured to the new task using the skills it has acquired from prior tasks. Recently, meta-learning has been used successfully in fields like reinforcement learning, natural language processing, and computer vision. Machine learning’s expanding topic of meta-learning has a wide range of intriguing applications [3].
8.2 Ensemble learning A machine learning technique called ensemble learning combines many learners to enhance predicted performance. It employs more learning algorithms than any of the individual learning algorithms in order to get superior predicted performance. Both classification and regression problems use ensemble methods. The basic concept behind ensemble methods is to combine multiple weak or average performing learners to create a stronger learner. The principle behind ensemble learning is that weak learners, when combined, can result in a strong learner.
8.2.1 Introduction to ensemble learning and its benefits A form of machine learning technique called ensemble learning combines several learning algorithms to produce models that are more accurate than those produced by any one of the individual algorithms alone. Different types of algorithms, such as decision trees, support vector machines, or neural networks, can be used in ensemble learning. By mixing predictions from many models, ensemble learning aims to decrease variance and bias, and enhance prediction accuracy. Benefits of ensemble learning include improved performance, increased stability, better generalizability, improved scalability, reduced risk of overfitting, and improved accuracy. Ensemble learning improves generalizability by taking advantage of multiple sources of information and aggregating them in a single model. It also reduces the risk of overfitting by combining different models, allowing for better generalization to unseen data [4]. Additionally, ensemble learning improves scalability by training multiple models in parallel, allowing more data to be processed faster. Finally, ensemble learning significantly improves the accuracy of models by combining multiple models with different characteristics.
8 Meta-learning through ensemble approach
169
8.2.2 Combining multiple models for improved performance and generalization Combining multiple models for improved performance and generalization is a powerful method in machine learning that can be used to boost overall performance. By combining models, the weaknesses of one individual model can be mitigated and improved by the strengths of other models. As a result, the system can generalize in new ways, which eventually helps the combined model perform better. For instance, a combination model may combine various models with various methods into a single ensemble model or it may take the average of numerous distinct models. It is possible to find patterns among models by merging several models that might be too complex for an individual model to find [5]. This implies that the ensemble model’s performance may be superior to that of the individual models and that the model may be better able to generalize trends and spot novel data patterns. By utilizing this technique, it is possible to achieve faster, more accurate predictions and better overall performance.
8.2.3 Different types of ensemble learning techniques A sort of machine learning technique, known as ensemble learning, combines the predictions of several independent models to produce a single, improved model. This method is applied to increase the model’s accuracy and robustness. There are five standard ensemble learning methods: – Boosting: Boosting builds up a strong classifier gradually, based on the weak classifiers or the simple base classifiers. It works by giving more weight to the wrongly classified examples so that those examples are given more attention in the next round of training. – Bagging: A sort of ensemble learning technique, bagging creates a single prediction by averaging the model outputs. It functions by using the same algorithm to train numerous models, which are then combined to produce a forecast. – Stacking: As an ensemble technique, stacking combines the outcomes of various models into a single outcome. It functions by using the output of one model as input for another model. – Blending: Blending functions by averaging the results of various models’ forecasts. It is frequently used to lessen the variation in a model’s predictions. – Voting: Voting is an ensemble learning technique that integrates the forecasts from various models. Every model makes a prediction, and the final prediction is determined by taking the vote of the majority of all the predictions.
170
S. Ramasamy et al.
8.3 Bagging Bagging is a type of ensemble method in which the same classification algorithm is used repeatedly on different subsets of the training data. This is an approach used to increase the accuracy of a predictive model by reducing the variance in its predictions. Bagging helps improve accuracy by combining the results of different classifiers and reducing their variance [6]. This process averages out the errors of individual classifiers and thus reduces the overall effects of noise. Bagging can be used with any kind of predictive model, although it is most frequently used with decision trees since it helps lessen the instability of a single decision tree.
8.3.1 Explanation of bagging as an ensemble learning method Bagging is an ensemble learning technique used to develop predictive models that, when compared to a single model, are less prone to overfitting. A variety of models are trained using various subsets of the input data in this ensemble technique, and the results are then pooled to produce a more precise forecast. This technique works because it averages out the effects of each of the models, reducing the chances of having an overly complex model [7]. Additionally, it reduces the variance by combining a range of different models. This leads to more stable predictions that are less prone to overfitting.
8.3.2 Bootstrap aggregating (bagging) algorithm The process known as bootstrap aggregating, often referred to as bagging, reduces the variance of individual estimates by merging them, producing a more accurate aggregate estimate. This is done by creating multiple bootstrapped versions of the original data set and then combining the different estimates to create an aggregated value. Bagging can also help control overfitting in machine learning applications. The bagging algorithm works by randomly selecting from the available data points and then combining the independent estimations that are returned. This reduces the variance of the overall estimate by averaging out the individual estimates.
8.3.3 Bagging implementation and benefits A machine learning ensemble approach called bagging or bootstrap aggregating makes use of numerous models to increase the precision of predictions generated by a single model. A subset of the dataset is used to train a model and predictions are then made using the model [8]. The outcomes of the several models are then inte-
8 Meta-learning through ensemble approach
171
grated, typically using an average. This generally results in more accurate forecasts than a single model, and helps to lower the variance of the predictions. Bagging helps improve the learning algorithm’s performance in a number of ways. Firstly, it increases the generalization of the model, meaning that the model can more accurately predict values for unseen data based on the data it has learned from. Secondly, it helps reduce overfitting, since it shows the model different training sets and therefore leads to more robust models [9]. Finally, it defends against data distortion, since the different training sets randomized to be created during the bagging process make it more difficult for any given batch of data to overly skew the model’s predictions.
8.4 Boosting Boosting is a type of machine learning algorithm that iteratively learns from data, allowing models to remain flexible and produce more accurate results. It does this by combining a set of weak learners (e.g., shallow decision trees) and creating a strong learner. Weak learners are instructed in stages, with each one attempting to correct the one before it. The final prediction is then created by adding up each individual’s predictions using a weighted majority vote (or sum).
8.4.1 Introduction to boosting as an ensemble learning technique Boosting is an ensemble learning strategy where several weak learners are combined to form one strong learner. It operates by building a progressively additive model, where each new model is built using the series of residuals from the prior one. The boosting process consists of adding each successive model to the ensemble to reduce the overall error of the final model [10]. Each successive model is an attempt to correct the errors of the previous model. The objective is to produce a strong learner that will improve the predictive performance on unseen data. Boosting algorithms have the advantage of not making strong assumptions about the data shape, and being able to absorb bias from previous models.
8.4.2 Adaboost algorithm and its working principles An ensemble learning technique called AdaBoost, short for Adaptive Boosting, is used to create predictive models. It operates by integrating several weak learners into one strong learner and is trained iteratively. Then, predictions are made by using this powerful learner. It functions by initially creating a weak classifier using a subset of
172
S. Ramasamy et al.
data. This classifier is employed to predict outcomes based on the provided data [11]. If it makes mistakes, the algorithm increases the weight of that instance and tries to create a better classifier. The process is repeated until a strong learner is formed. The AdaBoost algorithm is often used in conjunction model with decision trees, which provide the weak learners used by the algorithm. It is a popular choice of algorithm due to its simplicity, effectiveness, and robustness.
8.4.3 Gradient boosting algorithms Gradient boosting is a machine learning technique that assembles various decision trees into an ensemble to create the best possible model. It operates by sequentially training weak prediction models, each one aiming to fix the mistakes caused by the preceding one. It is a form of boosting, and is an iterative process in which the weak models are added one at a time until a satisfactory result is achieved. The basic idea behind gradient boosting is to create multiple weak learning algorithms – known as “base learners” – and then combine their predictions into a single, strong predictive model. The weak learners are decision trees of varying depths and complexities, which are trained and used to predict the target variable [12]. The predictive models are then combined by giving each tree its own “weight” or coefficient, which is determined by how much that tree contributed to the overall prediction. A potent, predictive machine learning model is the end product. Applications for gradient boosting include computer vision, natural language processing, web search ranking, and many more.
8.4.4 Boosting implementation and advantages In order to generate an ensemble strong learner model, the boosting machine learning method combines a number of weak learners, each of which concentrates on identifying a certain feature of the dataset. It works by training the weak learners sequentially, each time refocusing on the samples that were misclassified by the previous learner. At the end of the process, the ensemble is provided with a majority vote on the class label assigned to a given sample [13]. Many sophisticated learning algorithms, including XGBoost, are built using this technique of integrating multiple weak predictive models to produce a powerful statistical model. Advantages of boosting – It is a sequential learning technique that uses previous versions of the model to improve upon itself.
8 Meta-learning through ensemble approach
– – – – –
173
Boosting algorithms work particularly well with noisy data due to the use of weak learners. They are able to handle a large variety of data types. They are relatively insensitive to model parameter specifications. They allow for the definition of complex relationships between variables. Boosting algorithms come with fewer parameters to tune compared to deep learning algorithms.
8.4.5 Real-world applications of boosting Object detection, anomaly detection, online advertising, credit scoring, fraud detection, natural language processing (NLP), and recommendation systems are just a few of the real-world applications that make use of boosting techniques. Boosting algorithms are effective techniques for enhancing the dependability and accuracy of machine learning models. They are used in many applications to find patterns in data and generate predictions that are more accurate and reliable than those produced by conventional techniques. Boosting algorithms can also be used to reduce bias and variance in a model’s predictions. For example, they may be used to identify correlations in data that would otherwise be overlooked in traditional models. Boosting algorithms can also be used to improve accuracy when classifying data and reduce the likelihood of the occurrences of false positives or false negatives [14]. The ability to identify patterns in data is especially useful for medical diagnosis, fraud detection, and anomaly detection. Additionally, boosting techniques are employed in NLP applications to enhance the precision and speed of word, phrase, and syntax recognition. Finally, online advertising and recommendation systems use boosting algorithms to better target adverts and items to customers, depending on their preferences.
8.5 Random forests A supervised learning approach, known as a “random forest,” is used in machine learning to solve classification and regression issues. They are founded on decision trees, which draw conclusions from data using a decision tree. As an ensemble learning technique, random forests produce predictions using a variety of decision trees. Each tree in the forest is made up of a random assortment of data points and the replies that go with them, collectively known as data bags. Random forests are able to provide predictions that are superior to those produced by any single tree because they combine the forecasts from each of these individual trees. Random forests also help to reduce the variance when compared to single decision trees. Random forests
174
S. Ramasamy et al.
are an effective technique for many applications, such as customer segmentation, predictions, and fraud detection.
8.5.1 Overview of random forests as an ensemble learning approach Multiple decision trees are randomly trained on subsets of the data using the ensemble learning technique known as random forests, and the votes from each individual tree are then combined to provide the final forecast. This approach reduces variance from an individual tree by introducing a diversity of models that have no correlation with each other. The level of randomness is adjustable and the process is iterative, so the algorithm can hone in on the most predictive variables for a data set [15]. Random forests provide a tremendous boost in accuracy compared to a single decision tree, and are widely used in areas such as recommendation systems and computer vision.
8.5.2 Decision tree construction and aggregation in random forests Decision tree construction in random forests is the process of building individual decision trees that make up the forest. This process begins by randomly selecting observations from the data set to become the “root node” or “decision node.” For each node, the best split point is determined by algorithm and splitting criteria, such as information gain or a Gini index. The next step is to divide the training data into two new child-nodes, each representing a subset of the data. This process continues until each node is pure (all observations belong to the same class) or until all nodes have the same values [16]. Lastly, the algorithm builds the optimal model by pruning away nodes that do not significantly add to the accuracy of the model, and weighing the nodes based on their importance. In Random forests, the process of aggregation entails integrating the forecasts of the various decision trees that comprise the forest. This is done by averaging the predictions made by the various trees. A consensus prediction is produced as a result of this averaging, and it is thought to be more accurate and less prone to error than predictions produced by a single tree.
8.5.3 Random forests algorithm and its variations The random forests ensemble learning method builds a large number of decision trees during the training phase and outputs the class that represents the mean of the classes (classification) or mean prediction (regression) of the individual trees. Random
8 Meta-learning through ensemble approach
175
forests can be used for classification, regression, and other tasks. Due to their capacity to restrict overfitting of a single decision tree and their efficiency in minimizing variance in the predictions, random forests have been demonstrated to be a highly accurate and robust method for a range of situations [17]. Extra trees (sometimes referred to as extremely randomized trees), which randomly choose threshold and feature values, and sparse random forests, which are intended for high-dimensional datasets, are variations of the random forests algorithm. Other variations include tailored random forests for specific tasks, such as random forests for image classification and random forests for anomaly detection.
8.5.4 Advantages and limitations of random forests Advantages of random forests – Random forests is a nonparametric learning method that works well with large datasets. – Random forests can handle both categorical and numerical values and work well even without feature scaling. – Random forests provide good accuracy and generalize well. They are resistant to overfitting. – Random forests perform feature selection and help in reducing computational cost. – Random forests are robust to noise and outliers. Limitations of random forests – Model interpretability is quite poor in random forests. It is difficult to interpret and explain why a certain prediction is made. – Random forests can be sensitive to hyperparameter tuning. This means that if the model is not properly tuned, it may not perform well. – Random forests can be slower to train than other algorithms, especially for large datasets. – Random forests may not be the best choice for regression problems, as they are designed for classification.
8.6 Ensemble strategies and techniques In order to enhance the performance of a predictive model relative to the performance of individual models, ensemble tactics and techniques integrate numerous diverse models or algorithms. These strategies are used in machine learning algorithms and are popularly known as ensemble methods [18]. By combining diverse set of
176
S. Ramasamy et al.
learners, ensemble methods can provide better performance and a better generalization accuracy than a single model. The use of ensemble methods is applicable to both classification and regression issues. Ensemble approaches frequently employ bagging, boosting, stacking, and other strategies.
8.6.1 Voting ensembles A voting ensemble is a form of machine learning technique that combines predictions from various models, including neural networks, support vector machines, and decision trees, to make predictions that are more accurate than those that could be made using any one of the individual models alone. Each input is given a majority vote from all the models and the most popular response is then output as the final result. This technique has proven to be effective in reducing the overall error rate from individual models.
8.6.2 Stacking and blending ensemble techniques Machine learning employs the ensemble techniques of stacking and blending. Stacking is an ensemble method that trains numerous base models on various subsets of a single dataset. A single forecast for the dataset is then created by combining the predictions of each of these foundation models. Blending is an ensemble method that combines estimates from various base models and trains a separate model to generate the final estimate [19]. The weights used to combine the predictions from the basis models are determined by how well the base models performed. Both stacking and blending techniques can improve the accuracy of predictions as they combine the power of multiple different models.
8.6.3 Ensemble pruning and model selection methods Ensemble pruning and model selection methods are techniques for selecting the best possible model from among many different choices. The goal of pruning and model selection is to either improve a model’s accuracy or to reduce the complexity of the model, ensuring it is computationally tractable and efficient. Pruning involves removing irrelevant or noisy parameters from a model [20]. Model selection methods attempt to select the best set of hyperparameters for a specific problem. Ensembles methods involve combining multiple models that are connected together by various techniques. These techniques can include methods such as stacking, averaging, sandboxing, boosting, and bagging. Ensemble pruning involves selecting the subset of models that will provide the best performance. Model selection techni-
8 Meta-learning through ensemble approach
177
ques are used to select the most accurate model, given an input dataset. This can be achieved by assessing accuracy scores, feature importance, and other metrics on the trained model.
8.7 Meta-learning and model selection A type of machine learning called meta-learning is used to increase the capacity of the current models to pick up new skills. It uses existing models to provide information about how each model performs on a specific task. The performance of the several models can then be compared using meta-learning to determine which is best for a specific task. This helps to reduce the workload in selecting models for new tasks, making it easier to find the most suitable model quickly and efficiently. The process of choosing the best model from a group of competing models is known as model selection. This decision is made based on the model’s predictive accuracy, computational complexity, and prediction time. It is a crucial stage in the machine learning process that aids in choosing the model that is most appropriate for a given task. Model selection can be done manually or using automated algorithms such as cross-validation, bootstrapping, or model selection methods such as sequential model selection [21]. In meta-learning, model selection is automatically done using the data collected from the previous experiments. This helps to choose a model quickly and accurately without manual effort.
8.7.1 Leveraging ensemble learning for meta-learning tasks Ensemble learning is a powerful method for meta-learning tasks. Its goal is to combine multiple models to create a consistent prediction. It works by aggregating multiple weak learners to form a more powerful model. The general objective of ensemble learning in meta-learning tasks is to combine multiple models into a single, stronger model. The idea is to leverage the strengths of the individual models and to increase the accuracy, while minimizing the risk of overfitting. This can be done by incorporating different techniques such as boosting, bagging or stacking, or combining different types of algorithms, such as decision trees or neural networks [22]. Ensemble learning can be used in meta-learning tasks to improve the accuracy of prediction. It can be used to build comprehensive models with better performance, which can be accessed more quickly. This makes it especially useful for time-critical tasks or tasks with high volumes of data. Moreover, its robustness and ability to handle complex data sets make it a great tool for meta-learning tasks.
178
S. Ramasamy et al.
8.7.2 Using ensembles for model selection and hyperparameter optimization Ensemble methods are a type of model selection and hyperparameter optimization technique that combines different models and algorithms into a single model. Using an ensemble method, individual models and algorithms are used in combination to form a “stronger” model. Such models are usually more accurate or better predictors than single models. Due to its capacity to combine numerous models and algorithms, ensemble methods have been employed in a range of applications, including computer vision, natural language processing, and speech recognition. In model selection and hyperparameter optimization, ensemble methods are used to identify the best combination of models, algorithms, and hyperparameters for a given performance measure. Typically, a search strategy is used to choose the optimal model combination for a particular measure. To find the ideal set of parameters for a particular model, ensemble approaches can also be utilized [23]. To make superior predictions, ensemble methods frequently combine several algorithms, including boosting, bagging, and stacking. A sequence of weak learners are merged to generate a strong model in a process known as “boosting,” a sort of model combination. Bagging is a form of ensemble method that combines algorithms to lower the variation in the forecasts made by the different models. Stacking is where multiple models are combined to create a single model. All three of these methods are commonly used in ensemble methods for model selection and hyperparameter optimization.
8.7.3 Cross-validation and ensemble-based model evaluation Without the requirement for a second validation set or for dividing the initial dataset into training and test sets, the supervised learning technique of cross-validation enables model evaluation and selection. Cross-validation is a technique that involves randomly dividing the data into training and test sets; using the training set to build and modify a model, and then using the test set to assess the model’s performance [24]. Cross-validation can be used to identify the best parameters and hyperparameters for a given model, or to compare different algorithms. Ensemble-based model evaluation is a technique for model evaluation in which multiple models are trained and tested with different subsets of data. Each model is then compared to a baseline model, and the performance of the ensemble is determined by its accuracy and precision on the held-out dataset. Ensembling techniques can be used to combine prediction results from a variety of models, such as machine learning algorithms, to improve predictive performance.
8 Meta-learning through ensemble approach
179
8.8 Comparison and trade-offs Comparison and trade-offs is a way of looking at two or more options and determining which one is better, based on your own criteria. It involves examining the advantages and disadvantages of each option and then making an informed decision based on the data. This process can be used in many different areas, including business decisions, personal choices, and technology choices. It involves weighing the costs and benefits of each option and comparing them to find the best one. It is frequently crucial to take into account each decision’s possible risks and rewards as well as its short- and long-term implications. Ultimately, comparison and trade-offs help to ensure that the best possible decision is made.
8.8.1 Comparing bagging, boosting, and random forests strategies Bootstrap, or bagging AGGregatING is a method for reducing variance in predictions made by a model, thus increasing the predictive accuracy of machine learning models. It works by creating numerous models, each of which is trained on a bootstrapped dataset, and randomly picking or bootstrapping the training dataset. The final prediction is then calculated by averaging the combined data. An ensemble technique called “boosting” is used to raise the precision of machine learning models. A set of base learners (weak learners) are constructed and then “boosted” or “added to” one another to form a strong learner. Boosting works by starting with a random prediction and then iteratively improving the prediction by making small corrections. This process is repeated until the accuracy of predictions is improved. Random forests are a combination of bagging and boosting. A base model is constructed using the bagging technique and then the model is “boosted” or improved using the random decision trees method. The improved data set is then used to construct a stronger model. Decision trees are developed into random forests, which employ various randomization techniques to increase the model’s accuracy.
8.8.2 Differences in their underlying principles and performance characteristics The underlying principles of a relational database and a NoSQL database are quite different. A relational database is built upon the principles of relational-algebra, which is a mathematical model for representing and manipulating data in a relational way. This is why relational databases are often referred to as SQL, short for struc-
180
S. Ramasamy et al.
tured query language. Relational databases are organized into tables with rows and columns for optimal data storage and retrieval speeds. NoSQL databases are not built upon relational-algebra as their underlying principles. Instead, they are built upon the concepts of key-value stores, document databases, and graph databases [25]. Compared to conventional relational databases, these NoSQL databases offer a more flexible way to store and retrieve data. This adaptability may result in quicker data storage and retrieval times. The performance characteristics of both types of databases also vary significantly. Relational databases are generally more robust when it comes to handling large amounts of data. They are better suited for storing complex data structures and are more resilient when it comes to scaling. On the other hand, NoSQL databases are more suitable for handling unstructured data and are better at retrieving data quickly.
8.8.3 Choosing the appropriate ensemble method for different scenarios A type of learning algorithms, known as the ensemble technique, mixes different machine learning algorithms to increase the output’s accuracy and stability. An adequate ensemble strategy can considerably increase the model’s accuracy when the dataset is sizable and varied [26]. Random forest, boosting, bagging, and stacking are some of the ensemble methods in machine learning that are most frequently utilized. An ensemble technique called random forest uses random subsets of features and decision trees to make predictions. It is useful for both classification and regression problems, and is straightforward to use. Boosting is an iterative ensemble algorithm in which each successive round of predictions is built upon the previous predictions. The algorithm is beneficial when aiming for accuracy in predictions, as it can account for both bias and variance in a model. Bagging is a type of ensemble method used when the dataset is homogeneous and diverse. The method combines multiple models while maintaining independence among them to produce a more accurate result. Stacking is an ensemble technique that combines predictive techniques, such as bagging and boosting, to produce more accurate results. It is useful when the dataset is heterogeneous and contains features that are difficult to capture with a single model. Based on the dataset being used, the best ensemble approach should be chosen. Given the complexity and variety of the data collection, it may be best to combine several different techniques to obtain the appropriate level of accuracy. It is important to explore all of the available ensemble methods to determine which one provides the best result.
8 Meta-learning through ensemble approach
181
8.9 Advanced ensemble approaches Advanced ensemble techniques combine multiple machine learning algorithms to create a more powerful predictive model. In order to create an aggregate model that is more precise and reliable than any single individual model, this typically entails integrating various models. For classification, regression, and clustering problems, advanced ensemble approaches can be applied. This strategy makes use of well-liked methods like bagging, boosting, and stacking. These methods combine data from various models to get a final prediction, which is intended to reduce overfitting of a model. The prediction performance of the individual models is enhanced by the use of an ensemble technique since the combination of numerous models can capture more pertinent information and suppress irrelevant information.
8.9.1 Gradient boosting machine (GBM) variants A potent machine learning approach called Gradient boosting machine (GBM) is used to create predictive models from an ensemble of weak learners. It is a supervised learning algorithm that hones on a weak learner, and then combines them in an ensemble to create a stronger, more accurate model. GBM works by taking the errors from the previously produced model as inputs in the new model. This in turn reduces error and improves the overall model [27]. GBM comes in a variety of forms, including XGBoost, LightGBM, and CatBoost. A distributed gradient-boosting toolkit called XGBoost uses parallelization to speed up model training times. A novel gradient-boosting system called LightGBM makes use of a tree-based learning technique to speed both training and inference. It also includes a variety of cost functions, regularization, and shrinkage that further assist in reducing the model run time. Finally, CatBoost is a version of GBM that is specifically designed to work with categorical data, allowing information to be uniquely encoded in order to learn more effectively from a dataset.
8.9.2 Stacked ensembles with metalearners A form of machine learning technique called stacked ensembles with metalearners is used to ensemble models and increase prediction accuracy. This method combines multiple base models (usually referred to as ‘learners’) to create a single, more robust model (called ‘metalearner’). The base models can be of any type, such as deep learning, random forests, or logistic regression. The metalearner is supervised-trained on the predictions of all these base models, while the other learners in the ensemble are trained on a subset of the data (or even separately) [28]. In many machine learning
182
S. Ramasamy et al.
issues, particularly those where a single model lacks adequate accuracy or predictive ability, stacked ensembles are employed to increase prediction accuracy.
8.9.3 Ensemble pruning techniques Ensemble pruning is a method for reducing the size of a predictive model by removing unnecessary input variables and increasing its generalization performance. By eliminating unnecessary attributes or lowering the amount of attributes used in the model, this method can be used to simplify prediction models. The ensemble model can be more effective and accurate by choosing the most pertinent attributes. The pruning process involves identifying the least important features and removing them from the model. It can also be used to improve model interpretability by removing attributes that do not have a significant contribution to the prediction [29]. Thresholds can be used to define which attributes should be removed – such as removing attributes if the model accuracy is not decreased by a certain percentage. By using an ensemble of models, the pruning process can be performed in parallel; multiple pruned models can be combined to form a better-performing model than each individual model alone.
8.10 Applications and case studies The broad category of machine learning algorithms, known as “ensemble learning,” use many models to produce predictions or results. It can be used in a wide variety of applications, from predicting medical outcomes to designing autonomous vehicles. This kind of machine learning is predicated on the notion that integrating various models will produce superior outcomes to using only one stand-alone model. This is because each model may make different mistakes, and correcting those errors by combining them will ultimately lead to better predictions. Ensemble learning has been used in a number of case studies. In the field of healthcare, researchers have developed a variety of models that combine multiple data sources to predict medical outcomes. For example, one 2015 case study in France used ensemble learning to develop a system that predicted patient survival based on their medical history, laboratory tests, and other clinical parameters [30]. Ensemble learning has also been used in the field of autonomous vehicles, where different sensory inputs like camera data, lidar data, and radar data are combined to make decisions about how to navigate without the need for human input. Another common application of ensemble learning is computer vision, where models are used to identify and classify images. Ensemble learning has a wide variety of applications in many fields, and case studies of successful implementations demonstrate its potential utility.
8 Meta-learning through ensemble approach
183
8.10.1 Real-world applications of ensemble learning in various domains In order to build better models than could be obtained by utilizing just one algorithm or dataset, ensemble learning is a potent technique that integrates numerous algorithms and/or various sources of data. Many practical issues in computer vision, natural language processing, and complicated analytics can be resolved via ensemble learning. In the realm of computer vision, ensemble learning can be used to detect objects (e.g., faces or cars) in images or videos, as well as to classify the contents of images. For example, it could be used to detect faces for facial recognition applications or to classify cars in a traffic monitoring system [31]. In natural language processing, ensemble learning can be used to classify text into different topics or to categorize documents for information retrieval tasks. For instance, it can be used to classify text into categories such as sports, finance, or politics. Ensemble learning is also widely used in complex analytics applications, such as credit scoring and stock market analysis. Specifically, ensemble learning techniques can be used to develop models that accurately predict customer credit scores and stock prices. The use of ensemble learning in healthcare applications can increase the precision of patient diagnoses and provide individualized treatment plans for them. Using a patient’s medical history and biological data, for instance, ensemble learning can be used to create models that properly forecast the probability of developing specific diseases.
8.10.2 Success stories and examples showcasing the effectiveness of ensemble approaches Ensemble approaches, which involve combining or combining components of different algorithms or techniques, have been very successful in many different application areas. For instance, ensemble techniques can be used to aggregate the results of many machine learning algorithms to improve accuracy in the area of picture classification [32]. Another well-known illustration of an ensemble approach is the use of layered generalization, a method for combining the results of various models to produce predictions with higher accuracy. This method has been utilized successfully in the interpretation of medical images, where it has demonstrated increasing the accuracy compared to depending on a single model. Furthermore, ensemble approaches have also been used in the field of financial forecasting, in which multiple models are combined in order to provide more accurate and reliable investment decisions.
184
S. Ramasamy et al.
8.11 Challenges and future directions Challenges: – Developing an efficient way of providing access to large amounts of data – Integrating data from different sources and providing flexibility to users for data manipulation – Improving scalability to enable faster and more efficient processing of data – Designing a secure environment to ensure data privacy and security – Implementing cost-effective solutions to meet the growing demands of data analysis Future directions: – Implementing machine learning algorithms for predictive analytics – Increasing the speed and scalability of processing large volumes of data – Use of natural language processing technologies for extracting insights from data – Creating an effective data governance system to manage the data – Improving data visualization tools to increase user engagement
8.11.1 Addressing challenges and potential pitfalls in ensemble learning While ensemble learning is an effective technique for improving the performance of machine learning models, it is not without its difficulties and dangers. There are a number of ways that can be used to guarantee good ensemble learning. The first challenge for any ensemble learning system is incorporating the knowledge from multiple models into a single result [33]. This can be done in many ways, such as averaging the model predictions, weighting individual models differently, or creating a “voting” system where each individual model’s predictions are combined into one prediction. Another challenge of ensemble learning is selecting the “optimal” combination of models. To do this, models must be evaluated and compared in a systematic manner. Accuracy, precision, recall, the F1 score, and the area under the ROC curve are among the metrics that can be used to compare models. Building a good model ensemble requires understanding which metric to apply in each specific circumstance. Ensemble models can also suffer from model bias and overfitting. To prevent model bias, all models should be exposed to the same data and same preprocessing techniques. Additionally, when using techniques such as bagging or boosting to construct the ensemble model, the individual models must be carefully monitored to ensure that the ensemble does not overfit to the training data. The ensemble learning can be computationally expensive as the individual models in the ensemble must all be trained. To mitigate this challenge, techniques such as distributed computing, where many machines are used to train the models in parallel, can be used. The performance of machine learning models can be improved
8 Meta-learning through ensemble approach
185
by using ensemble learning, but its effectiveness depends on being aware of the difficulties and dangers the method presents.
8.11.2 Emerging research trends and future directions in meta-learning Meta-learning is an area of research that uses machine learning models to enable machines to learn how to learn. It has the ability to generalize previously learned tasks and apply them to quickly learn new tasks. One of the most active study fields in machine learning and AI in recent years is meta-learning [34]. Research in meta-learning focuses on bridging the gap between the recently acquired knowledge and out-ofsample generalization on never seen tasks. The approaches explored in this area mainly fall into three categories: 1. Meta-learning algorithms: These algorithms learn a set of parameters by taking into account the past solutions (known as archived solutions). It helps the algorithm to enhance its performance in dealing with out-of-sample tasks. 2. Meta-representation learning: This approach focuses on learning the representation that can facilitate fast adaptation to new tasks. 3. Meta-optimization: These algorithms focus on leveraging the acquired knowledge of solutions of previously solved tasks such as reinforcement learning algorithms. In the future, meta-learning research will focus on improving the learning speed and efficiency of meta-learning models while retaining the power of meta-knowledge. Researchers are also trying to understand the role of transferable meta-knowledge by combining it with domain knowledge, and this could lead to a better ability to handle new tasks. Another topic of interest is the use of meta-learning for multi-task learning, where the same meta-knowledge is used to solve different tasks. Finally, the ability to use meta-learning methods in other areas of artificial intelligence is gaining traction, such as in natural language processing, knowledge representation, and robotics.
8.11.3 Potential advancements and improvements in ensemble techniques In the field of machine learning, ensemble techniques have grown in popularity in an effort to produce more accurate models that are more generalizable than individual models[35]. Potential advancements and improvements in ensemble techniques include increasing the efficiency with which ensemble models can be constructed, developing new algorithms that can more effectively combine multiple models, exploring ways of automatically determining the best ensemble combination, and enhancing the ability of ensemble techniques to identify and exploit nonlinear relationships in the data. Addi-
186
S. Ramasamy et al.
tionally, further research could focus on boosting the predictive accuracy of ensemble models by including more data sources, incorporating additional sources of domain knowledge, or utilizing improved features.
References [1]
Olowookere, T. A., & Adewale, O. S. (2020). A framework for detecting credit card fraud with costsensitive meta-learning ensemble approach. Scientific African, 8, e00464. [2] de Paiva, B. B. M., Pereira, P. D., de Andrade, C. M. V., Gomes, V. M. R., Souza-Silva, M. V. R., Martins, K. P. M. P., . . . & Marcolino, M. S. (2023). Potential and limitations of machine meta-learning (ensemble) methods for predicting COVID-19 mortality in a large inhospital Brazilian dataset. Scientific Reports, 13(1), 3463. [3] Ma, Z., Guo, S., Xu, G., & Aziz, S. (2020). Meta learning-based hybrid ensemble approach for shortterm wind speed forecasting. IEEE Access, 8, 172859–172868. [4] Yadav, S. P., Zaidi, S., Mishra, A., & Yadav, V. (2022). Survey on machine learning in speech emotion recognition and vision systems using a recurrent neural network (RNN). Archives of Computational Methods in Engineering, 29(3), 1753–1770. [5] Guo, N., Di, K., Liu, H., Wang, Y., & Qiao, J. (2021). A metric-based meta-learning approach combined attention mechanism and ensemble learning for few-shot learning. Displays, 70, 102065. [6] Ahmed, K., Nadeem, M. I., Li, D., Zheng, Z., Al-Kahtani, N., Alkahtani, H. K., . . . & Mamyrbayev, O. (2023). Contextually Enriched Meta-Learning Ensemble Model for Urdu Sentiment Analysis. Symmetry, 15(3), 645. [7] Logeshwaran, J., & Karthick, S. (2022, April). A smart design of a multi-dimensional antenna to enhance the maximum signal clutch to the allowable standards in 5G communication networks. ICTACT Journal on Microelectronics, 8(1), 1269–1274. [8] Zhang, J., Song, J., Gao, L., Liu, Y., & Shen, H. T. (2022). Progressive meta-learning with curriculum. IEEE Transactions on Circuits and Systems for Video Technology, 32(9), 5916–5930. [9] Yadav, S. P., & Yadav, S. (2020). Image fusion using hybrid methods in multimodality medical images. Medical and Biological Engineering and Computing, 58, 669–687. [10] Gastinger, J., Nicolas, S., Stepić, D., Schmidt, M., & Schülke, A. (2021, July). A study on ensemble learning for time series forecasting and the need for meta-learning. In 2021 International Joint Conference on Neural Networks (IJCNN), (pp. 1–8). IEEE. [11] Gu, K., Zhang, Y., & Qiao, J. (2020). Ensemble meta-learning for few-shot soot density recognition. IEEE Transactions on Industrial Informatics, 17(3), 2261–2270. [12] Rajalakshmi, M., Saravanan, V., Arunprasad, V., Romero, C. T., Khalaf, O. I., & Karthik, C. (2022). Machine learning for modeling and control of industrial clarifier process. Intelligent Automation and Soft Computing, 32(1), 339–359. [13] Raja, S., Logeshwaran, J., Venkatasubramanian, S., Jayalakshmi, M., Rajeswari, N., Olaiya, N. G., & Mammo, W. D. (2022). OCHSA: Designing energy-efficient lifetime-aware leisure degree adaptive routing protocol with optimal cluster head selection for 5G communication network disaster management. Science Programme, 2022, 1–11 [14] Yang, A., Lu, C., Li, J., Huang, X., Ji, T., Li, X., & Sheng, Y. (2023). Application of meta-learning in cyberspace security: A survey. Digital Communications and Networks, 9(1), 67–78. [15] Anderson, R., Koh, Y. S., Dobbie, G., & Bifet, A. (2019). Recurring concept meta-learning for evolving data streams. Expert Systems with Applications, 138, 112832.
8 Meta-learning through ensemble approach
187
[16] Cawood, P., & Van Zyl, T. (2022). Evaluating state-of-the-art, forecasting ensembles and metalearning strategies for model fusion. Forecasting, 4(3), 732–751. [17] Shen, X., Dai, Q., Zhu, G., & Zhang, J. (2023). Dynamic ensemble pruning algorithms fusing metalearning with heuristic parameter optimization for time series prediction. Expert Systems with Applications, 225, 120148. [18] Hussain, S., & Ibrahim, N. (2022, June). Empirical investigation of role of meta-learning approaches for the improvement of software development process via software fault prediction. In Proceedings of the International Conference on Evaluation and Assessment in Software Engineering 2022, (pp.413–420). [19] Zhang, J., Song, J., Yao, Y., & Gao, L. (2021, October). Curriculum-based meta-learning. In Proceedings of the 29th ACM International Conference on Multimedia, (pp.1838–1846). [20] Saravanan, V., & Neeraja, A. (2013, January). Security issues in computer networks and stegnography. In 2013 7th International Conference on Intelligent Systems and Control (ISCO), (pp. 363–366). IEEE. [21] Logeshwaran, J., Ramkumar, M., Kiruthiga, T., & Sharan Pravin, R. (2022, February). SVPA – the segmentation based visual processing algorithm (SVPA) for illustration enhancements in digital video processing (DVP). ICTACT Journal on Image and Video Processing, 12(3), 2669–2673. [22] Kousik, N., Natarajan, Y., Raja, R. A., Kallam, S., Patan, R., & Gandomi, A. H. (2021). Improved salient object detection using hybrid Convolution Recurrent Neural Network. Expert Systems with Applications, 166, 114064. [23] Yadav, S. P., Mahato, D. P., & Linh, N. T. D. (Eds.). (2020). Distributed artificial intelligence: A modern approach. Boca Raton: CRC Press. [24] Ma, S., & Fildes, R. (2021). Retail sales forecasting with meta-learning. European Journal of Operational Research, 288(1), 111–128. [25] Irfan, D., Tang, X., Narayan, V., Mall, P. K., Srivastava, S., & Saravanan, V. (2022). Prediction of Quality Food Sale in Mart Using the AI-Based TOR Method. Journal of Food Quality, 2022, 1–13. [26] Natarajan, Y., Kannan, S., Selvaraj, C., & Mohanty, S. N. (2021). Forecasting energy generation in large photovoltaic plants using radial belief neural network. Sustainable Computing, 31, 100578. [27] Yadav, S. P., Agrawal, K. K., Bhati, B. S., Al-Turjman, F., & Mostarda, L. (2022). Blockchain-based cryptocurrency regulation: An overview. Computational Economics, 59(4), 1659–1675. [28] Leka, H. L., Fengli, Z., Kenea, A. T., Hundera, N. W., Tohye, T. G., & Tegene, A. T. (2023). PSO-based ensemble meta-learning approach for cloud virtual machine resource usage prediction. Symmetry, 15(3), 613. [29] Khan, S., Saravanan, V., Gnanaprakasam, C. N., Jaya Lakshmi, T., Deb, N., Othman, N. A. (2022). Privacy protection of healthcare data over social networks using machine learning algorithms. Computational Intelligence and Neuroscience, 2022, 8, Article ID 9985933. [30] Yuvaraj, N., Kousik, N. V., Jayasri, S., Daniel, A., & Rajakumar, P. (2019). A survey on various load balancing algorithm to improve the task scheduling in cloud computing environment. Journal of Advanced Research in Dynamical and Control Systems, 11(08), 2397–2406. [31] Jain, A., Kumar, A., & Susan, S. (2022). Evaluating deep neural network ensembles by majority voting cum meta-learning scheme. In Soft Computing and Signal Processing: Proceedings of 3rd ICSCSP 2020, (Vol. 2, pp. 29–37). Singapore: Springer. [32] Logeshwaran, J., & Shanmugasundaram, R. N. (2019, December). Enhancements of resource management for device to device (D2D) communication: A review. In 2019 Third International conference on I-SMAC (IoT in Social, Mobile, Analytics and Cloud)(I-SMAC), (pp. 51–55). IEEE.
188
S. Ramasamy et al.
[33] Hannah, S., Deepa, A. J., Chooralil, V. S., BrillySangeetha, S., Yuvaraj, N., Arshath Raja, R., . . . & Alene, A. (2022). Blockchain-based deep learning to process IoT data acquisition in cognitive data. BioMed Research International, 2022, 1–7 [34] Dvornik, N., Schmid, C., & Mairal, J. (2019). Diversity with cooperation: Ensemble methods for fewshot classification. In Proceedings of the IEEE/CVF International Conference on Computer Vision (pp.3723–3731). [35] Zang, H., Cheng, L., Ding, T., Cheung, K. W., Wei, Z., & Sun, G. (2020). Day-ahead photovoltaic power forecasting approach based on deep convolutional neural networks and meta learning. International Journal of Electrical Power and Energy Systems, 118, 105790.
Abhisek Sethy, Nazeer Shaik, Pavan Kumar Yadavalli, S. P. Anandaraj
9 AI: issues, concerns, and ethical considerations Abstract: Artificial intelligence (AI) ethics are the ethical considerations arising from the design and use of AI systems. Its focus is on understanding the ethical implications of developing, deploying, and using AI systems, and on developing strategies to ensure that AI systems are deployed responsibly and safely. AI ethics deals with issues such as fairness, data privacy, and security. It also covers topics such as ethical algorithms, ethical decision-making, trustworthiness, and transparency in AI. Keywords: Artificial, intelligence, ethics, design, developing, algorithms
9.1 Introduction Artificial intelligence (AI) ethics is an area of philosophy that deals with the moral implications of AI and how it affects society. It covers a broad range of topics including privacy, autonomy, bias, fairness, ownership, responsibility, trust and transparency. It is a critical field to consider because AI is increasingly being used in many different fields – from healthcare to finance – and its effects are far-reaching and have a potential for significant impact. Many ethical considerations should be taken into account in the development and deployment of AI technologies, such as ensuring data privacy, preventing discrimination, and ensuring fairness [1]. Additionally, ethical principles should be embedded in the design and implementation of AI systems and processes in order to ensure compliance with relevant laws and regulations as well as overcome the potential for unintended consequences.
9.1.1 Importance of ethical considerations in AI development and deployment AI systems are increasingly being used in many domains ranging from the health sciences to education, making the ethical considerations associated with AI development and deployment increasingly important. Examples of ethical considerations include privacy protection, data accuracy, fairness, transparency, and accountability. Privacy protection is important because people may be hesitant to use technology if their personal information is not protected. Data accuracy is also important for ensuring that AI systems are pulling data from accurate datasets and are making correct decisions based on that data. Fairness in AI means that all participants in the system are treated https://doi.org/10.1515/9783111323749-009
190
Abhisek Sethy et al.
equally and decisions are not based on bias or prejudice. Transparency makes sure that users have visibility into how AI systems are operated and accountable allows users to hold AI systems accountable for their decisions [2]. In essence, ethical considerations in AI development and deployment protect people, businesses, and the technology itself from potential misuse or manipulation. As such, it is essential that ethical considerations be taken into all aspects of AI development and deployment to ensure the accurate and ethical use of AI technology.
9.2 Bias and fairness in AI Bias and fairness in AI refers to the idea that algorithms, models, and other automated systems should be able to make decisions without bias or prejudice towards any person or group. This includes not only recognizing, understanding, or reflecting societal biases, but also having the capability to intentionally reduce them in order to create fair outcomes. An example of bias in AI is when an algorithm is trained to favor one type of customer more than others. Bias can be intentional or unintentional, and can often be embedded into algorithms due to processes like data collection and data preprocessing. Being aware of biases and working towards reducing them is key to creating fair AI systems. This can include taking steps to identify, measure, and adjust biases in the software, as well as creating checks and balances that allow for external review and accountability. Additionally, ethical standards such as fairness-aware machine learning (ML) can be incorporated throughout the software development life cycle to address bias issues as they arise.
9.2.1 Understanding bias in AI algorithms and data Understanding bias in AI algorithms and data is essential for ensuring fairness in AI-based decision making. Bias can be introduced into algorithms due to a number of sources including sampling errors, poor data sources, incorrect labels in datasets, and even the algorithm design process itself. Poorly designed AI algorithms can rely on inaccurate or incomplete datasets resulting in incomplete models that are biased against certain individuals or populations [3]. For example, facial recognition algorithms that are trained with data that does not properly represent various gender, age, or ethnic groups will be less accurate and exhibit bias when tested on those groups. Companies should be aware of how bias can creep into AI algorithms and datasets and work to mitigate it. This can involve the use of techniques such as active learning, which involve humans in the model building process, or strategies such as test-time augmentation to increase the diversity of the training data. Companies should also pay
9 AI: issues, concerns, and ethical considerations
191
attention to fairness metrics during the development process and formalize best practices for data collection and model testing.
9.2.2 Implications of biased AI systems on fairness and equity Biased AI systems can have a serious effect on fairness and equity. AI systems can be biased due to the data used to train them, the algorithms used to run them, and the decision-making process. Biased AI systems can result in decision-making that is unfair, unjust, or exclusionary. For example, AI systems used for hiring decisions may be more likely to select certain candidates based on gender or race. This can lead to a lack of diversity in the workforce and reduce opportunities for those not included in the algorithm [4]. It can also lead to decisions that are not in the best interests of the individuals affected, such as through automated admissions systems. AI bias can have far-reaching consequences for fairness and equity, both at an individual and societal level. It can lead to a lack of equal opportunity, in terms of access to education, jobs, or other opportunities. Additionally, AI bias can result in economic and social disparity, as advantages and disadvantages can be magnified across different groups. AI bias can also lead to a reduced level of trust between people and organizations, as people may feel their rights, security, or privacy are not being respected. In sum, AI bias has implications for fairness and equity that are far-reaching and can lead to serious consequences.
9.2.3 Strategies for detecting and mitigating bias in AI models –
– –
– –
–
Establish human-centered AI governance structures: Establish guidelines for AI models that focus on human values and ethical considerations. Integrate human oversight into the AI model design. Use representation analysis: Audit AI models and algorithms for fairness, accuracy, and bias. Analyze data to identify potentially biased inputs or results. Implement fairness algorithms: Use fairness algorithms to reduce the impact of bias. Algorithms that detect and correct for bias may be used in conjunction with supervised learning techniques to improve fairness. Evaluate and adjust models: Use testing tools to evaluate the effectiveness of models prior to deployment and adjust when necessary. Leverage diverse and quality datasets: Use high quality datasets that are reflective of real-world populations and behaviors. Include considerations with regard to minority populations and under served populations. Share information and practices: Provide transparent and clear information about how AI is being used and disseminate best practices across organizations to reduce bias.
192
Abhisek Sethy et al.
9.2.4 Fairness metrics and approaches in AI development Fairness metrics and approaches in AI development are used to improve the fairness of datasets and algorithms. They measure the performance of an algorithm on protected, sensitive characteristics such as gender or race [5]. By using fairness metrics, AI developers can assess whether an AI model is producing different results for similarly situated individuals and groups and is not treating some of them differently. This helps ensure that the outcomes of AI models are fair and unbiased. Some of the most popular fairness metrics and approaches include parity metrics, error rate parity, power parity, goodness of fit, predictive value parity, and statistical parity.
9.3 Transparency and explainability Transparency and explainability are two critical aspects of ensuring a responsible development and use of any AI system. Transparency means the ability of an AI system to provide an understanding of how it works, while explainability means the capacity of AI to explain itself and the decisions it makes. Transparency and explainability ensure that AI-developed systems are able to be monitored and monitored effectively without violating user privacy or security norms. Transparency and explainability also ensure that AI-developed systems are held accountable for decisions and outcomes, which in turn rely on the ability of the system to explain why it has taken certain decisions [6]. Transparency and explainability help in ensuring that AI systems are fair and compliant with ethical and legal norms. Consequently, transparency and explainability are important for trust in AI systems and for their acceptance and adoption by the larger society.
9.3.1 The need for transparent and explainable AI systems Transparent and explainable AI systems provide an improved understanding for how autonomous systems are making decisions, enabling all stakeholders to build trust and respect for AI. This includes better climate modeling for weather and environmental predictions, improved medical diagnosis use-cases, and increased autonomous driving accuracy and safety. Additionally, regulatory and legal compliance requirements can be met more easily when using transparent and explainable AI systems, as the decision-making process can now be more easily understood and tracked. In all areas, less bias can be introduced into models and decision-making when explanation is available. Finally, transparent and explainable AI systems can improve many automation and self-learning systems by providing enhanced debugging and troubleshooting capabilities [7].
9 AI: issues, concerns, and ethical considerations
193
9.3.2 Challenges in understanding and interpreting AI decisions There are three main challenges to understanding and interpreting AI decisions: 1. Explainability: AI decisions are frequently made using complex algorithms and mathematical models, making it difficult for humans to understand the underlying logic. As a result, AI developers and researchers are tasked with designing systems that can explain their decision-making processes. 2. Fairness and bias: AI decision-making systems can quickly become biased if they are not carefully monitored and managed. To ensure fairness, AI models should be tested and monitored closely for any signs of bias. This can include analyzing datasets for potential sources of prejudice or developing effective processes for understanding and mitigating bias. 3. Uncertainty and reliability: AI decision-making systems are often based on incomplete or uncertain datasets. As such, it can be difficult to accurately assess the accuracy and reliability of AI decisions. This challenge can be addressed through careful monitoring and testing of AI systems in order to confirm their accuracy and reliability.
9.3.3 Techniques for enhancing transparency and explainability in AI There are various techniques for enhancing transparency and explainability in AI systems. These techniques can help AI systems to provide explainable decisions, increase user trust, and mitigate potential acceptance and usage issues. 1. Post-hoc explanation method: This method focuses on the explainability after the result is produced. It is a set of methods used for providing explainable insights into the AI system’s operations and decisions. Such techniques include global explanations to understand the overall system decision-making, local explanations for Interpretable features, model debugging for locating important features, feature relevance and inputs sensitivity for understanding the system’s behavior/decision on an input, feature extraction and highlighting for understanding the decisions based on features, feature clustering for understanding the features driving a decision, visual representations to provide a visual explanation of the system’s decisions, and algorithm visualizations to understand the internals of a system such as input-output relations, etc. 2. Predictive model explainability: This technique, with the aid of certain algorithms such as generalized additive models (GAMs), rule-based models (RBMs), attempts to explain the AI system’s decisions and operations from before the operation or decision. GAMs allow the user to model a complex nonlinear system using linear models while RBMs express complex behavior using a set of rules.
194
3.
4.
Abhisek Sethy et al.
Explainable AI: This technique is a set of techniques leveraging user analytics, model-level interpretability, problem specification, and system dynamics to provide explainability into an AI system. It uses a combination of algorithms such as AutoML, deep learning, reinforcement learning, and explanation techniques such as feature importance and sensitivity analysis. Data-driven explainability: This technique tries to explain the decision before is made by using AI algorithms and techniques such as documents similarity analysis, information extraction, data analysis, and visualization. It is also used to demonstrate the responsible algorithm to the user.
These techniques are essential for developing and deploying explainable AI systems. They can enable AI systems to provide more transparent and understandable results. This would help in increasing user trust and well as fostering better acceptance and usage of AI systems.
9.3.4 Regulatory requirements and standards for AI transparency Regulatory requirements and standards for AI transparency focus on establishing rules and frameworks that ensure that automated decisions made by an AI system are explainable and traceable. They are intended to protect the rights of individuals and businesses by ensuring that AI-driven decisions are transparent and accountable. Increasingly, regulatory bodies are taking a proactive role in promoting industry standards around AI transparency [8]. AI transparency requirements vary across jurisdictions, but certain key elements are commonly included: disclosure of AI processes, limits on data usage and protection, safeguards to protect vulnerable populations and minorities, and ongoing monitoring of results. In some cases, these requirements can require specific measures such as establishing privacy impact assessments, creating public dispute resolution systems, and ensuring end-user access to algorithms and data. These measures help ensure that AI-driven decisions made by an entity are explainable to stakeholders, traceable to the data used, and auditable to ensure that algorithms are not biased or discriminatory.
9.4 Privacy and data protection Privacy and data protection is the practice of protecting data from unauthorized use, access, and disclosure. It involves the collection, storage, and use of information about individuals, organizations, and the public. It also includes protecting data from malicious actors, such as hackers and identity thieves. Some of the key concepts of privacy and data protection include respecting the privacy of individuals, securing data collected by organizations, and ensuring that data is used in a way that is compli-
9 AI: issues, concerns, and ethical considerations
195
ant with regulations and laws. Properly implementing data privacy and protection can protect businesses from data breaches, reputation damage, fines, and other negative consequences [9].
9.4.1 Privacy concerns in AI applications and data collection Privacy concerns in AI applications and data collection are becoming increasingly prevalent. AI can analyze large amounts of data to draw patterns and make predictions about user behavior that can be used to make sophisticated decisions that may not always be in the user’s best interests. Additionally, personal data that is collected and used to train AI models may be collected without the user’s knowledge. This can put users at risk of being targeted by malicious actors or their data being sold or used in ways they are unaware of [10]. The risk of data leakage and unauthorized access or use of data can also increase when AI models are trained on large datasets containing sensitive data. In order to ensure user privacy when deploying AI models, organizations must ensure that their data is properly encrypted and secured and that appropriate policies are in place to ensure the responsible use of data.
9.4.2 Data protection regulations and compliance Data protection regulations are rules established by governments worldwide to protect personal or sensitive data stored on electronic devices. These regulations set limits on how information can be collected and used, and define what must be done to protect the data when stored and transferred. Compliance with data protection regulations requires organizations to take specific steps such as implementing encryption, access controls, data classification, and security awareness training. Organizations must also notify individuals whenever their data is collected and provide ways for them to access, modify, or delete the information. Penalties for failing to comply with data protection regulations can include fines, disciplinary action, and in some cases, criminal prosecution. By following data protection regulations, businesses can demonstrate trust and legitimacy when handling sensitive information and maintain their customers’ trust.
9.4.3 Privacy-preserving AI techniques Privacy-preserving AI techniques refer to methods and algorithms used in AI to protect sensitive personal and corporate information such as credit card numbers, address details, and social security numbers, while allowing data to be analyzed to leverage the benefits of AI. AI systems can be used to process large amounts of data and generate
196
Abhisek Sethy et al.
useful insights, but much of this data may contain sensitive information that needs to be kept secure [11]. Privacy-preserving AI technology helps protect personal data while also allowing AI algorithms to generate useful insights. Some common techniques used include differential privacy, homomorphic encryption, federated learning, and secure multiparty computation.
9.4.4 Balancing data utility and privacy in AI development Balancing data utility and privacy in AI development is an important factor to consider when using AI to capture and analyze data. This means that AI systems must be designed with certain safeguards to ensure that data is not misused or abused in the process of development. It is critical to ensure that user data is secured and that the AI-based systems are not used to invade privacy or intentionally or inadvertently manipulate data. There are various approaches to balancing data utility and privacy in the AI development cycle [12]. One approach is to limit the access to the data that is critical in the development process to ensure that only those with the necessary expertise and permission to access it are able to do so. To further protect user privacy, the use of artificial neural networks and other privacy-centric technologies can be implemented as part of the development cycle. Additionally, data sanitization techniques such as data anonymization and data masking can be used to protect user privacy while still allowing for meaningful development and analysis. number of actual qubits to make it possible to develop robust error correction systems.
9.5 Accountability and responsibility Accountability and responsibility are key concepts to consider when developing AI systems. Accountability is the ability to explain the decisions and answers given by a system, which increases transparency and trust in the system’s accuracy and reliability. Responsibility, on the other hand, involves ensuring that the system’s decisionmaking process is supervised properly, adhering to principles such as data privacy and security, fairness, and ethical considerations [13]. This ultimately makes sure that the system does not become a source of harm and establishes trust in the process. Both accountability and responsibility are paramount to the successful development of AI systems, creating a framework to insure that the system is reliable, safe, and ethical.
9 AI: issues, concerns, and ethical considerations
197
9.5.1 Determining accountability for AI system outcomes Determining accountability for AI system outcomes is a complex and important challenge in the development and deployment of AI systems. Accountability means taking responsibility for actions and outcomes, including the possibility of punishment if these are found to be incorrect. For AI systems, this means that those responsible for developing and deploying such systems must be held responsible for the outcomes they produce. This involves not only ensuring that the technology works as intended but also that its use does not cause harm. A critical component in the process of accountability is the ability to trace the decisions made by the AI system back to their creators. This requires the use of appropriate systems of record-keeping, documentation, and auditing [14]. Additionally, AI systems must be built and tested according to best practices, and have systems in place for responding to mistakes or unforeseen consequences. Finally, organizations must be transparent in the ways in which their systems are used, including making results and decisions open to external scrutiny.
9.5.2 Ethical considerations in autonomous AI systems Ethical considerations in autonomous AI systems must be taken into account in order to ensure the safety, privacy, and fairness of people affected by their use. This includes considering questions such as whether or not the input data is biased or unbalanced, how will decisions be made, how those decisions impact people, and whether or not they are in compliance with laws and regulations. Other considerations include questions about privacy, such as: how will any data collected be used and where will it go? The use of autonomous AI systems has the potential to cause tremendous changes in the way we live and interact with the world around us. As such, it is important to consider and assess the ethical implications of their use [15–18]. An ethical framework should be applied when developing or using autonomous AI systems, considering the potential unethical impacts of their use and methods for managing those risks. This could include considering the potential for the AI system to cause harm to people, to be used for illegal or immoral purposes, or to infringe upon someone’s rights or privacy. As AI technologies continue to develop and become more pervasive, it is important to pay attention to ethical considerations in the development and use of autonomous AI systems.
9.5.3 Legal and ethical frameworks for AI responsibility Legal and ethical frameworks for AI responsibility are based on laws and codes of ethical conduct that govern the use of AI technology. These frameworks indicate how AI systems should be developed, used, and deployed, in order to respect and uphold
198
Abhisek Sethy et al.
ethical standards and comply with laws and regulations related to AI. The frameworks provide guidance on various aspects of AI technology and the responsible use of AI, such as privacy, data security and safety, transparency, algorithmic decisionmaking, accountability, and fairness. They may also address legal questions about liability and responsibility, as well as how scenarios involving AI technology and people should be addressed. Laws and frameworks governing AI may also extent to AIenabled products, including those related to autonomous vehicles, healthcare technologies, facial recognition, and other services [19]. At the heart of any development, deployment, or usage of AI technology is the importance of ensuring that user safety, privacy, and security are primary considerations. AI systems must also be transparent about their purpose, design, and function, while still safeguarding personal privacy and guaranteeing anonymity when appropriate. Finally, AI systems must be held accountable and evaluated for fairness and accuracy in automated decision-making processes. The legal and ethical frameworks for AI responsibility seek to both protect individuals and create an environment of fairness, trust, and responsibility when it comes to how AI systems are developed, used, and deployed.
9.5.4 Implications of AI decision-making and potential harms AI decision-making has the potential to lead to a variety of harms, such as decisions that are nonobjective, unfair, and undesirable. Conversely, AI decision-making can also lead to unintended harm to individuals and society, such as discrimination against protected classes of people, such as gender, race, etc. Without proper checks and balances, AI decision-making can also lead to unethical decisions or negligence in decision-making [20]. Furthermore, AI may lead to unintended consequences due to its reliance on large datasets, which can lead to inaccurate or incomplete results if not properly checked or monitored. Finally, AI may pose privacy concerns if data is not collected or shared responsibly, leading to misuse of information.
9.6 Robustness and security Robustness is the ability of a system to resist failure and to continue to operate when exposed to potential attacks or threats. Robustness is often achieved by thorough testing and rigorous security measures. Security measures such as authentication and authorization, encryption, and virus protection are particularly important for robust systems [21]. Security is the practice of safeguarding information and resources from unauthorized access, disclosure, destruction, or modification. It is the ability to protect system components from harm as well as the ability to prevent and detect unauthorized ac-
9 AI: issues, concerns, and ethical considerations
199
cess. Proper security measures can help protect data from unauthorized use, while also protecting the system from malicious actors and attacks[22].
9.6.1 Ensuring robustness and resilience of AI systems Ensuring robustness and resilience of AI systems is one of the most important challenges facing the AI community. Robustness is the ability of an AI system to perform consistently across a variety of operating conditions. Resilience is the ability of an AI system to be able to adapt and recover after being subjected to sudden changes in its environment. To ensure robustness and resilience of AI systems, several approaches can be followed. One approach involves testing AI systems during development to ensure they are able to perform in a variety of operating conditions. This helps identify areas of improvement that can be worked on prior to deployment. Additionally, researchers can use techniques such as reinforcement learning and deep learning to develop AI models that are robust and resilient to changes in their environment. Such models can be used to ensure that AI systems are able to adapt to sudden changes and can continue to provide accurate outputs after being subjected to such changes [23]. Furthermore, organizations can use safety protocols and protocols for monitoring AI systems to ensure that they are operating within acceptable parameters. Additionally, organizations can use techniques such as data partitioning to help AI systems identify anomalies more effectively, allowing them to adapt to changes in their environment more quickly. Finally, AI systems can be designed to take into account the fallibility of humans involved in their operation to ensure that AI systems are able to respond to sudden changes.
9.6.2 Adversarial attacks and defenses in AI Adversarial attacks in AI refer to malicious alterations of data used as inputs, which are designed to manipulate an AI system’s output or threaten its safety, security, accuracy, and trustworthiness. These modifications may be intentional or unintentional, but in either case can be used to produce unpredictable and potentially damaging results [24]. Adversarial defenses in AI refer to methods of protecting against adversarial attacks. These defenses can be either proactive, such as through the use of data augmentation or robust AI models; or they can be reactive, through detection and classification of adversarial samples. Proactive defenses are geared towards ensuring that AI systems are robust and secure from the outset, while reactive defenses aim to detect and identify malicious alterations as they occur. These defenses will become increasingly important as AI systems become more widely used and deployed [25].
200
Abhisek Sethy et al.
9.6.3 Security considerations in AI deployment When deploying AI, organizations need to consider a range of security considerations. These include: – Ensuring the data used to train AI models is collected securely and is free from bias. – Ensuring the AI system does not leak sensitive information such as user’s identity or their personal data. – Validating the AI system is working as designed without it being maliciously or unintentionally misused or hacked – Ensuring the system is compliant with regulations such as General Data Protection Regulation (GDPR) – Identifying potential data privacy and compliance risks associated with the AI system – Implementing firewalls, encryption methods, and other security measures to protect the AI system from attack – Adopting security protocols and best practices for using the AI system – Reviewing and updating security policies to ensure they are keeping pace with AI technology
9.6.4 Safeguarding AI systems against malicious intent Safeguarding AI systems against malicious intent involves securing the data, assumptions, algorithms, and infrastructure components used to create and deploy AI systems. Security teams should use a variety of techniques including access control, identifying critical elements, and monitoring system operations to secure AI systems. Furthermore, organizations should ensure that data used to train, test, and deploy AI systems should be secure and that the data contains no malicious content [26]. This includes rigorously testing for malicious code and protecting sensitive data. They should also have robust procedures to identify malicious actors and thwart threats. Finally, they should ensure that the infrastructure used to deploy AI systems is secure. This includes securing network infrastructure and enforcing authentication and authorization protocols. By taking these steps, organizations can ensure that AI systems are resilient against malicious intent.
9.7 Social and economic impacts The social and economic impacts of AI technology are considerable and varied. AI technology is being used by companies, organizations, and governments to automate processes, improve customer experiences, and inform decision-making. AI technology
9 AI: issues, concerns, and ethical considerations
201
has the potential to improve productivity, reduce costs, and open up new business opportunities. It can streamline processes, reduce labor costs, and generate more efficient operations. It can also automate data entry, improve accuracy, and cut down on paperwork. On the customer service side, AI technology can help create better customer experiences with personalized recommendations and automated problemsolving tools. In addition to the positive effects of AI technology, there are also potential risks. AI technology can increase unemployment, disrupt the labor market, and raise ethical and legal concerns [27]. There are also worries that AI technologies could lead to a loss of privacy as data is collected and used in unknown ways. There are social changes to consider. AI technology can change the way people interact, leading to a rise in virtual socializing. It could also have a wider impact on society by increasing inequality, disrupting human interactions, and introducing bias into decisionmaking processes. The social and economic impacts of AI technology are significant and varied. Ultimately, it is important for society to explore these impacts and ensure that the benefits outweigh any potential harms.
9.7.1 Evaluating the social and economic implications of AI Evaluating the social and economic implications of AI involves looking at how AI impacts individuals, organizations, and society as a whole. This means considering how AI might affect the labor market, income inequality, privacy, security, and access to services, as well as the ethical boundaries that should govern its use [28]. AI can lead to profound shifts in the structures of the global economy and changes in how society functions, enabling new possibilities in business and public services, but also leading to new threats and vulnerabilities. Understanding the implications of AI on society and the economy is essential for designing policies, regulations, and incentives that promote beneficial applications while preventing or limiting the risks associated with AI.
9.7.2 Impact of automation on employment and workforce The impact of automation on employment and the workforce is significant, as automation replaces manual labor in repetitive tasks. Automation has reduced the need for a large workforce and is expected to continue to reduce the need for manual labor jobs in the near future. Automation also contributes to an increasing number of repetitive tasks being automated, which further reduces the workforce needed by businesses and other organizations. In addition, automation can reduce some of the job roles within a company, such as in customer service or data entry, as digital technology can be deployed in many of these roles, often outperforming a human worker. This poses difficulties for those workers who might have been employed in these roles, as they will have to find new employment. Automation also leads to an increase
202
Abhisek Sethy et al.
in productivity as employers can reduce the amount of time it takes to complete tasks and increase return on investments. This increases the cost effectiveness of the business model and can lead to greater profits overall. It can also lead to the development of new products and services that would not have been possible to develop before. Automation is thus seen as an important component in business strategy. The automation can lead to increased job satisfaction as employees no longer need to complete tedious and repetitive tasks, freeing up more time for creative tasks and problem solving [29]. This can lead to increased morale in the workplace and enhance productivity. The automation has had a positive impact on businesses, from reducing the need for manual labor to increasing productivity and job satisfaction. However, it has also had a negative impact on some workers, reducing the number of manual labor jobs available and impacting those who may have been employed in those jobs.
9.7.3 AI and inequality: addressing disparities and access AI and inequality are two closely related concepts. AI can help address disparities and access for people across different segments of society. AI can help in mitigating disparities and granting access to marginalized populations – for example, by providing more personalized education. AI can also help in providing access to better healthcare services and financial services for underserved populations. Additionally, AI can be used to identify trends in a wide variety of datasets to better understand how inequality exists within communities. AI can also be used to develop policies and programs to reduce disparities and increase access for vulnerable populations. Finally, AI can be used to monitor the effectiveness of such policies and ensure that they are adequately addressing the needs of all groups in society.
9.7.4 Ethical considerations in AI deployment in developing countries The deployment of AI in developing countries presents ethical considerations that should not be overlooked. These ethical considerations include, but are not limited to: – Ensuring that AI is deployed in an equitable and ethical manner, free from bias and discrimination. This means ensuring that AI-based decision-making abides by the same ethical code of conduct that would be expected from humans. – Giving consideration to the impact that the deployment of AI might have on an individual or a community. AI has the potential to be incredibly powerful and this power should be used responsibly. – Allowing individuals and communities the right to decide whether or not they accept or reject technology. This should be done in such a way that gives autonomy and dignity to the people it affects.
9 AI: issues, concerns, and ethical considerations
–
–
203
Establishing safeguards against data privacy and security issues. Developing countries are often more vulnerable to cybersecurity threats as their technology infrastructure may not be as secure as that of developed countries. Therefore, appropriate safeguards should be put in place to ensure data privacy, integrity, and security. Educating people about AI and its implications. People in developing countries will benefit the most from the use of AI but should not be left in the dark about the implications it may have. It is important to ensure people understand the implications of AI so that they can make informed decisions about their own lives.
9.8 Ethical frameworks and guidelines Ethical frameworks and guidelines are focused on establishing strategies for decisionmaking and professional conduct that promote ethical practices and behaviors. These frameworks help ensure that people act ethically and responsibly in their professional lives, which can encompass elements such as rights and responsibilities, costs and benefits, accountability and self-discipline, and fairness and justice [30]. By using these frameworks, professionals can ensure that their decisions are based on principles of fairness and morality, and that their work is conducted in ways that uphold ethical standards and values.
9.8.1 Overview of existing ethical frameworks and guidelines for AI Existing ethical frameworks and guidelines for AI are sets of guiding principles created to ensure that AI systems are implemented in a responsible and ethical manner. Generally, these frameworks operate by outlining various ethical considerations related to the development and use of AI technologies, such as implications for privacy, safety, transparency, fairness, accountability, and human rights. They are designed to ensure that such technologies are implemented with consideration for how their usage may impact the lives of individuals, communities, and society as a whole. Furthermore, these frameworks may also provide suggestions for researchers and practitioners to ensure that AI applications are used responsibly and in accordance with the applicable laws and regulations [31]. Ultimately, these ethical frameworks are intended to provide guidance as AI technology continues to become more prominent in everyday life and serve to ensure that its use is beneficial and unharmful.
204
Abhisek Sethy et al.
9.8.2 Ethical considerations in AI research and development Ethical considerations in AI research and development refer to the expectations, guidelines, and principles that govern the development and implementation of AI applications and systems. These considerations include respecting people’s rights, privacy, security, and safety; considering the potential negative impacts; understanding the implications of the AI’s own objectives; and promoting transparency and accountability. Respect for human rights includes recognizing the importance of engaging with individuals and communities affected by AI, allowing humans to have control over how AI will be used, and protecting users from potential harm. Privacy and security should also be addressed, as well as shared responsibility between the developers and those affected by the AI [32]. Additionally, AI should be designed and developed in a manner that limits unintended negative impacts such as discrimination, inadequate decision-making, or a lack of safety and security. Transparency and accountability in AI research and development should be a key component of ethical considerations, as this will help build trust and ensure that AI is used responsibly.
9.8.3 Incorporating ethical practices throughout the AI life cycle Incorporating ethical practices throughout the AI life cycle is a process of aligning ethical principles with the development and deployment of AI. This involves identifying ethical hazards throughout the design, training, deployment, and use of AI systems, as well as actively promoting user safety and privacy. AI developers should strive to be mindful of potential ethical fallouts from their designs, which may have adverse effects on individuals, communities, and society as a whole. In order to mitigate such issues, developers can consider incorporating the following best practices into their AI life cycle: – Examine how the proposed AI system will affect users. AI researchers and developers should consider the potential risks posed by their AI system, such as the potential for overfitting data that affects a user’s privacy or security. – Consider ethical implications prior to deployment. AI developers should strive to identify potential ethical issues and limit adverse effects before releasing AI into the wild. This includes taking into account the potential unintended consequences of an AI system, such as its prejudice against certain minority groups or its potentially biased outcomes. – Establish oversight and review processes. AI development teams should design an oversight and review system in order to address ethical issues that may arise. This can include developing an audit mechanism to check data inputs as well as implementing measures for cross-checking outputs and making sure research outcomes are consistent with indicated ethical principles.
9 AI: issues, concerns, and ethical considerations
–
205
Develop accountability and transparency processes. Developers should take responsibility for the AI system’s performance and any ethical implications that arise. This can involve employing user feedback mechanisms to provide whistleblowing protection, in addition to regular communication to ensure users and the public are informed of any changes to the AI system.
9.8.4 Ethical decision-making models and frameworks Ethical decision-making models and frameworks provide a structured approach to navigating ethical issues. These models often provide step-by-step guidance to consider the ethical implications of a situation and make a decision based on a set of agreed-upon principles. Ethical models help us to move beyond our own biases and assumptions to consider the ethical implications of an action in a more systematic way. Ethical models are helpful for making consistent decisions that reflect the values and principles of a community or organization [33]. They may also help create accountability and allow for transparent decision-making. Common models and frameworks include the utilitarianism model, the rights framework, and the virtue framework. Each framework encourages decision-makers to consider different dimensions in making an ethical decision.
9.9 Ethical AI governance Ethical AI governance is a system of principles, laws, guidelines, and policies designed to ensure that organizations using or deploying AI-based technologies and solutions work ethically and responsibly in a socially responsible way. It includes setting policies on data collection, privacy, and usage, transparency, accountability, bias detection and remediation, and impact mitigation. AI governance also involves research, development, and deployment of AI systems within a framework that respects rights and values established by society, businesses, governments, and academic institutions. It ensures that AI is not used to inflict harm on or violate the rights of individuals [34].
9.9.1 Establishing ethical AI governance frameworks and mechanisms Establishing ethical AI governance frameworks and mechanisms is a process of developing a set of standards and guidelines that ensure that AI technologies are managed responsibly and safely. Ethical AI governance frameworks and mechanisms strive to mitigate potential risks and harms associated with the use of AI. These frameworks
206
Abhisek Sethy et al.
and mechanisms are critical for addressing the ethical, legal, and societal implications of AI technologies. They also provide guidance for building trust in AI systems, ensure the responsible use of data, and promote transparency in decision-making. Ethical AI governance frameworks and mechanisms aim to ensure that AI is used for the good of humanity by balancing the benefits and potential risks associated with the technology [35]. Examples of such frameworks and mechanisms include ethical AI codes of conduct, data protection standards, and algorithmic fairness initiatives.
9.9.2 Roles and responsibilities of stakeholders in AI governance The roles and responsibilities of stakeholders in AI governance can vary depending on the type of AI that is being used, the specific applications, the sector, and the jurisdiction. Generally, stakeholders with a vested interest in AI governance are corporate buyers and sellers, AI technology providers, policy analysts, government agencies, academics, and individual citizens. Each stakeholder plays a critical role in helping ensure responsible AI production, implementation, and usage in any given industry. For example, corporate buyers and sellers are responsible for understanding the risks associated with deploying AI technology, adhering to safety measures, and ensuring that the AI is reliable and accurate. AI technology providers need to be aware of all the implications of their products and ensure that they are being developed in compliance with legislation, industry standards, and ethical best practices. Policy analysts must ensure that any proposed legislation is properly understood and enforced, while government agencies are in charge of formulating laws and regulations [36]. Academics have an imperative role in researching the implications of AI and providing guidance on how to mitigate risks. Lastly, individual citizens have the responsibility of being aware of their rights and obligations when it comes to the use of AI technology. The aim of stakeholders in AI governance is to ensure that ethical principles are followed and that safety protocols are adhered to. Ultimately, stakeholders must work together in order to ensure that the responsible use of AI technology is encouraged and its potential to improve lives is unlocked.
9.9.3 Ensuring transparency and accountability in AI governance Ensuring transparency and accountability in AI governance refers to the need to ensure that AI governance frameworks are open and transparent, with clear lines of responsibility and decision-making authority. It also requires that there is an independent entity, responsible for the implementation of the AI system, to oversee the development and implementation process. In addition, any decision-making involving the use of AI should be based on objective evidence and the public should have access to the steps taken in the development, implementation, and decision-making process. This means
9 AI: issues, concerns, and ethical considerations
207
ensuring that data used in the AI system development is fair, unbiased, and accurate and that the AI system is effectively monitored and regulated for its accuracy and any adverse effects that it might produce. Ultimately, transparency and accountability in AI governance ensures that AI systems are responsible and beneficial for everyone.
9.9.4 International collaborations and initiatives for ethical AI International collaborations and initiatives for ethical AI are global collaborations and initiatives that focus on ensuring that AI systems are designed, developed, and deployed according to ethical and humane principles. These initiatives include collaborations between governments, corporations, nonprofits, and individuals, who are working together to ensure ethical AI development and use. These initiatives are led by a variety of respected global organizations, including the United Nations, the European Union, the World Economic Forum, and the World Organization for Human Rights. Examples of international collaborations and initiatives for ethical AI include the High-Level Panel on Digital Cooperation, the Global Partnership on Artificial Intelligence, the Global Network of Digital Cooperation Hubs, and the Global AI Council. These collaborations and initiatives focus on setting standards and guidelines for ethical AI development, sharing best practices and using data responsibly, ensuring responsible AI deployment, and developing international cooperation and coordination on AI ethics and governance.
9.10 Future challenges and directions The future of AI systems continues to be a source of excitement for many researchers and developers. The potential of AI systems to improve various aspects of life has already been demonstrated, yet the full power of AI systems and technologies remains largely untapped. The next few years will witness a significant increase in research efforts aimed at developing and deploying even more powerful AI systems. One of the major challenges future AI systems will have to address is improving the reliability of decision-making processes. In order for AI systems to be trusted and accepted, it is necessary to ensure that they are able to make accurate decisions even in complex or unforeseen circumstances. This involves teaching AI systems to detect and respond to unexpected events, changing patterns, and new types of data. The development of systems that are able to explain the decisions they make is another important challenge for future AI systems. This is especially important when it comes to applications where human lives and safety are at stake, such as autonomous vehicles or medical decision support systems. In order for humans to trust and accept the decisions of AI systems, it is necessary that users are able to understand how and why the decisions
208
Abhisek Sethy et al.
are being made [37]. AI systems and solutions must be designed with privacy and security in mind. As more sensitive data is being collected and shared with AI systems, there is increased need for systems with strong security measures that protect the data and personal information of individuals. This includes the development of secure infrastructure, authentication measures, encryption, and other tools and techniques that ensure the security of data.
9.10.1 Emerging ethical challenges in AI Emerging ethical challenges in AI are those questions and dilemmas that arise as AI and ML are increasingly used in our society and economy. These ethical questions have become increasingly important as AI and ML algorithms are used to make decisions and take action in areas ranging from medical care to financial services. The questions include addressing potential risks around privacy, data security, accuracy, accountability, and algorithmic bias, as well as the broader implications of applying AI for social and economic purposes. It also necessitates a better understanding of the legal, moral, and ethical implications of the technology. AI must be subject to standards that ensure outcomes are produced in a responsible and trustworthy manner, taking into account the potential risks and benefits when a computer is deciding on behalf of a person.
9.10.2 Addressing ethical considerations in emerging AI technologies Addressing ethical considerations in emerging AI technologies refers to taking a proactive stance towards understanding and managing the potential ethical challenges posed by the widespread implementation of AI. This involves creating and adopting ethical frameworks, policies, and regulations to minimize potential harms and ensure that the intended benefits of the technology are realized. It also involves understanding how AI can impact society and working to ensure that any negative effects are minimized. This includes developing standards for collecting, analyzing, and managing data used for AI, determining appropriate levels of transparency around AI systems, and addressing issues such as bias and privacy. Additionally, it means creating mechanisms for monitoring and regulating AI systems in order to ensure that they remain compliant with established ethical principles.
9 AI: issues, concerns, and ethical considerations
209
9.10.3 The role of interdisciplinary collaboration in ethical AI development Interdisciplinary collaboration is essential for the development of ethical AI systems. This collaboration should involve a range of experts and stakeholders from different disciplines including computer science, philosophy, law, ethics, and social sciences, as well as industry experts [38]. This ensures a wide range of perspectives are brought to bear and a more thorough consideration of ethical issues. Such collaboration helps inform the current best practice in developing ethical AI systems, identify potential weaknesses and develop solutions, and create industry standards. Additionally, it helps ensure the development of AI systems, which are developed and used responsibly in the right circumstances, and which have the capability to make ethical decisions. To be effective, interdisciplinary collaboration must be ongoing to ensure that the development of ethical AI systems keeps up with advances in AI technology.
References [1]
Munoko, I., Brown-Liburd, H. L., & Vasarhelyi, M. (2020). The ethical implications of using artificial intelligence in auditing. Journal of Business Ethics, 167, 209–234. [2] Fiske, A., Henningsen, P., & Buyx, A. (2019). Your robot therapist will see you now: Ethical implications of embodied artificial intelligence in psychiatry, psychology, and psychotherapy. Journal of Medical Internet Research, 21(5), e13216. [3] Keskinbora, K. H. (2019). Medical ethics considerations on artificial intelligence. Journal of Clinical Neuroscience, 64, 277–282. [4] Varlamov, O. O., Chuvikov, D. A., Adamova, L. E., Petrov, M. A., Zabolotskaya, I. K., & Zhilina, T. N. (2019). Logical, philosophical and ethical aspects of AI in medicine. International Journal of Machine Learning and Computing, 9(6), 868. [5] Giuffrida, I. (2019). Liability for AI decision-making: Some legal and ethical considerations. Fordham Law Review, 88, 439. [6] Schönberger, D. (2019). Artificial intelligence in healthcare: A critical analysis of the legal and ethical implications. International Journal of Law and Information Technology, 27(2), 171–203. [7] Holmes, W., Porayska-Pomsta, K., Holstein, K., Sutherland, E., Baker, T., Shum, S. B., . . . & Koedinger, K. R. (2021). Ethics of AI in education: Towards a community-wide framework. International Journal of Artificial Intelligence in Education, 2021, 1–23. [8] Safdar, N. M., Banja, J. D., & Meltzer, C. C. (2020). Ethical considerations in artificial intelligence. European Journal of Radiology, 122, 108768. [9] Hancock, J. T., Naaman, M., & Levy, K. (2020). AI-mediated communication: Definition, research agenda, and ethical considerations. Journal of Computer-Mediated Communication, 25(1), 89–100. [10] Ouchchy, L., Coin, A., & Dubljević, V. (2020). AI in the headlines: The portrayal of the ethical issues of artificial intelligence in the media. AI and Society, 35, 927–936. [11] Borenstein, J., & Howard, A. (2021). Emerging challenges in AI and the need for AI ethics education. AI and Ethics, 1, 61–65. [12] Liebrenz, M., Schleifer, R., Buadze, A., Bhugra, D., & Smith, A. (2023). Generating scholarly content with ChatGPT: Ethical challenges for medical publishing. The Lancet Digital Health, 5(3), e105–e106.
210
[13] [14]
[15] [16]
[17]
[18]
[19] [20]
[21] [22] [23] [24] [25] [26]
[27]
[28]
[29] [30] [31]
[32]
Abhisek Sethy et al.
Garrett, N., Beard, N., & Fiesler, C. (2020, February). More than “If Time Allows” the role of ethics in AI education. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society (pp. 272–278). Murphy, K., Di Ruggiero, E., Upshur, R., Willison, D. J., Malhotra, N., Cai, J. C., . . . & Gibson, J. (2021). Artificial intelligence for good health: A scoping review of the ethics literature. BMC Medical Ethics, 22(1), 1–17. Floridi, L., Cowls, J., King, T. C., & Taddeo, M. (2020). How to design AI for social good: Seven essential factors. Science and Engineering Ethics, 26, 1771–1796. Yadav, S. P., & Yadav, S. (2019). Mathematical implementation of fusion of medical images in continuous wavelet domain. Journal of Advanced Research in Dynamical and Control System, 10(10), 45–54. Logeshwaran, J., Kiruthiga, T., Kannadasan, R., Vijayaraja, L., Alqahtani, A., Alqahtani, N., & Alsulami, A. A. (2023). Smart load-based resource optimization model to enhance the performance of deviceto-device communication in 5G-WPAN. Electronics, 12(8), 1821. Rani, P., Verma, S., Yadav, S. P., Rai, B. K., Naruka, M. S., & Kumar, D. (2022). Simulation of the lightweight blockchain technique based on privacy and security for healthcare data for the cloud system. International Journal of E-Health and Medical Communications (IJEHMC), 13(4), 1–15. Yadav, S. P. (2022). Blockchain security. Blockchain Security in Cloud Computing, 1–18. Pujahari, R. M., Yadav, S. P., & Khan, R. (2022). Intelligent farming system through weather forecast support and crop production. In Application of Machine Learning in Agriculture (pp. 113–130). India: Academic Press. Morris, M. R. (2020). AI and accessibility. Communications of the ACM, 63(6), 35–37. Hagendorff, T. (2020). The ethics of AI ethics: An evaluation of guidelines. Minds and Machines, 30(1), 99–120. Belk, R. (2021). Ethical issues in service robotics and artificial intelligence. The Service Industries Journal, 41(13–14), 860–876. Char, D. S., Abràmoff, M. D., & Feudtner, C. (2020). Identifying ethical considerations for machine learning healthcare applications. The American Journal of Bioethics, 20(11), 7–17. Morley, J., Machado, C. C., Burr, C., Cowls, J., Joshi, I., Taddeo, M., & Floridi, L. (2020). The ethics of AI in health care: A mapping review. Social Science and Medicine, 260, 113172. Subramanian, B., Saravanan, V., Nayak, R. K., Gunasekaran, T., & Hariprasath, S. (2019). Diabetic retinopathy-feature extraction and classification using adaptive super pixel algorithm. International Journal of Engineering and Advanced Technology, 9, 618–627. Singh, J., Deepika, J., Sathyendra Bhat, J., Kumararaja, V., Vikram, R., Jegathesh Amalraj, J., . . . & Sakthivel, S. (2022). Energy-efficient clustering and routing algorithm using hybrid fuzzy with grey wolf optimization in wireless sensor networks. Security and Communication Networks, 2022, 1–10 Tripathy, R., Nayak, R. K., Saravanan, V., Mishra, D., Parasa, G., Das, K., & Das, P. (2021). Spectral clustering based fuzzy C-means algorithm for prediction of membrane cholesterol from ATP-binding cassette transporters. In Intelligent and Cloud Computing: Proceedings of ICICC 2019 (Vol. 2, pp. 439–448). Singapore: Springer. Esmaeilzadeh, P. (2020). Use of AI-based tools for healthcare purposes: A survey study from consumers’ perspectives. BMC Medical Informatics and Decision Making, 20(1), 1–19. Reddy, S., Allan, S., Coghlan, S., & Cooper, P. (2020). A governance model for the application of AI in health care. Journal of the American Medical Informatics Association, 27(3), 491–497. Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). A novel architecture of intelligent decision model for efficient resource allocation in 5G broadband communication networks. ICTACT Journal On Soft Computing, 13(3), 2986–2994. Mittelstadt, B. (2019). Principles alone cannot guarantee ethical AI. Nature Machine Intelligence, 1(11), 501–507.
9 AI: issues, concerns, and ethical considerations
211
[33] Morley, J., Floridi, L., Kinsey, L., & Elhalal, A. (2020). From what to how: An initial review of publicly available AI ethics tools, methods and research to translate principles into practices. Science and Engineering Ethics, 26(4), 2141–2168. [34] Ponnusamy, M., Bedi, P., Suresh, T., Alagarsamy, A., Manikandan, R., & Yuvaraj, N. (2022). Design and analysis of text document clustering using salp swarm algorithm. The Journal of Supercomputing, 78(14), 16197–16213. [35] Kannan, S., Yuvaraj, N., Idrees, B. A., Arulprakash, P., Ranganathan, V., Udayakumar, E., & Dhinakar, P. (2021). Analysis of convolutional recurrent neural network classifier for COVID-19 symptoms over computerised tomography images. International Journal of Computer Applications in Technology, 66 (3–4), 427–432. [36] Manikandan, V., & Yuvaraj, N. (2019, May). Advanced expert system using particle swarm optimization based adaptive network based fuzzy inference system to diagnose the physical constitution of human body. In Emerging Technologies in Computer Engineering: Microservices in Big Data Analytics: Second International Conference, ICETCE 2019, Jaipur, India, February 1–2, 2019, Revised Selected Papers (Vol. 985, p. 349). Springer. [37] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). Energy‐efficient resource allocation model for device‐to‐device communication in 5G wireless personal area networks. International Journal of Communication Systems, 2023, 1–14. [38] Schiff, D., Biddle, J., Borenstein, J., & Laas, K. (2020, February). What’s next for AI ethics, policy, and governance? A global overview. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society (pp. 153–158).
Abhisek Sethy, Nazeer Shaik, V. Saravanan, S. Karthika
10 The future with AI and AI in action Abstract: The future of artificial intelligence (AI) is bright and filled with possibilities. AI has the potential to revolutionize many areas, including medicine, finance, customer service, and transportation. It can help create more efficient systems, enhance decision-making, and ultimately improve people’s lives. With increased computing power and faster networks, AI can become more sophisticated. Eventually, AI could be used to automate mundane and labor-intensive tasks such as filing paperwork or scheduling meetings. AI can also help identify new opportunities for businesses and increase their competitiveness. For example, AI can be used for data mining to allow organizations to make more informed decisions. The AI can be used to create more realistic simulations and virtual environments to help with the development of products, software, and autonomous cars. In the future, AI might even be able to take on more complex tasks such as visual recognition, language translation, and natural language processing. Ultimately, AI will play an increasingly important role in how businesses operate and how people interact with each other. As AI technology evolves, so too will its applications in the future. Keywords: Artificial, intelligence, medicine, finance, customer service, transportation
10.1 Introduction to the future of AI The future of artificial intelligence (AI) is an exciting one. AI has the potential to revolutionize the way we live, work, and play. AI technologies are creating new products and services and transforming existing ones, eliminating mundane tasks, automating complicated processes, speeding up decision-making, and improving customer service. AI can also help us solve some of the world’s most pressing problems, like climate change, healthcare, and poverty. While AI is still in its infancy, experts expect it to become even more powerful in the coming years [1]. As AI technologies become more prevalent, we can expect to see more applications in areas like autonomous vehicles, healthcare, smart homes, finance, and the legal profession. Ultimately, AI will help us lead a better life and make the world a better place.
10.1.1 The rapid advancement of AI technologies AI is an area of computer science that has seen rapid advancements in recent years due to the availability of data, cloud computing, and powerful algorithms. AI technologies are being used to solve a variety of complex problems related to automation, natural https://doi.org/10.1515/9783111323749-010
214
Abhisek Sethy et al.
language processing (NLP), facial recognition, and more. As AI continues to become more powerful, it can be used to solve increasingly complex problems [2]. For example, AI can now detect and respond to malicious online activity, detect objects in images, and automate routine tasks that once required human intervention. By leveraging machine learning (ML) techniques, AI continues to scale and get more efficient, allowing it to tackle ever more complex tasks. AI is being integrated into various aspects of our lives and is quickly transforming how people interact with machines and computers.
10.1.2 Anticipated impact of AI on various industries and sectors AI is increasingly being adopted by businesses as a way of increasing efficiency and reducing costs. AI can be used to automate tasks, improve customer service, and provide insights into customer behavior and preferences. It can also be used to increase productivity and optimize production processes. As AI continues to evolve, it has the potential to revolutionize every sector and industry, leading to massive economic and societal changes. In the transportation sector, AI has already been deployed in various ways. Advanced driver-assistance systems (ADAS) are already being used to avoid collisions and help drivers navigate traffic. Autonomous vehicles are already being tested in limited capacity in some cities, with the hope that they will be available to the public in the near future. In the healthcare sector, AI is being used to help diagnose illnesses, detect cancer and Alzheimer’s, and improve patient outcomes. AI is also being used to create virtual nurses, which can provide personalized advice and healthcare to patients. In the manufacturing sector, AI is being used to optimize production processes, reduce defects, and increase production speeds. AI-powered robots are being used to assemble products, while AI-assisted robots are beginning to be used to perform more dangerous tasks with greater accuracy [3]. In the financial sector, AI is being used to process transactions faster and more efficiently, detect fraudulent activities, and provide tailored investing advice. AI is also being used to automate bookkeeping and compliance tasks, allowing businesses to be more efficient and cost-effective. In the retail sector, AI is being used to create personalized marketing campaigns, identify customer preferences, and recommend tailored product or service offerings. AI is also being used to optimize pricing and predict customer demand, allowing retailers to maximize profits. The impact of AI on these sectors is expected to be significant and far-reaching. With data becoming increasingly accessible and AI developing rapidly, AI-based solutions will become more ubiquitous in the coming years, leading to improved efficiency, better decisionmaking, and higher ROI (return of interest) for businesses.
10 The future with AI and AI in action
215
10.2 AI in healthcare AI in healthcare is the utilization of advanced AI techniques to provide innovative and more efficient medical solutions. AI can help improve patient outcomes, enhance the safety of medical treatments, reduce costs, and improve patient satisfaction. AI can also help healthcare organizations improve diagnostic accuracy, patient flow, provider communication, and operational efficiency of their organizations [4]. AI can be used to identify patterns in medical records, develop better prediction models to diagnose and treat diseases, automate some administrative and care processes, and provide personalized and preventive care. AI also allows for easier data sharing, increases the accuracy of medical records, allows for more effective patient-provider communication, and can provide streamlined medical care coordination.
10.2.1 Applications of AI in healthcare diagnostics and treatment AI is a rapidly growing field and has the potential to revolutionize the healthcare industry. AI can be used in healthcare for a variety of purposes, including diagnostics, prognostics, personalized healthcare, and advanced medical treatments. AI can be employed to detect and diagnose medical conditions, utilize data and predict outcomes, monitor patients, provide personalized treatments, and even suggest the best treatment options for certain diseases. AI can also be used to analyze vast amounts of patient data to improve disease management and help determine the most suitable patient-specific treatments [5]. Additionally, AI can be used to improve the performance and accuracy of medical imaging techniques such as CT scans and MRIs. AI can even be used to plan and deliver targeted drug delivery and treatment. AI can also be used to build intelligent robots and machines to assist surgeons and other medical personnel in performing medical procedures. AI has already been implemented in some hospitals and clinics, and its uses in healthcare are growing fast. As AI-equipped robots, machines and algorithms become increasingly advanced, they can be employed to perform a greater range of tasks, from diagnostics to surgery.
10.2.2 AI-enabled medical imaging and diagnostics AI-enabled medical imaging and diagnostics is the use of AI to analyze medical images, such as MRI scans and X-rays, for the purpose of diagnosing or detecting medical problems. AI-enabled medical imaging and diagnostics can help reduce the time spent manually analyzing images, improve accuracy and reproducibility of diagnoses, and even highlight abnormalities in a patient’s medical images that may otherwise be missed. AI-enabled medical imaging and diagnostics can also reduce errors associated with misinterpretation of medical images [6]. Many large medical institutions have
216
Abhisek Sethy et al.
already implemented AI-enabled medical imaging and diagnostics in their practices, while research continues into further applications of AI to medical imaging, such as 3D reconstruction of medical images as well as automatic segmentation of tumors and other abnormalities.
10.2.3 AI-assisted drug discovery and personalized medicine AI-assisted drug discovery and personalized medicine are two closely related concepts in healthcare that are driven by advances in AI. AI-assisted drug discovery is an emerging field that combines ML and AI to enable drug discovery and development faster and more efficiently. This can be used to identify and develop new treatments for a wide range of illnesses with fewer resources and in less time, leading to fewer expensive failures and a greater likelihood of success. Personalized medicine is an approach to healthcare that uses an individual’s genetic and nongenetic traits to better anticipate, diagnose, and treat various illnesses [7]. This approach can provide personalized treatments to fit an individual’s specific needs, which can improve patient outcomes and lead to lower health costs. AI-assisted drug discovery can be combined with personalized medicine to identify treatments that are tailored to each individual patient, allowing for improved efficacy and better patient outcomes.
10.2.4 Ethical considerations and challenges in AI healthcare applications Ethical considerations and challenges in AI healthcare applications mostly revolve around data collection and privacy, patient safety, and transparency. The ethical considerations in AI healthcare applications are mainly concerned with data privacy, usage, and sharing. A key concern is that the patient’s personal data is not being abused or misused. Additionally, ethical considerations when using AI in healthcare involve ensuring that the AI is safe for the patients, being transparent on how decisions are being made, and understanding the implications of AI-based system on healthcare access and quality. Data privacy and usage is an important ethical consideration when using AI healthcare applications. In order for an AI healthcare system to be effective, it needs to be able to collect and analyze large amounts of patient data. This data must be both secure and anonymous so that the patient’s privacy is respected. Additionally, the data must be ethically collected, used, and shared among healthcare providers [8]. Patient safety is also an important ethical consideration when using AI healthcare applications. AI healthcare applications must be tested and monitored to make sure that they are accurate and that they do not put patients in any risk. Additionally, AI healthcare systems must be designed with safeguards to protect against incorrect diagnoses and outcomes. There must be transparency regarding
10 The future with AI and AI in action
217
the decision-making process of the AI healthcare application. It must be made clear how the AI is making decisions and why certain decisions were made. This transparency is necessary in order to ensure trust in the system and that the AI is making the best possible decisions for the patient’s well-being.
10.3 AI in finance AI in finance has become a major factor in how companies, financial institutions, and organizations manage their resources. AI provides a more efficient way to make decisions and process data, allowing financial institutions to identify and seize market opportunities. AI can also be used to automate mundane and time-consuming tasks in the financial industry, such as compliance and fraud prevention, capital management, and portfolio management. Additionally, AI has been used to create algorithmic trading strategies, simulate risk scenarios, and develop optimized stock portfolios. AI can also be used to automate customer service and provide predictive analytics for customer segmentation and personalization. AI is quickly becoming an essential tool for financial institutions to stay competitive and maximize their profits.
10.3.1 AI applications in financial forecasting and risk management AI applications are increasingly being used in financial forecasting and risk management. AI can help financial forecasters and risk managers to make better decisions based on past data. AI can also be used to analyze a large amount of data quickly and accurately, providing more accurate and timely predictions compared to traditional analysis methods [9]. AI can also help identify and analyze trends in financial data and can be used to detect fraud and help detect and prevent money laundering. AI can also be used to create more reliable financial models, which can help financial managers to better assess risks and make informed decisions. AI can help financial managers automate processes and identify new opportunities for investments.
10.3.2 Algorithmic trading and AI-driven investment strategies Algorithmic trading and AI-driven investment strategies are types of computer-based strategies used by financial firms to improve the speed, accuracy, or efficiency of investing decisions. Algorithmic trading and AI-driven strategies rely on computer algorithms and AI to analyze large amounts of data and use it to make decisions on where and when to trade. Algorithmic systems are also used to monitor and adjust positions,
218
Abhisek Sethy et al.
analyze market conditions and trends, and identify potential trading opportunities. AI algorithms can also assist with optimizing allocations to help maximize profits and reduce risk [10].
10.3.3 Fraud detection and prevention using AI techniques Fraud detection and prevention using AI techniques is a popular technology used in many industries. Fraudsters can use sophisticated tactics to commit fraud, so it is important for companies to use the latest methods and technologies to detect and prevent fraud. AI is proving to be an effective tool in this regard as it can detect patterns and anomalies in large data sets that may not be apparent to the human eye. AI technologies can also be used to analyze large amounts of data to detect fraudulent transactions quickly and efficiently. Additionally, AI techniques can be used to identify potential fraudulent activity in real time based on changes in data patterns, customer behavior, and other factors. This makes it easier to spot fraud sooner and take action to prevent it.
10.3.4 Regulatory considerations and challenges in AI finance Regulatory considerations and challenges in AI Finance (AI FinReg) are the legal, regulatory, and ethical requirements and potential implications associated with the use of AI in finance. AI-based financial systems create new risks associated with transparency, fairness, reliability, safety, and security. Financial regulators are beginning to develop regulations to address these risks. For instance, the European Union’s General Data Protection Regulation (GDPR) provides guidelines for data protection and privacy considerations when developing and deploying AI-based financial services. In the United States, the Office of the Comptroller of the Currency (OCC) has developed an anticipatory framework for handling regulatory issues associated with the use of AI in financial services [11]. In addition, regulators are actively engaged in conversations and consultations with industry stakeholders to discuss the use of AI in financial services. These conversations often reveal potential blind spots in the existing regulatory framework and suggest potential solutions. Despite these proactive measures, the regulatory landscape is difficult to navigate because AI technology is still so new and dynamic. Regulators will continue to assess financial services driven by AI, so the exact regulations and enforcement of these regulations are continuously evolving. As such, it is important for businesses to remain abreast of relevant regulations and stay connected with industry leaders to ensure compliance.
10 The future with AI and AI in action
219
10.4 AI in transportation AI is increasingly being used in the transportation sector to create smarter and more efficient transportation networks. AI can be used to enhance safety, reduce traffic accidents, expand capacity, improve customer experience, and reduce costs. It can help enable autonomous, self-driving vehicles; optimize route planning and maintenance, and facilitate the integration of alternative energy sources. AI has also been used to analyze data collected from vehicle sensors, detect traffic patterns, and predict future conditions. By leveraging AI-based algorithms, transportation networks can be made more reliable, efficient, and environmentally friendly.
10.4.1 Autonomous vehicles and self-driving technologies Autonomous vehicles and self-driving technologies are the future of transportation. Autonomous vehicles are able to operate without a human driver and with minimal human intervention. They are equipped with advanced sensors, AI, and a suite of other technologies that allow them to navigate both on-road and off-road. Self-driving technologies such as AI, ML, computer vision, and radar technology are used to help the vehicle detect and navigate its environment. The vehicle is then able to make decisions based on what it senses and plan a safe path of travel [12]. Autonomous vehicles offer the potential to reduce car accidents, free up passengers’ time, and improve overall living conditions through expanded transportation options.
10.4.2 AI-based traffic management and optimization AI-based traffic management and optimization makes use of AI algorithms to optimize and manage traffic on roads and highways. It is designed to improve road safety, reduce traffic congestion and pollution, and reduce costs for individuals and businesses. AI-based traffic management and optimization systems may include algorithms that monitor traffic flow, predict roadways with greater than usual congestion, plan and adjust routes, monitor speed limits, and notify drivers of possible delays. These systems also help assign resources such as personnel and equipment to areas of need in order to reduce congestion [13]. Additionally, the data collected from these traffic management systems can help cities and local governments plan and develop new strategies for improving transportation in their areas.
220
Abhisek Sethy et al.
10.4.3 Predictive maintenance and AI in logistics and supply chain management Predictive maintenance and AI in logistics and supply chain management are critical components of a supply chain management system. These tools help warehouse managers, shippers, and other supply chain stakeholders optimize operations and manage their supply chain more efficiently. Predictive maintenance solutions analyze data from warehouse operations, shipments, or other sources to identify breakdowns in operations and anticipate them before they cause costly delays or costly repairs. AI in logistics and supply chain management can be used to identify patterns in customer orders, plan routes, and make recommendations for optimizing transportation costs and delivery times. This technology can also be used to monitor inventory and optimize replenishment [14]. AI can also be used to automate many of the tasks in order processing, which can help speed up orders and reduce manual errors. Additionally, AI can be used to optimize the delivery chain, including load consolidation and timely loading or scheduling of shipments. AI can also reduce paperwork and make warehouse operations more efficient and accurate.
10.4.4 Safety and ethical considerations in AI transportation systems Safety and ethical considerations need to be considered when using AI transportation systems as they are meant to support and improve the safety of vehicles and their occupants. AI transportation systems work with data-driven algorithms that make decisions on behalf of the driver. As a result, ethical considerations should always be factored into the design of these systems to ensure that they are operating with an appropriate level of trust, reliability, accountability, and privacy. Safety considerations should also be taken into account when using AI transportation systems. This includes making sure that the system is able to detect and respond to potential hazards and that it can be paused or disabled if necessary [15]. Additionally, it is important to consider how well these systems can evaluate driver behavior and intervene if necessary, as well as how useful AI-driven warnings and suggestions are and how easy they are for drivers to understand. In terms of ethical considerations, it is important to consider how the decisions taken by AI-driven systems could affect individuals, governances, and society, and how transparent the system is in communicating their decisions. Questions of fairness and accountability must also be addressed and the data used in the training of AI-driven systems should be monitored and managed carefully to ensure that it is free from any bias.
10 The future with AI and AI in action
221
10.5 AI in manufacturing and robotics AI in manufacturing and robotics refers to a wide range of technologies and approaches that use AI techniques such as ML and NLP to automate and augment operations in the manufacturing industry. Some of these technologies, such as robotic process automation (RPA), are being used to automate complex production and assembly processes. Other AI-driven technologies, such as computer vision and NLP, are being used to enable autonomous robots and unmanned aerial vehicles (UAVs). AI is also being used in the manufacturing industry to improve existing processes, such as predictive analytics, anomaly detection, and prescriptive analytics [16]. Additionally, AI-enabled tools, such as computer-aided design (CAD) and autonomous robots, are being used to improve quality assurance and reduce costs.
10.5.1 AI-driven automation and robotics in manufacturing AI-driven automation and robotics in manufacturing is the use of automated robotics equipped with AI to improve the production processes. This technology is used to automate mundane and laborious tasks and can also be applied to improve the performance of existing production processes by optimizing the scheduling of robot movements and improving the precision of execution. The use of AI-driven automation and robotics can reduce the amount of manual labor required in manufacturing while increasing the quality and efficiency of production [17]. AI-driven automation and robotics can also measure and analyze product and process data to identify trends and make adjustments for process improvement. With the incorporation of AI, robots can automatically identify patterns, find solutions, and take corrective actions to avoid errors, and prevent costly production delays.
10.5.2 Quality control and defect detection using AI Quality control and defect detection using AI is a process aimed at ensuring that manufactured products or services meet the required specifications. AI-based solutions can help identify quality and defect problems quickly and accurately, reducing the time and cost associated with conventional methods. AI-based quality control tools incorporate object recognition algorithms that can autonomously process the visualized output of digital inspection equipment and compare it against predefined criteria in order to detect defects. AI-based systems can also learn over time from production data and detect patterns in the occurrence of quality or defects across products and processes, leading to improved process control and quality assurance.
222
Abhisek Sethy et al.
10.5.3 Collaborative robots (cobots) and human-robot interaction A collaborative robot, also known as cobot, is an intelligent robotic platform equipped with sensors and other features to allow it to interact with humans in a shared workspace in a safe and effective manner. The cobots can be used for various applications ranging from industrial manufacturing to healthcare, retail, construction, logistics, and other industries. They can also be used in hybrid systems, where humans and robots work alongside each other to optimize production processes. In order to ensure safety and effectiveness of their operations, cobots are generally programmed with safety protocols to protect those that operate them. Human-robot interaction (HRI) is the way in which humans communicate, interact with, and collaborate with robots. HRI plays an important role in the development and deployment of any type of service and industrial robot, especially those that are meant to be used with humans in close proximity [18]. Different forms of HRI can be implemented in order to optimize safety, performance, and productivity. These forms include robotic teach and playback, tactile feedback, voice or visual recognition, and haptic interfaces.
10.5.4 Workforce implications and ethical considerations in AI manufacturing Workforce implications of AI in manufacturing depend largely on the types of jobs that will be created by its implementation. Some job functions, such as those requiring physical labor, may be replaced entirely by AI solutions while other job functions such as those requiring higher-level skills may be augmented or enhanced. As AI continues to develop, it is likely that current job functions may become obsolete as new job functions are created to support AI operations [19]. Ethical considerations in AI manufacturing involve the use and deployment of AI solutions, which leads to questions such as who has the right to control and benefit from the technology, what are the implications for workers’ rights and safety, and how does one ensure responsible use of AI systems. While there are many ethical considerations to be made around the use of AI in manufacturing, there are also potential advantages, such as improved quality control, enhanced production capabilities, and improved customer service.
10.6 AI in customer experience and marketing AI is becoming increasingly important in customer experience (CX) and marketing. AI can help businesses capture, interpret, and use customer data more accurately and quickly than ever before. AI technologies such as NLP, sentiment analysis, predictive
10 The future with AI and AI in action
223
analytics, and ML can provide businesses with data-driven insights into customer preferences and trends that help marketers develop more effective strategies. AI can also be used to customize customer experiences based on specific needs and interests, making them more personalized and engaging. Additionally, AI can automate marketing tasks such as segmentation, targeting, and personalization, saving businesses time and resources. By leveraging the power of AI, businesses can drive customer engagement and loyalty while creating more efficient and effective marketing strategies.
10.6.1 Personalization and recommendation systems powered by AI Personalization and recommendation systems powered by AI allow websites and applications to provide users with tailored offers, content, and product recommendations that are tailored to each individual user. By leveraging user data, such as search queries, purchase history, and past interactions, AI can power personalized content and recommendations that are better and quicker than manual processes [20]. AI-powered systems can also use collaboration filtering techniques to group together people with similar interests and offer tailored recommendations. This kind of personalization can help increase user engagement and satisfaction, as customers feel like their needs are being better understood.
10.6.2 Sentiment analysis and AI in customer sentiment tracking Sentiment analysis is a technique used in the field of NLP to identify the opinion or attitude of a speaker. AI-driven sentiment analysis works by interpreting text or audio data to detect underlying sentiment. In customer sentiment tracking, AI-driven sentiment analysis can be used to quickly and accurately identify customer attitudes toward a brand or product. AI’s ability to quickly process large amounts of data from multiple sources allows for detailed analysis and reporting to be generated within minutes. For example, customer sentiment tracking can be used to analyze customer feedback from social media posts, product and service reviews, surveys, and more to gain an understanding of customer attitudes [21]. AI-driven sentiment analysis can also detect sentiment in real time, allowing for more effective customer service and issue resolution. By combining sentiment analysis with AI-driven algorithms and ML, businesses are able to quickly identify customer pain points, complaints, and feedback. This data can then be used to inform product or service changes, customer service strategies, and marketing campaigns. Ultimately, AI-driven sentiment analysis is an invaluable tool for businesses aiming to provide better customer experiences.
224
Abhisek Sethy et al.
10.6.3 AI-powered chatbots and virtual assistants A chatbot (also known as a virtual assistant) is a computer program that uses AI to simulate human conversation. It can help automate customer service, sales, marketing, and other tasks. It can act as a virtual assistant to conduct conversations and learn user preferences in order to provide personalized, context-aware recommendations and information. It can also be used to generate leads, respond to customer queries, take surveys, or provide product recommendations. Many companies are using AI-powered chatbots and virtual assistants to improve customer experience, drive engagement, and provide valuable insights [22].
10.6.4 Privacy concerns and ethical considerations in AI marketing Privacy concerns and ethical considerations in AI marketing involve the use of AI to collect, analyze, and use human data to target ads, create personalized experiences, or otherwise influence purchase decisions. With AI marketing, there may be a lack of transparency around how data is collected, stored, and used, with possible violations of personal privacy. Additionally, AI can generate biased or inaccurate results due to racial, gender, or other stereotypes embedded in the algorithms used to create marketing profiles. This may lead to customers being targeted for unwanted or irrelevant products or services, creating potential ethical issues. Other ethical concerns may arise when AI is used to create predictive models or influence shopping habits, with customers being duped into making purchases that they may not have otherwise made [23].
10.7 AI in education AI in education is the use of AI technologies in the classroom. AI technologies can be used to personalize learning experiences, improve student engagement, and automate routine activities. AI can also be used to support teaching by providing real-time feedback to students on their progress and identifying areas of improvement. It can also help instructors be more efficient and provide a better learning environment for students. AI can also be used to create digital tutors and virtual classrooms that assist in delivering personalized instruction. AI-driven technologies such as bots, virtual assistants, and intelligent tutoring systems allow for more efficient learning and knowledge acquisition of key educational concepts.
10 The future with AI and AI in action
225
10.7.1 Adaptive learning platforms and AI tutors Adaptive learning platforms are systems that use AI to determine when and how to deliver content to students. AI tutors use NLP and ML techniques to understand a student’s current abilities, goals, and educational preferences. They then generate assignments, activities, and resources that are tailored to the student’s individual needs. AI tutors can deliver content in various forms, such as videos, interactive games, online worksheets, and more. In addition, they provide feedback, take tests, and monitor student progress and engagement. AI tutors are used to supplement or replace traditional instructors, allowing for personalized learning experiences and real-time personalized instruction.
10.7.2 Intelligent educational content generation Intelligent educational content generation is the process of automating the creation of educational content, using AI techniques in order to create unique, personalized content for educational use. This can enable the creation of targeted educational content that is tailored to the needs of each individual learner in a more efficient and costeffective manner. AI can be used to organize and structure educational material into a series of activities or lessons, generate new content, and identify and recommend appropriate material for review and further study. AI technologies such as NLP, speech recognition, ML, search algorithms, and computer vision may be used to better understand what material an individual may be looking for and make recommendations accordingly.
10.7.3 AI-enabled student assessment and performance tracking AI-enabled student assessment and performance tracking is a technology that enables educators to measure and track student performance by using AI. AI algorithms are used to analyze students’ performance data such as test scores, assignments, and even classroom participation to measure student progress over time and identify areas where students need the most help. This technology can also be used to tailor instruction based on individual student needs and provide personalized guidance to students to help them achieve their academic goals [24]. AI-enabled student assessment and performance tracking not only allows educators to quickly assess student progress but also provides them with meaningful insights to inform instruction.
226
Abhisek Sethy et al.
10.7.4 Ethical considerations and challenges in AI education Ethical considerations and challenges in AI education refer to the ethical implications associated with the development, use, and teaching of AI. These ethical considerations involve questions of fairness, trust, privacy, security, and the role of humans in a society increasingly occupied by intelligent machines. AI education must ensure that students learn about these ethical issues and that things such as bias and potential for misuse are addressed. Additionally, educators must be aware that AI can cause unintended consequences and must be careful to explain the implications of AI’s use. Furthermore, educators must be alert to potential ethical issues when teaching AI due to its potential to automate certain tasks and take the place of humans in many areas. Lastly, teachers must develop methods for teaching AI that are safe and transparent to ensure that students are aware of their own bias and the potential risks of AI.
10.8 AI in agriculture and food industry AI is increasingly being used in the agriculture and food industry to increase efficiency by automating processes and improving overall production throughout the supply chain. AI can help identify problems in the supply chain that humans may not have noticed, allowing farmers to adjust and improve their processes accordingly. AI can also help increase crop yields and reduce food waste, which can help reduce global poverty. By improving the accuracy of inventory management, AI can also reduce food contamination and help combat foodborne illnesses. AI can also help with more accurate labeling of food products so that customers can select healthy options. Finally, AI can help farmers predict weather patterns more accurately, allowing them to adjust their farming practices accordingly.
10.8.1 Precision farming and AI-driven crop management Precision farming is a farming management concept based on observing, measuring, and responding to inter- and intra-field variability in crops. It uses technology such as GPS, remote sensing, tractor-mounted sensors, and computer modelling to identify and respond to variability in crops. AI-driven crop management leverages the use of AI and ML technologies to help farmers monitor, analyze, and manage their crops in real time [25]. The technology can process large amounts of data rapidly and accurately. This helps farmers improve the yield and quality of their crops, while reducing labor, cost, and environmental impact. AI-driven crop management systems enable farmers to collect data from sensors on their fields, and using ML algorithms, provide valuable insights into the health of
10 The future with AI and AI in action
227
their crops. Farmers can adjust their irrigation, fertilization, and management strategies accordingly, to ensure that their crops are not only healthy but also perform optimally.
10.8.2 Pest detection and disease monitoring using AI Pest detection and disease monitoring using AI is a process whereby AI is used to detect, diagnose, and track pests and diseases in agriculture. This can be done in a variety of ways. AI can gather data about pests and diseases from multiple sources and use the data to identify potential outbreaks and track changes in the environment. It can also use algorithms to identify patterns in the data, helping to identify when action needs to be taken. AI can also provide detailed maps of the risk factors associated with a particular pest or disease and help develop strategies for prevention and management.
10.8.3 AI applications in food processing and quality control AI applications in food processing and quality control involve the use of AI-powered analytics, such as computer vision and ML to automate and optimize the processes of food production and inspection. For example, AI can be used to detect quality anomalies in food products and enable faster and more accurate quality decision-making. AI can also be used to automate the production process and reduce labor costs. Additionally, AI-driven analytics can be used to improve predictive analytics and supply chain performance. AI can help reduce food waste by detecting anomalies in production operations and making sure only the highest quality products reach consumers.
10.8.4 Sustainability and ethical considerations in AI agriculture Sustainability and ethical considerations in AI agriculture are two very important aspects when looking to utilize AI to improve agricultural practices. AI-powered solutions can provide increased efficiency, decreased cost, and improved yields, but it is critical to consider the environmental and ethical implications of these decisions. First, it is important to consider how the implementation of AI will impact the environment in terms of energy use, water use, and soil conservation. Additionally, ethical considerations such as labor practices, human rights, and animal welfare must be taken into account when deciding how best to deploy AI-enabled technologies [26]. Finally, governments and stakeholders must come together to create rules and regulations that will guarantee the ethical use of AI in the agricultural sector. Doing so will ensure that the application of AI-powered solutions will benefit farmers and
228
Abhisek Sethy et al.
the environment while allowing AI-powered systems to continue to transform the way agriculture is operated.
10.9 AI in smart cities AI in smart cities refers to the use of AI techniques such as ML, deep learning, NLP, and robotics to enable automated control and monitoring of the different aspects of city operations. AI technologies can enable cities to become smarter by providing automated support for decision-making, by recognizing and responding to the needs of citizens, by optimizing the operation and maintenance of networks, and by making processes more efficient and cost-effective [27]. AI can help manage things such as water consumption, energy consumption, flow of traffic and transportation, and public safety. It can also make citizens’ lives easier by providing personalized services such as smart home systems and virtual assistants. AI-based services and applications can ultimately lead to a more efficient, sustainable, and livable city.
10.9.1 AI for urban planning and resource optimization AI for urban planning and resource optimization involves using AI algorithms and systems to improve the planning and management of urban areas and to optimize the use of resources. The goal is to create more efficient and sustainable cities. This includes the use of ML algorithms and models to analyze population size, zoning laws, transit patterns, economic development, and other factors to determine the best way to plan and build a city. AI can also be used to streamline and optimize the allocation of resources to areas with a higher demand. This can then help reduce costs and provide better quality of life to citizens. Additionally, AI can reduce and minimize the amount of waste produced by cities by optimizing processes like energy consumption, water management, and waste management.
10.9.2 Intelligent transportation systems in smart cities Intelligent transportation systems (ITS) are information and communication technologies applied to transportation and traffic management. With the help of near realtime information and automated location-based management, ITS can help reduce traffic congestion, improve safety, and provide a better user experience. ITS technologies are often used in smart cities to improve traffic flow, reduce emissions, and reduce the cost of travel. Examples of ITS technology include traffic lights, adaptive signal timing, variable speed limit signs, and traffic enforcement cameras. Addition-
10 The future with AI and AI in action
229
ally, ITS systems can predict traffic route recommendations, route guidance, and trip times. By helping optimize traffic flow, reduce emissions, and save time and money, ITS is a key component in creating sustainable and dynamic smart cities.
10.9.3 AI-powered energy management and sustainability AI-powered energy management and sustainability is a term used to refer to the use of AI and ML algorithms to empower organizations to become more responsible and use resources more efficiently. AI-powered energy management and sustainability allows businesses to optimize their processes by analyzing a variety of data points related to energy consumption and corporate sustainability. The result is a system that can identify and implement better practices that reduce energy costs and minimize environmental impact [28]. AI-powered energy management and sustainability solutions provide companies with tremendous control over usage and consumption of energy and resources. This type of solution can be used to monitor and track energy usage, environmental disturbance, and irregularities in energy usage from areas without proper tracking. By monitoring these metrics, businesses can identify potential problems and resolve them in a timely fashion. Additionally, AI-powered energy management and sustainability solutions can help organizations optimize energy usage and consumption preferences to maximize energy efficiency and reduce costs. AI-powered energy management and sustainability solutions can also provide global oversight that aids in regulatory compliance, which allows businesses to be proactive rather than reactive. Such a system can enable companies to implement new strategies for reducing energy consumption while working to achieve corporate sustainability objectives. AI-powered energy management and sustainability solutions are becoming increasingly popular due to their comprehensive features and scalability [29].
10.9.4 Privacy, security, and ethical concerns in AI smart cities Privacy, security, and ethical concerns in AI smart cities are issues that need to be addressed as cities become increasingly reliant on technology to provide services such as transportation and waste management. With the increased reliance on AI, ethical concerns have arisen, such as potential impacts on human rights, privacy, data protection, algorithmic accountability, and market competition. Ethical issues are particularly important, as AI systems can have significant effects on people’s lives [30]. The implications of AI and the technology that is used to create these cities must be studied to assess whether or not they are adequately protecting individuals’ privacy and other human rights. A key concern is the accuracy of data used by AI, as
230
Abhisek Sethy et al.
inaccurate data can lead to biased decisions that could affect people’s lives [31]. Additionally, AI algorithms have the potential to strengthen existing inequality by amplifying existing patterns of discrimination or favoring certain user interests. This bias in decision-making could further marginalize certain groups if not considered early on. In terms of security, AI smart cities are particularly vulnerable to attack due to the reliance on digital technologies and the speed at which information is relayed [32]. As a result, cities will need to develop extensive security systems to protect themselves from malicious actors using cyberattacks to disrupt services, steal data, or create security vulnerabilities. Ultimately, AI smart cities have the potential to provide significant benefits for their citizens as long as they consider the privacy, security, and ethical implications of their technology. It is vital that cities plan carefully in order to ensure that their AI algorithms and systems are fair and accountable, while simultaneously protecting the privacy of individuals and creating secure networks [33].
References [1]
Haenlein, M., & Kaplan, A. (2019). A brief history of artificial intelligence: On the past, present, and future of artificial intelligence. California Management Review, 61(4), 5–14. [2] Zhang, C., & Lu, Y. (2021). Study on artificial intelligence: The state of the art and future prospects. Journal of Industrial Information Integration, 23, 100224. [3] Schwalbe, N., & Wahl, B. (2020). Artificial intelligence and the future of global health. The Lancet, 395(10236): 1579–1586. [4] Schiff, D. (2021). Out of the laboratory and into the classroom: The future of artificial intelligence in education. AI and Society, 36(1): 331–348. [5] Belfield, H. (2020, February). Activism by the AI community: Analysing recent achievements and future prospects. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society (pp. 15–21). [6] Biswas, S., Carson, B., Chung, V., Singh, S., & Thomas, R. (2020). AI-Bank of the future: Can banks meet the AI challenge. New York: McKinsey & Company. [7] Cave, S. (2020, February). The problem with intelligence: Its value-laden history and the future of AI. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society (pp. 29–35). [8] Sur, J., Bose, S., Khan, F., Dewangan, D., Sawriya, E., & Roul, A. (2020). Knowledge, attitudes, and perceptions regarding the future of artificial intelligence in oral radiology in India: A survey. Imaging Science in Dentistry, 50(3): 193. [9] Aggarwal, K., Mijwil, M. M., Al-Mistarehi, A. H., Alomari, S., Gök, M., Alaabdin, A. M. Z., & Abdulrhman, S. H. (2022). Has the future started? The current growth of artificial intelligence, machine learning, and deep learning. Iraqi Journal for Computer Science and Mathematics, 3(1): 115–123. [10] Sangeetha, R., Logeshwaran, J., Rocher, J., & Lloret, J. (2023). An improved agro deep learning model for detection of Panama Wilts disease in banana leaves. AgriEngineering, 5(2), 660–679. [11] Singh, P. K., Chauhan, S. S., Sharma, A., Prakash, S., & Singh, Y. (2023). Prediction of higher heating values based on imminent analysis by using regression analysis and artificial neural network for bioenergy resources. In Proceedings of the Institution of Mechanical Engineers, Part E: Journal of Process Mechanical Engineering. 09544089231175046.
10 The future with AI and AI in action
[12]
[13] [14]
[15]
[16] [17] [18]
[19] [20]
[21]
[22]
[23] [24]
[25]
[26] [27]
[28]
231
Chauhan, S. S., Yadav, S. P., Awashthi, S., & Naruka, M. S. (2023). Sentimental analysis using cloud dictionary and machine learning approach. In: Kaushal Kishor, Neetesh Saxena & Dilkeshwar Pandey. Cloud-based Intelligent Informative Engineering for Society 5.0 (pp. 157–169). Boca Raton: Chapman and Hall/CRC. Awasthi, S., Naruka, M. S., Yadav, S. P., & De Albuquerque, V. H. C. (Eds.). (2023). AI and IoT-based Intelligent Health Care & Sanitation. United Arab Emirates: Bentham Science Publishers. Yadav, S. P., Zaidi, S., Nascimento, C. D. S., de Albuquerque, V. H. C., & Chauhan, S. S. (2023, January). Analysis and design of automatically generating for GPS Based Moving Object Tracking System. In 2023 International Conference on Artificial Intelligence and Smart Communication (AISC) (pp. 1–5). IEEE. Saravanan, V., Sumathi, A., Shanthana, S., & Rizvana, M. (2013). Dual mode mpeg steganography scheme for mobile and fixed devices. International Journal of Engineering Research and Development, 6, 23–27. Cihon, P., Maas, M. M., & Kemp, L. (2020). Fragmentation and the future: Investigating architectures for international AI governance. Global Policy, 11(5): 545–556. Stahl, B. C. (2021). Artificial intelligence for a better future: An ecosystem perspective on the ethics of AI and emerging digital technologies (p. 124). New York: Springer Nature. Allam, Z., Dey, G., & Jones, D. S. (2020). Artificial intelligence (AI) provided early detection of the coronavirus (COVID-19) in China and will influence future Urban health policy internationally. Ai, 1(2): 156–165. Saravanan, V., & Chandrasekar, C. (2012). Qos-continuous live media streaming in mobile environment using vbr and edge network. International Journal of Computer Applications, 53(6), 1–3. Lakshminarayanan, R., & Yuvaraj, N. (2020). Design and analysis of an improved deep learning algorithm on classification of intervertebral discs. International Journal of Advanced Science and Technology, 29, 4019–4026. Gupta, S., Saravanan, V., Choudhury, A., Alqahtani, A., Abonazel, M. R., & Babu, K. S. (2022). Supervised computer-aided diagnosis (CAD) methods for classifying Alzheimer’s disease-based neurodegenerative disorders. Computational and Mathematical Methods in Medicine, 2022, 1–10 Indhumathi, R., Amuthabala, K., Kiruthiga, G., Yuvaraj, N., & Pandey, A. (2023). Design of task scheduling and fault tolerance mechanism based on GWO algorithm for attaining better QoS in cloud system. Wireless Personal Communications, 128(4): 2811–2829. Zhang, K., & Aslan, A. B. (2021). AI technologies for education: Recent research & future directions. Computers and Education: Artificial Intelligence, 2, 100025. Vaishya, R., Javaid, M., Khan, I. H., & Haleem, A. (2020). Artificial Intelligence (AI) applications for COVID-19 pandemic. Diabetes and Metabolic Syndrome: Clinical Research and Reviews, 14(4): 337–339. Aruna, R. D., Surendran, S. D., Yuvaraj, N. D., & Debtera, B. (2022). An enhancement on convolutional Artificial Intelligent based diagnosis for skin disease using nanotechnology sensors. Computational Intelligence and Neuroscience, 2022, 1–7. Berente, N., Gu, B., Recker, J., & Santhanam, R. (2021). Managing artificial intelligence. MIS Quarterly, 45(3), 1–18. Yuvaraj, N., Praghash, K., & Karthikeyan, T. (2022). Privacy preservation of the user data and properly balancing between privacy and utility. International Journal of Business Intelligence and Data Mining, 20(4): 394–411. Ramesh, G., Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). Prediction of energy production level in large PV plants through AUTO-Encoder Based Neural-Network (AUTO-NN) with restricted Boltzmann feature extraction. Future Internet, 15(2), 46.
232
Abhisek Sethy et al.
[29] Saravanan, V., Sreelatha, P., Atyam, N. R., Madiajagan, M., Saravanan, D., & Sultana, H. P. (2023). Design of deep learning model for radio resource allocation in 5G for massive IoT device. Sustainable Energy Technologies and Assessments, 56, 103054. [30] Borges, A. F., Laurindo, F. J., Spínola, M. M., Gonçalves, R. F., & Mattos, C. A. (2021). The strategic use of artificial intelligence in the digital era: Systematic literature review and future research directions. International Journal of Information Management, 57, 102225. [31] Lu, Y. (2019). Artificial intelligence: A survey on evolution, models, applications and future trends. Journal of Management Analytics, 6(1): 1–29. [32] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). L-RUBI: An efficient load-based resource utilization algorithm for bi-partite scatternet in wireless personal area networks. International Journal of Communication Systems, 36(4), 1–17. [33] Fouad, H., Hassanein, A. S., Soliman, A. M., & Al-Feel, H. (2020). Analyzing patient health information based on IoT sensor with AI for improving patient assistance in the future direction. Measurement, 159, 107757.
S. P. Sreekala, S. Revathy, S. Rajeshwari, B. Girimurugan
11 A survey of AI in industry: from basic concepts to industrial and business applications Abstract: Artificial intelligence (AI) has the potential to revolutionize nearly every industry, and is already making a significant impact on many. AI in the enterprise has proven to be a major influence on the way products and services are developed, delivered, and sold. AI has helped to automate mundane manual processes, as well as enable companies to provide personalized customer experiences. In the manufacturing industry, AI has enabled robots to replace manual labor, resulting in increased output and improved efficiency. In the retail sector, AI is being used for product recommendations, increased customer personalization, and predictive analytics for inventory management. AI is also being used in the healthcare industry to improve patient diagnosis, predict medical outcomes, and analyze images for potential anomalies. AI has made many areas of industry more efficient and has opened up the possibility of many more advancements. Keywords: Artificial, intelligence, industry, service, robots
11.1 Introduction to AI in industry AI in industry involves the implementation of artificial intelligence (AI) technologies and applications in various business scenarios for improved efficiency, accuracy, cost savings, and operational lifecycle [1]. AI is widely used in industry across a range of sectors including marketing, customer service, manufacturing, logistics, healthcare and finance, among others. Its applications offer solutions for automation of mundane tasks, improved customer experience, fraud and anomaly detection, predictive analytics, and more. AI-driven automation can empower companies to streamline their entire process operations and provide actionable insights for faster decision-making. AI also manages large and complex datasets and drives user engagement, resulting in more comprehensive business intelligence.
11.1.1 Overview of the chapter scope and objectives AI is playing a crucial role in today’s industrial environment. AI is a broad term that describes computer systems designed to learn, understand, reason, and act autonomously to solve complex problems. It has been widely adopted in various industries https://doi.org/10.1515/9783111323749-011
234
S. P. Sreekala et al.
for a variety of applications, such as automatic production line optimization, predictive maintenance, process automation, and supply chain forecasting [2]. The scope and objectives of AI in industry are to increase operational efficiency, enhance customer experiences, and reduce costs. AI can automate complex processes and tasks across multiple departments and optimize the workflow. It can help with predictive analytics, allowing organizations to make accurate forecasts for better business decisions. AI can also improve customer experience by automating customer service tasks, such as answering common customer questions. Last but not least, AI can reduce costs by implementing automated systems that can replace human labor.
11.1.2 Importance of AI in industrial and business settings The use of AI in industrial and business settings is rapidly growing and becoming increasingly important due to the potential it offers for improving efficiency, productivity, and cost savings. AI can match the demand for personalized services and tailored products that are tailored to the specific needs of customers. AI also helps in making more accurate analyses and forecasts, thereby helping to identify areas for improvement and avoid potential risks. Moreover, AI can speed up decision-making, automate operations, and provide guidance on areas that could be improved [3]. AI is used in a number of industries, including finance and healthcare, and a wide range of business functions such as marketing, human resources, and operations. AI is becoming ubiquitous in businesses and industries as it has the potential to revolutionize the way companies conduct their business.
11.2 Fundamentals of AI for industry AI for Industry is the application of AI technologies to industrial processes. It involves automating the aspects of a process, such as decision making, workload optimization, predictive analysis, and machine learning. AI for Industry can increase efficiency, save costs, and improve accuracy in areas such as customer service, supply chain management, and operations. Additionally, AI can be used to enhance processes in safety, security, and compliance. AI can help companies analyze data, gain insights, and develop strategies to stay competitive. AI can also play a role in developing new products and services [4]. AI techniques such as deep learning and reinforcement learning (RL) can be used to improve decision-making, anticipate customer needs, and create predictive models. AI for industry can also provide personalized customer experiences and enable businesses to provide better customer service.
11 A survey of AI in industry
235
11.2.1 Brief overview of AI principles and techniques AI principles and techniques refer to the specific ways that AI can be used to solve problems and provide solutions. These principles and techniques can range from algorithms, machine learning, natural language processing (NLP), robotics, and more [5]. The principles of AI, such as problem solving, search, planning, knowledge representation, and learning, are fundamental to the development of successful AI systems. AI techniques, such as supervised and unsupervised learning, neural networks, Bayesian networks, and RL, allow for more refined and sophisticated applications of these principles. Along with the advancements of deep learning, more complex AI systems can be developed to tackle complex problems and derive highly accurate predictive models.
11.2.2 Machine learning algorithms and their applications Machine learning algorithms are a set of algorithms used to identify patterns in data and learn from such data. The main purpose is to enable computers to learn from data and use these learnings to create predictive models and automate decisions. These algorithms are broadly categorized into supervised learning, unsupervised learning, and RL: – Supervised learning algorithms are used to predict the output, given a certain input. Examples of supervised algorithms include linear regression, decision trees, logistic regression, and support vector machines. – Unsupervised learning algorithms is used to identify patterns in data without labels. Examples of unsupervised algorithms include clustering, dimensionality reduction, and anomaly detection. – RL algorithms are used to optimize decisions in an environment. Examples of reinforcement algorithms include Q-learning, Markov decision processes, and Monte Carlo tree search. Applications of machine learning algorithms include seeking patterns in customer data for marketing purposes, analyzing medical records to predict diseases, and identifying objects in surveillance footage [6].
11.2.3 Deep learning and neural networks in industrial settings Deep learning and neural networks have become the go-to solutions for many challenging industrial settings. Deep learning is a subset of machine learning, which seeks to replicate the way humans absorb and process information and data. Neural networks are a form of deep learning that imitate the structure and functions of the
236
S. P. Sreekala et al.
human brain. In industrial settings, deep learning algorithms can be used to categorize and classify large amounts of data, detect patterns, and make predictions and generate insights [7]. Neural networks have the ability to automate and streamline decision-making processes, and can be used to improve operational efficiency and reduce costs. In some industrial settings, deep learning algorithms can be used to detect anomalies or patterns with high accuracy as compared to traditional methods. By feeding these algorithms with large datasets, the algorithms can be trained to identify features, and even small changes or instances that would otherwise be undetected. They can also be used to detect objects in images, recognize speech, and model dynamic systems. In terms of predictive maintenance, neural networks can be used to detect subtle anomalies or small changes in the parameters of a system over time. The neural network will then be able to predict when a system is either likely to fail or requires maintenance. Deep learning and neural networks have shown some promise in industrial settings [8]. By combining these technologies with traditional methods, it is possible to automate and streamline decision-making processes, improve operational efficiency, and reduce costs.
11.2.4 Reinforcement learning and its relevance to industry RL is an important subfield of AI that is gaining popularity in many industries. It is used to create agents, which act and make decisions within an environment, based on collected data and feedback. The agent learns how to optimize its behavior over time in order to achieve higher levels of success. Examples of RL applications can be seen in robotics, autonomous racing, finance, customer service, and even medical diagnosis [9]. The key concept of RL can best be described as a form of trial-and-error learning. The agent is given a goal and it takes actions that will interact with its environment in order to receive a reward or penalty. Based on this feedback, the agent can adjust its strategy and learn what kind of behavior generates the highest rewards. This can be compared to how humans learn a new skill, like how to ride a bike. We try several different methods and adjust our behavior until we find the most successful one. In the business world, RL can be used to solve a variety of problems such as optimizing pricing, inventory control, product design, and customer targeting. As the agent learns more about its environment, it can find new ways to exploit existing resources and outmaneuver the competition. For example, an online retailer could use a RL agent to identify customer segments based on their buying patterns and target them with specialized offers to improve sales [10]. RL can be a powerful tool for companies of any size to make better decisions. Its ability to learn from its environment and adapt to changing conditions makes it a great asset to any business. And, with its increasing popularity in the industry, the potential for businesses to use RL agents to automate and optimize business processes is huge.
11 A survey of AI in industry
237
11.3 AI in manufacturing and production AI in manufacturing and production has become one of the most important tools helping companies to stay competitive. AI is used to automate and optimize processes for increased quality, cost savings, and better customer service. AI is used to identify, analyze, and predict patterns in the production process in order to detect problems and automate decisions, such as the timing of production [11]. AI also helps to improve customer service by providing more accurate estimates of delivery dates and helping to identify customer needs by analyzing customer data. Additionally, AI can be used to improve the safety at the workplace by autonomously monitoring production lines or identifying potential hazards. Overall, AI is used in a variety of ways to reduce costs, improve quality, and provide better customer service.
11.3.1 Predictive maintenance and condition monitoring using AI Predictive maintenance and condition monitoring using AI is a technology focused on the prediction of component and machinery failure before it happens. This type of maintenance relies on advanced analysis of data to identify potential problems and facilitate predictive maintenance tasks in order to stop failures before they happen or before they become serious. For predictive and condition monitoring, AI leverages machine learning algorithms in order to detect changes in the condition of components or machinery based on sensor data, pressure readings, vibration analysis and other parameters [12]. This helps in early prediction and prevention of failure, thereby reducing the downtime and associated costs. Additionally, AI can be used to optimize the scheduling of maintenance tasks, further improving safety, cost efficiency and maximizing efficiency.
11.3.2 Quality control and defect detection in manufacturing processes Quality control is the process of ensuring that products are free from defects and meet a certain standard of quality through inspections and testing. It is a critical component of the manufacturing process and is usually done at different stages of production to identify and prevent the manufacture of defective parts and products. Defect detection is the process of detecting defects in products during production before they reach the retail market or customer. It is the first step in catching and fixing any defects before products reach the customer [13]. This ensures the highest quality possible and that customers receive only defect-free products. Methods for defect detection can range from manual physical inspections to automated machine-vision systems.
238
S. P. Sreekala et al.
11.3.3 Optimization of production processes with AI techniques Optimization of production processes with AI techniques is the use of AI techniques such as neural networks to streamline and optimize the production processes. AI techniques can be used to reduce the amount of waste in the production process, improve efficiency, and reduce quality defects. AI can predict failures, identify maintenance needs, and detect anomalies in production data [14]. AI can also help in the automation of processes and the identification of bottlenecks. AI can provide real-time insights and make better decisions regarding manufacturing schedules and resource allocations. Furthermore, AI techniques can be used to reduce production cost and increase yield.
11.4 AI in supply chain and logistics There are technologies that use AI to help companies in the supply chain and logistics sectors to optimize their operations. AI can be used to automate some or all of the processes involved in the supply chain, such as warehousing, inventory management, order fulfillment, and scheduling truck routes. AI can also be used to analyze large amounts of data related to logistics, such as customer order information, real-time transport feedback, and customer trends. This data can then be used to predict future supply chain performance, enabling companies to improve inventory and warehousing management, optimize transportation and route planning, and improve customer service [15]. AI can also be used for predictive maintenance to detect faults and predict repairs, ensuring machinery is kept running efficiently. AI can also be used to power digital assistants and chat systems, providing users with personalized service, 24/7.
11.4.1 Demand forecasting and inventory optimization using AI Demand forecasting and inventory optimization using AI is a process of predicting customer demand for certain products or services, and using AI algorithms to recommend the optimal quantity of inventory to be held and forecasting future customer demand. AI algorithms can analyze large amounts of data from customer purchase histories, open and closed databases, and other sources to accurately predict customer demand and recommend the best levels of inventory to meet customer needs [16]. By using AI to evaluate customer demand, businesses can make more informed decisions about inventory, lower the cost of holding inventory, and improve customer service levels.
11 A survey of AI in industry
239
11.4.2 Route optimization and logistics planning with AI algorithms Route optimization and logistics planning with AI algorithms refers to the use of AI algorithms to optimize logistics-related tasks such as route planning, scheduling, inventory management, and fleet management. AI algorithms can be used to analyze large amounts of data to identify optimal routes and create logistics plans that reduce cost and increase efficiency. The algorithms can also review past performance to identify areas for improvement that can further increase efficiency [17]. AI algorithms can also be used to automate scheduling tasks and create optimized order fulfillment plans. AI algorithms can also be used to automate decision-making related to inventory and fleet management. In the context of route optimization, AI algorithms can be used to develop optimized plans to deliver goods with minimal cost and time. AI algorithms can also be used to continuously monitor and adjust logistics to changing conditions for improved efficiency.
11.4.3 Warehouse automation and AI-driven order fulfillment Warehouse automation and AI-driven order fulfillment are transforming fulfillment operations in warehouses of any size. Powered by automation systems and AI-driven analytics and decision-making, modern warehouses can now keep up with higher customer demands while saving time and money. Leading warehouse automation technologies include robotic material handling, automated storage and retrieval systems, automated sortation systems, pick-to-light systems, and automated guided vehicle systems. AI solutions can provide real-time tracking of inventory, optimize workflows and sequencing, and streamline order fulfillment processes. AI also enables warehouse operators to make data-driven decisions that can significantly reduce their operational costs and increase customer satisfaction. In light of the industry’s everincreasing customer demands, warehouse automation and AI-driven order fulfillment are the keys to success for many businesses today.
11.4.4 AI in several industries It have been used in several industries such as transportation, retail, and manufacturing. These applications help streamline the supply chain process by reducing paperwork, increasing efficiency, and lowering costs. In transportation, for example, AI can help optimize driver routes, reduce unnecessary stops, and identify optimal shipping methods. AI can also create efficient scheduling to maximize delivery time and reduce unplanned costs. In retail, AI can be used to automate the supply chain process for stocking shelves and warehousing inventory [18]. AI can also analyze customer data
240
S. P. Sreekala et al.
to predict trends and recommend items that customers may be more likely to purchase. In manufacturing, AI can be used to automate production scheduling and inventory management. AI can also be used to analyze variables such as cost, lead time, and quality to optimize production processes and minimize waste. AI applications in supply chain management can improve the efficiency and effectiveness of the entire process, helping organizations save time and money.
11.5 AI in customer relationship management Customer relationship management (CRM) involves managing customer relationships by informing clients of your company’s value and offering your customer service team helpful tools so they can provide superior customer experiences. Predictive analytics enable companies to quickly get insights into customer behavior and preferences, automate interactions with users, and personalize services. AI also allows companies to accurately segment customers, identify potential problems before they occur, and respond quickly and accurately to customer inquiries [19]. By leveraging AI, companies can create better customer experiences and build successful customer relationships into the future.
11.5.1 Customer segmentation and targeting using AI techniques Customer segmentation and targeting using AI techniques is a process of using AI algorithms to divide customers into distinct groups based on their behaviors, needs, and preferences. By understanding customers’ buying habits, businesses are able to personalize services and products, create more cost-effective marketing campaigns, and increase customer loyalty. AI can be used to develop deeper insights into consumer behavior and preferences, yielding improved accuracy and more targeted outcomes [20]. AI techniques include NLP to segment customers based on their interests and needs; Machine Learning to identify potential customers; and neural networks to personalize marketing campaigns for each segment. By using AI-driven customer segmentation and targeting, businesses can create a data-driven strategy for targeting the right customers and driving more sales.
11.5.2 Sentiment analysis and customer sentiment tracking Statements made on social media and other digital media platforms are an excellent source of customer data for sentiment analysis. Customer sentiment tracking is the systematic collection, analysis, and reporting of customer opinions over time. Companies use customer sentiment tracking to improve their customer experience and to
11 A survey of AI in industry
241
understand how their customers feel about their products and services. This data can be collected and analyzed using sentiment analysis techniques, such as NLP. Companies monitor both qualitative (e.g., surveys, customer reviews, and feedback) and quantitative (e.g., social media posts and sentiment scores) sources to track customer sentiment. The data is then analyzed in order to identify trends and underlying problems in customer satisfaction [21].
11.5.3 Personalization and recommendation systems in CRM AI-based personalization and recommendation systems in CRM are a type of software that uses AI algorithms to provide insights into customer behavior as well as tailored content recommendations for customers. These systems help companies tune in to customers’ needs in a way that was previously not possible, with insights into their preferences and patterns of behavior. AI-based personalization and recommendation solutions analyze data from customers’ interactions with the company’s products and services to gain insight into customer needs and preferences. This data is then used to recommend tailored content and offers to customers based on their individual profiles [22]. In addition to providing customers with tailored content and offers, AI-based personalization and recommendation systems can also be used to identify customer service issues in real time and quickly recommend solutions before customers have time to become frustrated. As customers get more and more used to using personalization and recommendation systems in their daily life, companies can benefit from using these systems to provide a personalized, tailored customer experience.
11.6 AI in financial services AI in financial services leverages the power of AI to create a more efficient and optimized financial experience. AI technology allows organizations to increase efficiency, automate processes, and uncover consumer insights to improve performance and customer experience. AI can help financial firms to streamline processes, improve access to data, and detect fraud, as well as offer personalized product and recommendations to customers. AI helps in carrying out the roles of many financial professionals, from banking to stock market operations [23]. By using AI, companies can scan existing financial documents to detect anomalies, detect fraudulent activities, and create investment portfolios for personalized experiences. Additionally, AI can help financial advisors to personalize products and recommendations to cater to the customers’ needs in a faster, more efficient way.
242
S. P. Sreekala et al.
11.6.1 Fraud detection and prevention using AI algorithms Fraud detection and prevention using AI algorithms is the process of utilizing Machine Learning, Deep Learning, and other AI algorithms to automatically detect and respond to fraud and other malicious activities. AI algorithms can also be used to detect potential threats, such as content manipulation, and cybercrime. By detecting fraud in near real time, organizations can reduce their chances of costly and damaging fraud. AI-enabled fraud prevention systems can be used to continuously monitor millions of data points to identify patterns in transactions that may be indicative of fraudulent activity. AI algorithms can also be used to classify such patterns into risk ratings, and alert organizations of potential risks [24].
11.6.2 Credit scoring and risk assessment with AI models Credit scoring is a process that evaluates an individual’s financial history in order to assess their creditworthiness. It includes collecting and analyzing a range of personal and financial data – such as credit history, income, liabilities, and assets – to produce an overall score that helps lenders decide whether to approve a loan or a line of credit. Risk assessment using AI models is the process of using AI-based predictive models to assess a borrower’s risk of default. AI models use data-driven techniques to identify patterns in applicant data that help predict the likelihood of a borrower paying back a loan. This includes taking into account a variety of financial variables, such as credit history, income level, employment history, and more. AI models can also help lenders identify risky behavior, such as multiple loan applications in a short period of time. The predictive power of AI models is invaluable in reducing risk and making smarter lending decisions.
11.6.3 Algorithmic trading and AI-driven investment strategies Algorithmic trading and AI-driven investment strategies are a set of rules and algorithms used to make financial decisions. Algorithmic trading and AI-driven investment strategies are used to trade stocks and other securities in real time based on various market conditions such as news, market sentiment, and market trends. AI-driven investment strategies also allow for the automation of all trading decisions and the optimization of portfolio management. This can help to reduce costs and improve returns. Additionally, AI-driven investment strategies can leverage complex algorithms and data analysis to identify patterns, detect anomalies, and make decisions regarding investments.
11 A survey of AI in industry
243
11.7 AI in healthcare and medicine AI is becoming increasingly important in managing the growing complexity of healthcare and improving patient outcomes. AI is being used to help diagnose and treat diseases, streamline patient care processes, and to detect and prevent illness [25]. AI is also being used to help process large amounts of data quickly, allowing medical researchers to identify patterns and make better decisions. AI-enabled technologies can assist healthcare professionals in making faster, more informed decisions, while also reducing the risk of human error. AI-driven innovations in healthcare are expected to reduce costs and improve the quality of care for patients around the world.
11.7.1 Medical imaging and diagnostics with AI assistance Medical imaging and diagnostics with AI assistance is a new field of medical technology that utilizes AI to help analyze images and provide rapid, accurate diagnoses. AI-assisted systems use advanced algorithms to process and analyze medical images taken by imaging devices such as ultrasound, magnetic resonance imaging (MRI), X-rays, and computed tomography (CT) scans. By combining knowledge from large datasets, AI can quickly identify and categorize objects in an image, such as tumors. Medical imaging and diagnostics with AI assistance can help to reduce the time and cost associated with diagnosing, monitoring, and managing medical conditions by quickly and accurately analyzing and interpreting medical imaging and other data. Additionally, AI can be used to reduce falsepositive results and eliminate human error, helping to make diagnoses more accurate.
11.7.2 AI-enabled drug discovery and personalized medicine AI-enabled drug discovery and personalized medicine are two emerging areas of healthcare that are revolutionizing the way diseases are treated. AI-enabled drug discovery uses data and machine learning to identify and test new potential drugs that can be used to treat diseases. It can also be used to form clinical decision-making and help scientists understand why certain diseases occur and how they can be treated. Personalized medicine is a practice that customizes treatments based on a patient’s specific needs and genetic markers. By taking into account individual traits, such as genetic susceptibility or responses to medications, personalized medicine allows physicians to tailor treatments that are specifically tailored to a patient’s unique needs. This helps reduce the risk of taking an ineffective or potentially dangerous treatment and allows physicians to make more informed treatment decisions.
244
S. P. Sreekala et al.
11.7.3 Patient monitoring and health analytics using AI algorithms Patient monitoring and health analytics using AI algorithms are a form of AI technology that uses algorithms to monitor, assess, and analyze health data in an effort to improve patient outcomes and the overall healthcare system. By analyzing large amounts of patient data, AI algorithms can detect patterns and anomalies in the data that may indicate health risks, potential infections, or the effectiveness of treatments [26]. AI-powered health analytics can be used to predict health care outcomes, identify preventive care opportunities, and improve resource allocation. Additionally, AI can be used to efficiently track patient health, providing real-time prevention and monitoring of chronic conditions. AI can also help healthcare professionals develop personalized, customized treatment options tailored to each patients’ unique circumstances.
11.8 AI in energy and utilities AI in energy and utilities is revolutionizing the way in which energy and utilities providers interact and operate. It is driving innovation by allowing for automated decisionmaking, predictive analytics, and personalized customer interactions. AI-based solutions in the energy and utility sector help with utility forecasting, smart home energy monitoring, asset optimization, and more. AI technology can be used to improve the energy efficiency of homes, reduce energy consumption, generate important insights into customer behaviors and preferences, and help utilities run their operations more effectively. AI can also be used to identify potential supply chain disruptions and detect fraudulent energy use. By leveraging AI, energy and utilities companies can become more efficient, reduce costs, and provide personalized customer experiences.
11.8.1 Energy demand forecasting and load optimization with AI Energy demand forecasting and load optimization with AI is the process of predicting the energy demand in a particular area and then optimizing the energy supply to meet that demand using AI. AI models can use various data sources such as historical data of energy use, temperature, population, economic activity, etc. to make accurate forecasts of energy demand. AI models can then use complex optimization algorithms to optimize the energy supply for maximum efficiency, reliability, and cost savings. AI can track real-time changes and adjust the energy supply accordingly. This enables organizations to reduce their energy costs, increase energy efficiency, and reduce their carbon footprint.
11 A survey of AI in industry
245
11.8.2 Smart grid management and energy efficiency using AI Smart grid management and energy efficiency using AI is the use of advanced AI to optimize the efficiency of energy utilization in smart grids. Smart grid management is a technology-enabled system that enables improved communication between electricity customers and suppliers. AI helps optimize energy usage in smart grids by evaluating customer usage data in order to identify and act upon opportunities to conserve energy or reduce operational costs. AI can also enable energy suppliers to optimize the electricity resources allocated to customers with greater accuracy, helping to reduce electricity costs for both consumers and suppliers, and ensuring that electricity is utilized more efficiently on the grid [27]. AI can also identify potential issues in the energy grid before they happen and suggest solutions to improve reliability and safety. Additionally, AI can proactively forecast customer usage demands, allowing electricity suppliers to anticipate needs and plan for future energy needs accordingly. By leveraging AI, smart grids can optimize energy efficiency, reduce costs, enhance reliability, and increase customer satisfaction.
11.8.3 Asset maintenance and predictive analytics in the utilities sector Asset maintenance and predictive analytics in the utilities sector is the practice of utilizing predictive analytics to analyze operational data from assets in the utilities sector to identify potential maintenance issues before they occur. This practice helps to reduce downtime and mitigate the risk associated with asset failure, while enabling utilities to optimize asset performance and ensure reliable service. Predictive analytics combines data from existing systems and databases, such as a utility’s asset management system and geographic information systems, with general patterns of failure, environmental conditions, and other related factors. Based on this data-driven approach, predictive analytics can predict potential maintenance issues before they happen, allowing utility companies to take proactive action to address maintenance requirements and minimize asset risk [28]. In addition to early detection and prevention of asset failures, predictive analytics can also provide valuable insights into asset utilization and performance. By identifying asset usage patterns, utilities can improve their operational efficiency and ensure more reliable service. Predictive analytics can also track the performance of assets over time, providing detailed feedback on asset health and performance, and allowing for optimization of maintenance schedules. Predictive analytics can play an important role in ensuring the reliable, safe, and efficient operation of utility assets. It is a versatile tool that can alert utilities to potential maintenance issues before they occur, allowing for proactive preventive action. Additionally, predictive analytics can optimize the efficiency and reliability of assets, ultimately leading to improved customer satisfaction and enhanced financial performance.
246
S. P. Sreekala et al.
11.9 AI in marketing and advertising AI is increasingly being used for marketing and advertising purposes. It can perform tasks such as personalizing content to target audiences, making more effective use of existing data, and optimizing customer experiences. AI can also provide useful insights into customer behavior patterns by automatically analyzing big data. AI-based platforms can be used to automate media buying, and deliver ads to the right audience at the right time [29]. AI can also provide support for marketers and digital marketers to research, optimize, and integrate campaigns across multiple channels. This type of support allows for more relevant and tailored content to satisfy customer needs, and leads to increased customer engagement. AI can help to reduce costs while providing better customer insights and optimizing campaigns for maximum ROI. AI marketing and advertising are expected to be even more prevalent in the future as businesses continue to move toward digital platforms and automation.
11.9.1 AI-powered marketing analytics and customer insights AI-powered marketing analytics and customer insights are automation techniques used to analyze customer data and customer behavior to form marketing decisions and strategies. AI-driven marketing analytics and customer insights allow companies to better understand their customers and make more informed decisions about engaging with them. AI helps companies to identify trends in customer data, identify customers’ purchasing behavior, and develop marketing campaigns that are tailored to individual customer segments. AI-driven customer insights can also help companies create more effective marketing campaigns, better target customers, and even predict future customer behavior with greater accuracy [30]. AI-powered marketing analytics and customer insights provide customers with a more personalized marketing experience, helping companies to better understand their customers’ needs and develop tailored campaigns and products to meet those needs.
11.9.2 Programmatic advertising and AI-driven ad campaigns Programmatic advertising is a form of advertising that uses automated methods to serve digital advertisements to potential customers [31]. This type of advertising allows ads to be targeted to a specific audience through automation technology such as AI. The technology collects data from online behavior and bridges the digital divide between the customer and the product. AI-driven ad campaigns rely on algorithms to ensure ads are monitored and tailored to the right audiences at the right times – without human input as much as possible [32]. AI-driven campaigns allow marketers to automate complex decision-making in real time, allowing them to reach a larger audi-
11 A survey of AI in industry
247
ence [33]. AI-driven ad campaigns can also enable automated bidding on impressions, ensuring that ads are served to the most relevant target demographic in the most cost-effective way. Programmatic advertising and AI-driven ad campaigns are important for businesses looking to optimize ROI on their marketing efforts [34].
11.9.3 Social media monitoring and sentiment analysis with AI Social media monitoring and sentiment analysis with AI is a process of using AI algorithms to track, classify, analyze, and interpret public sentiment and online conversations about a particular product, brand, service, person, or organization [35]. For instance, AI-driven sentiment analysis can help brands identify how key users are interacting with them and how they are strategically positioning themselves on social media channels. AI can also help businesses better understand the overall sentiment expressed by users, and which topics and trends are the most important to them. Additionally, AI-powered sentiment analysis can be used to detect and respond to customer complaints, identify customer service issues, uncover marketing opportunities, and monitor competitor strategies [36].
References [1]
[2]
[3] [4]
[5]
[6]
[7] [8]
Pallathadka, H., Ramirez-Asis, E. H., Loli-Poma, T. P., Kaliyaperumal, K., Ventayen, R. J. M., & Naved, M. (2023). Applications of artificial intelligence in business management, e-commerce and finance. Materials Today: Proceedings. 80, 2610–2613. Kumar, S., Lim, W. M., Sivarajah, U., & Kaur, J. (2023). Artificial intelligence and blockchain integration in business: Trends from a bibliometric-content analysis. Information Systems Frontiers, 25(2), 871–896. Ribeiro, J., Lima, R., Eckhardt, T., & Paiva, S. (2021). Robotic process automation and artificial intelligence in industry 4.0 – a literature review. Procedia Computer Science. 181, 51–58. Burström, T., Parida, V., Lahti, T., & Wincent, J. (2021). AI-enabled business-model innovation and transformation in industrial ecosystems: A framework, model and outline for further research. Journal of Business Research. 127, 85–95. Kaur, D. N., Sahdev, S. L., Sharma, D. M., & Siddiqui, L. (2020). Banking 4.0:‘The influence of artificial intelligence on the banking industry & how ai is changing the face of modern day banks’. International Journal of Management, 11(6), 1–9. Akter, S., Michael, K., Uddin, M. R., McCarthy, G., & Rahman, M. (2022). Transforming business using digital innovations: The application of AI, blockchain, cloud and data analytics. Annals of Operations Research, 2022, 1–33. Chen, H., Li, L., & Chen, Y. (2021). Explore success factors that impact artificial intelligence adoption on telecom industry in China. Journal of Management Analytics, 8(1), 36–68. Lee, D., & Yoon, S. N. (2021). Application of artificial intelligence-based technologies in the healthcare industry: Opportunities and challenges. International Journal of Environmental Research and Public Health, 18(1), 271.
248
[9] [10]
[11]
[12]
[13]
[14]
[15] [16] [17]
[18]
[19]
[20]
[21]
[22]
[23] [24]
[25]
S. P. Sreekala et al.
Paschen, U., Pitt, C., & Kietzmann, J. (2020). Artificial intelligence: Building blocks and an innovation typology. Business Horizons, 63(2), 147–155. Zhang, X., Ming, X., Liu, Z., Yin, D., Chen, Z., & Chang, Y. (2019). A reference framework and overall planning of industrial artificial intelligence (I-AI) for new application scenarios. The International Journal of Advanced Manufacturing Technology 101, 2367–2389. Wamba-Taguimdje, S. L., Fosso Wamba, S., Kala Kamdjoug, J. R., & Tchatchouang Wanko, C. E. (2020). Influence of artificial intelligence (AI) on firm performance: The business value of AI-based transformation projects. Business Process Management Journal, 26(7), 1893–1924. Dash, R., McMurtrey, M., Rebman, C., & Kar, U. K. (2019). Application of artificial intelligence in automation of supply chain management. Journal of Strategic Innovation and Sustainability, 14(3), 43–53. Tyagi, A. K., Fernandez, T. F., Mishra, S., & Kumari, S. (2021, June). Intelligent automation systems at the core of industry 4.0. In Intelligent Systems Design and Applications: 20th International Conference on Intelligent Systems Design and Applications (ISDA 2020) held December 12–15, 2020 (pp. 1–18). Cham: Springer International Publishing. Chien, C. F., Dauzère-Pérès, S., Huh, W. T., Jang, Y. J., & Morrison, J. R. (2020). Artificial intelligence in manufacturing and logistics systems: Algorithms, applications, and case studies. International Journal of Production Research, 58(9), 2730–2731. Yadav, S. P., & Yadav, S. (2020). Image fusion using hybrid methods in multimodality medical images. Medical and Biological Engineering and Computing. 58, 669–687. Irfan, D., Tang, X., Narayan, V., Mall, P. K., Srivastava, S., & Saravanan, V. (2022). Prediction of Quality Food Sale in Mart using the AI-based TOR Method. Journal of Food Quality, 2022, 1–13. Yuvaraj, N., Srihari, K., Dhiman, G., Somasundaram, K., Sharma, A., Rajeskannan, S. M. G. S. M. A., . . . & Masud, M. (2021). Nature-inspired-based approach for automated cyberbullying classification on multimedia social networking. Mathematical Problems in Engineering. 2021, 1–12. Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). Energy‐efficient resource allocation model for device‐to‐device communication in 5G wireless personal area networks. International Journal of Communication Systems, 2023, 1–14. Yadav, S. P., Zaidi, S., Mishra, A., & Yadav, V. (2022). Survey on machine learning in speech emotion recognition and vision systems using a recurrent neural network (RNN). Archives of Computational Methods in Engineering, 29(3), 1753–1770. Rajalakshmi, M., Saravanan, V., Arunprasad, V., Romero, C. T., Khalaf, O. I., & Karthik, C. (2022). Machine learning for modeling and control of industrial clarifier process. Intelligent Automation and Soft Computing, 32(1), 339–359. Sivaram, M., Porkodi, V., Mohammed, A. S., Manikandan, V., & Yuvaraj, N. (2019). Retransmission DBTMA protocol with fast retransmission strategy to improve the performance of MANETs. IEEE Access. 7, 85098–85109. Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. L.-R. U. B. I. 2023. An efficient load-based resource utilization algorithm for bi-partite scatternet in wireless personal area networks. International Journal of Communication Systems, 36(4), 1–17. Yadav, S. P., Mahato, D. P., & Linh, N. T. D. (Eds.). (2020). Distributed artificial intelligence: A modern approach. Boca Raton: CRC Press. Saravanan, V., & Neeraja, A. (2013, January). Security issues in computer networks and stegnography. In 2013 7th International Conference on Intelligent Systems and Control (ISCO) (pp. 363–366). IEEE. Kannan, S., Dhiman, G., Natarajan, Y., Sharma, A., Mohanty, S. N., Soni, M., . . . & Gheisari, M. (2021). Ubiquitous vehicular ad-hoc network computing using deep neural network with IoT-based bat agents for traffic management. Electronics, 10(7), 785.
11 A survey of AI in industry
249
[26] Logeshwaran, J., Kiruthiga, T., Kannadasan, R., Vijayaraja, L., Alqahtani, A., Alqahtani, N., & Alsulami, A. A. (2023). Smart load-based resource optimization model to enhance the performance of deviceto-device communication in 5G-WPAN. Electronics, 12(8), 1821. [27] Yadav, S. P. (2021). Emotion recognition model based on facial expressions. Multimedia Tools and Applications, 80(17), 26357–26379. [28] Narayanan, K. L., Krishnan, R. S., Robinson, Y. H., Julie, E. G., Vimal, S., Saravanan, V., & Kaliappan, M. (2022). Banana plant disease classification using hybrid convolutional neural network. Computational Intelligence and Neuroscience, 2022, 1–13. [29] Sivaram, M., Porkodi, V., Mohammed, A. S., Manikandan, V., & Yuvaraj, N. (2019). Retransmission DBTMA protocol with fast retransmission strategy to improve the performance of manets. IEEE Access. 7, 85098–85109. [30] Ramesh, G., Logeshwaran, J., & Kumar, A. P. (2023). The smart network management automation algorithm for administration of reliable 5G communication networks. Wireless Communications and Mobile Computing, 2023, 1–13. [31] Gupta, R., Tanwar, S., Al-Turjman, F., Italiya, P., Nauman, A., & Kim, S. W. (2020). Smart contract privacy protection using AI in cyber-physical systems: Tools, techniques and challenges. IEEE Access. 8, 24746–24772. [32] Sjödin, D., Parida, V., Palmié, M., & Wincent, J. (2021). How AI capabilities enable business model innovation: Scaling AI through co-evolutionary processes and feedback loops. Journal of Business Research. 134, 574–587. [33] Peres, R. S., Jia, X., Lee, J., Sun, K., Colombo, A. W., & Barata, J. (2020). Industrial artificial intelligence in industry 4.0-systematic review, challenges and outlook. IEEE Access. 8, 220121–220139. [34] Loureiro, S. M. C., Guerreiro, J., & Tussyadiah, I. (2021). Artificial intelligence in business: State of the art and future research agenda. Journal of Business Research. 129, 911–926. [35] Di Vaio, A., Palladino, R., Hassan, R., & Escobar, O. (2020). Artificial intelligence and business models in the sustainable development goals perspective: A systematic literature review. Journal of Business Research. 121, 283–314. [36] Cubric, M. (2020). Drivers, barriers and social considerations for AI adoption in business and management: A tertiary study. Technology in Society. 62, 101257.
S. Revathy, S. P. Sreekala, D. Praveenadevi, S. Rajeshwari
12 The intelligent implications of artificial intelligence-driven decision-making in business management Abstract: The rapid rate at which technology has progressed in the last few decades has resulted in emerging technological solutions such as artificial intelligence (AI). AI is increasingly being used in business management decision-making processes to make more informed, better-informed, and more intelligent decisions. It is important to consider the intelligent implications of AI-driven decision-making in business management in order to ensure that the decisions made are not only effective but are also ethical, responsible, and beneficial to all stakeholders involved. AI has enabled decision makers to access and analyze large amounts of data quickly and accurately. This can aid decision-makers in making better-informed decisions as well as reducing costs and saving time. AI can also assist decision-makers in identifying important patterns or trends within datasets and in finding solutions to problems that might have otherwise been overlooked. Additionally, AI can also help to identify potential opportunities to improve the efficiency of existing processes and systems. Keywords: Artificial, intelligence, technology, business, management, efficiency
12.1 Introduction to AI-driven decision-making The intelligent implications of artificial intelligence (AI)-driven decision-making must also be carefully considered and taken into account. AI can assist decision-makers in making informed decisions but may lack the sophistication and insight into the wider implications that a human decision-maker can bring. Furthermore, AI-driven decisions can be subject to bias and errors, and can be vulnerable to manipulation or attack [1]. Finally, AI-driven decisions may risk the loss of control over decision-making, which can have a significant impact on the company’s reputation and financial performance. The effective and responsible use of AI in business management requires a balanced look at the potential risks and benefits as well as careful consideration of the intelligent implications of AI-driven decision-making. The potential benefits of AI must be carefully weighed against the potential risks and unintended consequences in order to ensure that the decisions made are both beneficial and responsible.
https://doi.org/10.1515/9783111323749-012
252
S. Revathy et al.
12.1.1 Overview of the chapter scope and objectives The scope of the intelligent implications of AI-driven decision-making in business management is to examine the various ways in which AI can help to optimize and enhance business decision-making processes. It will consider the impact of AI on a range of managerial activities such as marketing, operational planning, and resource allocation [2]. It will also explore the implications for organizations, individuals, and society, concerning any potential disruption resulting from increased use of artificial intelligence-enhanced decision-making. The objectives of this research include: – Examining the role and limitations of AI in business decision-making processes – Assessing the potential benefits and costs associated with increased use of AIenabled decision-making – Analyzing the effect of AI-enabled decision-making on organizational structures and processes – Investigating the impact of AI on organizational power and decision-making – Exploring the implications of using AI for business decision-making on society
12.1.2 Importance of AI-driven decision-making in business management AI-driven decision-making in business management is becoming increasingly important in today’s economy. AI is being widely adopted by businesses across all industries and is being used to automate and optimize decision-making in areas ranging from operational processes to marketing campaigns. AI-driven decision-making is providing businesses with previously unobtainable insights that can help with everything, from financial forecasting to product price optimization. AI-driven decision-making can automate repetitive tasks that are time consuming and labor-intensive, freeing up resources for more innovative and creative endeavors. Moreover, AI can be used to optimize decision-making processes, helping to identify the best options for achieving goals quickly and efficiently [3]. AI can also provide insights into customer behavior, allowing businesses to determine the best strategies for marketing and sales. AIdriven decision-making can be used to analyze large amounts of data to identify patterns that may have otherwise gone unnoticed. By combining traditional analytics with AI, businesses can gain a greater understanding of their customer base and how to better reach them. Not only can this improve existing processes but it can also help businesses gain a competitive edge. AI-driven decision-making is quickly becoming an essential part of modern business operations, offering savings in time and resources while providing access to valuable insights. AI can help businesses make decisions more accurately and quickly than ever before, enabling them to stay ahead of the competition and succeed in the ever-changing market.
12 The intelligent implications of artificial
253
12.2 Understanding AI-driven decision-making AI-driven decision-making employs algorithms and AI systems to make decisions. These decisions are based on analytical data and insights from the machine learning (ML) processes that are used to analyze the data. AI-driven decision-making can be used to make decisions more quickly, accurately, and cost-effectively than if humans had to do the same task manually [4]. It can be used to help automate processes or recommend the best course of action, given a set of parameters or data. AI-driven decision-making often times takes into account a variety of different types of data sources, such as customer purchase history, market trends, and customer feedback, to make decisions. AI-driven decision-making can also be used to monitor and analyze operational performance and automate internal processes. Ultimately, AI-driven decision-making can lead to more informed decision-making that can save time, money, and resources.
12.2.1 Role of AI in decision-making processes AI plays an important role in decision-making processes. AI algorithms can be used to help businesses and organizations make decisions by examining large sets of past data and learning from it. AI algorithms can be trained to forecast future scenarios, optimize resource consumption, and provide insights for decision-making. AI-based systems can also be used for predictive analytics to forecast the potential outcome of different decisions [5]. AI can also help identify various decision-making options and weigh different pros and cons using ML algorithms. AI can also help organizations become more agile and responsive to changes in market conditions by quickly identifying potentially optimal solutions. AI can also assist in decision-making by enabling automated decision-making processes. This means that AI algorithms can be used to make decisions in real time with minimal human intervention, providing quick and proven results.
12.2.2 Benefits and challenges of AI-driven decision-making Benefits: – Increased efficiency: AI systems make decisions quickly and accurately, helping businesses achieve greater cost and time efficiency. – Improved accuracy: AI systems are capable of gathering, analyzing and interpreting data more accurately than humans can on their own, leading to more reliable decisions.
254
–
–
S. Revathy et al.
Increased ability to anticipate and respond to change: AI-driven decision-making is capable of finding patterns and trends in data to anticipate future changes in the market, allowing organizations to respond in a proactive manner. Enhanced customer service: AI systems are capable of gathering and analyzing customer data to offer more personalized services, helping businesses engage their customers more effectively.
Challenges: – Cost: Implementing AI systems can be quite expensive and time-consuming, especially in terms of data gathering, training, and maintenance. – Complexity: AI systems are complex and hard to manage, as they involve multiple algorithms and data sources. – Lack of explainability: AI systems are so complex sometimes that it can be difficult to explain their decisions and outcomes, which has become a focus of research for many companies. – Human involvement: AI systems still require human input and guidance to ensure accurate and ethical results, with the potential for bias in decision-making if the correct protocols are not in place.
12.2.3 Ethical considerations in AI-based decision-making Ethical considerations in AI-based decision-making refer to the principles and guidelines that are used to ensure that decisions made by AI algorithms are both fair and ethical. This is particularly important when AI algorithms are used to make decisions that affect a person’s life, such as in health care, hiring decisions, and judicial sentencing decisions [6]. Ethics in AI decision-making involves ensuring that AI algorithms, and the decisions they are making, reflect the values of the society in which the decision is being made. This means considering the values that are important to the people the decision can affect and ensuring that AI algorithms are programmed to adhere to those values. This involves scrutinizing the raw data that AI algorithms are trained on, making sure that it is free from bias, and double-checking the decisions that are made to reduce the chance of human error. The role of humans is critical in AI decision-making, since humans are ultimately responsible for the decisions that are made.
12.2.4 Integration of AI with human decision-makers Integration of AI with human decision-makers refers to the process of combining AI and data analysis with the judgement and insights of humans to enhance the accuracy of decisions [7]. This integration helps in improving the quality of decisions taken
12 The intelligent implications of artificial
255
from one of the multiple groups or organizations. AI and its related technologies can provide valuable data and analysis that can inform human decision-makers, eliminating bias and increasing accuracy. AI can help in uncovering patterns and associations that may not be apparent to human decision-makers or predict the probability of future events based on current trends and data. This integration increases the effectiveness of decision-making processes and reduces decision-making errors.
12.3 AI-driven decision-making in strategic planning AI-driven decision-making in strategic planning involves leveraging AI technology to assist in the strategic planning process. AI helps to identify high-value insights and trends from data, generate forecasts, optimize strategies, and assess risk. It can help organizations make more informed, data-driven decisions when planning their longterm strategies and scenarios [8]. AI can also be used to streamline processes, identify potential opportunities, automate decision-making, and more. By leveraging AI, organizations can identify resources more efficiently, improve collaboration, and increase the agility of their decision-making process.
12.3.1 AI-supported market analysis and trend forecasting AI-supported market analysis and trend forecasting is a way to use AI in order to analyze the market and identify trends. AI technology can be used to mine the data from various sources, and find patterns and correlations. This is helpful for businesses to make better decisions and have an insight into what will happen in the future [9]. AIsupported market analysis and trend forecasting can also be used to identify weak points in the market and opportunities for investment. With AI-assisted market analysis, businesses can save time, increase efficiency, and make smarter decisions.
12.3.2 Strategic decision-making using AI-based simulations and models Strategic decision-making using AI-based simulations and models is the process of using AI and computational modeling to simulate different scenarios and optimize future outcomes. AI-based simulations and models can help companies strategize and anticipate future scenarios, identify areas of opportunities or risks, and develop better strategies with the aid of AI algorithms [10]. This type of decision-making can provide companies with accurate and up-to-date data that can help optimize their decisions and increase their competitive advantage in the market. AI-based simulations and
256
S. Revathy et al.
models can also be used to anticipate market changes and consumer behavior. By using predictive analytics, businesses can make better informed decisions that can lead to improved outcomes.
12.3.3 AI-driven competitive intelligence and business strategy formulation AI-driven competitive intelligence and business strategy formulation is an approach to understanding and using data to build competitive advantage across an organization’s operations. It is an umbrella term for AI-driven tools and processes that collect and analyze data to identify opportunities and risks, and then recommend strategies to capitalize on them [11]. This approach combines traditional business strategy with insights from AI and advanced analytics to make more informed decisions and build better strategies. It examines all elements of the competitive landscape – including rivals, customers, suppliers, and environment – to determine where the organization should focus resources, guide strategic decision-making, and inform tactics and resource allocations. Moreover, AI-driven competitive intelligence and business strategy formulation provides powerful insights on how to differentiate yourself from your competitors, while helping you identify lucrative areas of growth.
12.4 AI-driven decision-making in operations management AI-driven decision-making in operations management is the use of AI to automate the decision-making process in operations management. AI-driven decision-making can be beneficial for many aspects of operations management, including demand forecasting, production planning, inventory optimization, and lean manufacturing. AI is used to analyze data from multiple sources to arrive at the best decisions. AI-driven decision-making can improve accuracy and speed of decision-making and allow operations managers to focus on more strategic management tasks [12]. AI can also identify patterns, trends, and opportunities that can be used to make decisions. In addition, AI-driven decision-making can help operations managers determine their production schedule more accurately and efficiently, and can help optimize delivery processes, which can ultimately lower costs and increase customer satisfaction.
12 The intelligent implications of artificial
257
12.4.1 AI-enabled supply chain optimization and demand forecasting AI-enabled supply chain optimization and demand forecasting is a process that uses AI technologies to help improve the efficiency of supply chain operations and ensure accuracy in the forecasting of customer demand. Supply chain optimization uses AI algorithms to model complex systems of supply and demand in order to identify opportunities for streamlining processes, reduce costs, and maximize revenue. Demand forecasting leverages AI techniques such as time-series analysis, ML, and natural language processing (NLP) to glean insight from customer behavior and predict future demand. By using AI-enabled techniques in supply chain optimization and demand forecasting, businesses can gain a better understanding of supply chain operations, better anticipate customer demand, and improve operational efficiency.
12.4.2 AI-based inventory management and procurement decisions AI-based inventory management and procurement decisions are decisions related to the efficient management of inventory and procurement of goods and services used by a company. In this type of decision-making, AI algorithms are used to aid decisionmakers in making decisions related to inventory and procurement [13]. AI algorithms are capable of predicting customer or business needs, analyzing inventory data, and optimizing purchasing decisions based on cost, availability, location, and other related factors. AI algorithms can also provide predictive insights into potential inventory management and procurement decisions that can potentially reduce costs and increase profits for a company. Furthermore, AI-based decisions can be used to automate certain aspects of the procurement process such as price negotiation and purchase order generation.
12.4.3 Automation of production planning and scheduling with AI The automation of production planning and scheduling with AI involves the use of AI technologies for managing the planning and scheduling of production activities in manufacturing and industrial operations. AI-based production planning and scheduling systems can automate a wide range of tasks such as resource optimization, capacity planning, job sequencing, and task assignment. Not only this, production planners and schedulers can use AI-driven technologies to improve the accuracy and efficiency of their operations [14]. With advanced algorithms, these systems can make complex decisions and recommendations about how to improve the production process. AIbased production planning and scheduling systems also make it easier to coordinate
258
S. Revathy et al.
tasks across the entire value chain, enabling more efficient utilization of resources and better decision-making.
12.5 AI-driven decision-making in marketing and sales AI-driven decision-making in marketing and sales is the application of AI technologies to generate insights and enable automated decision-making for sales and marketing activities [15]. AI-driven decision-making can be used to automatically deliver personalized email campaigns, optimize a website for search results, detect customer trends, generate leads, predict customer behaviors, and more. Adopting AI technology in sales and marketing can result in increased efficiency while eliminating tedious manual tasks, enabling companies to refocus efforts on activities that drive growth. Examples of AI-driven decision-making in marketing and sales include predictive analytics, NLP, ML, and customer knowledge bases.
12.5.1 AI-powered customer segmentation and targeting strategies AI-powered customer segmentation and targeting strategies are techniques used to divide customers into distinct groups based on their individual characteristics. Each segment can be targeted with different products, marketing messages, and services to better meet its needs [16]. AI-powered strategies use ML algorithms to analyze customer data, including demographic factors, purchase history, and behavior, and develop customer segments with more accurate insights than traditional methods. This allows companies to tailor their offerings and messages to each segment with greater accuracy. Additionally, AI-powered segmentation and targeting strategies can save time and resources as the ML algorithms can quickly process large amounts of data to create segmentations that are more tailored and efficient than manual methods.
12.5.2 Personalized marketing campaigns using AI algorithms Personalized marketing campaigns using AI algorithms are designed to provide customers with tailored marketing messages or offers that are based on their needs and preferences. The AI algorithms analyze customer data such as demographic information, purchase history, website browsing behavior, and other customer interactions to identify customer interests and develop more tailored marketing messages or offers [17]. This helps businesses deliver more effective marketing campaigns that meet the
12 The intelligent implications of artificial
259
needs of each individual customer. Additionally, AI algorithms can optimize campaigns by adjusting the messages or offers to meet changing customer needs. Overall, personalized marketing campaigns that use AI algorithms provide businesses with heightened customer engagement, increased conversions, and more efficient time management.
12.5.3 Pricing optimization and dynamic pricing with AI Pricing optimization and dynamic pricing with AI are strategies that businesses can use to maximize their profits. With these strategies, businesses can use AI to determine the optimal price for each product they sell by considering the pricing history and the current market environment. AI can take into account customer preferences, the competitive landscape, and changes in demand patterns to automatically tweak prices accordingly, ensuring that the company is always in line with what the market will bear [18]. AI-powered dynamic pricing also allows companies to adjust prices quickly and accurately in response to customer activity, ensuring that they are always ahead of the competition and maximizing their profits.
12.6 AI-driven decision-making in financial management AI-driven decision-making in financial management involves the use of automated systems such as ML, NLP, and computer vision to automate the process of financial decision-making. These tools can be used to identify patterns in financial data, identify trends in financial markets, and make recommendations for sound financial decisions [19]. AI-driven decision-making can help reduce costs associated with manual analysis and decision-making processes. AI can be used to improve the speed and accuracy of financial decisions, allowing for faster reactions to changes in the market and improved overall performance. Additionally, AI can help users to better understand and optimize their investment portfolios, reducing the time and effort spent researching and analyzing various investments.
12.6.1 AI-based financial forecasting and risk assessment AI-based financial forecasting and risk assessment are built on ML algorithms that are able to look at an organization’s historical data and predict changes in market conditions, as well as assess potential risks. For example, an ML algorithm can be
260
S. Revathy et al.
used to look at past financial data and determine the likelihood of a future downturn in the market or the probability of a particular investment turning a profit. AI-based financial forecasting and risk assessment are useful as they can help organizations identify potential risks or opportunities before they happen and make more informed decisions. They also offer organizations a more detailed and accurate analysis of their financial situation. AI-based systems can be used to provide evaluations of risk, identify potential market trends, provide an early warning system of future financial changes, and even suggest optimal allocation of its resources. In this way, businesses can make better decisions and improve their financial performance.
12.6.2 Automated investment decision-making with AI models Automated investment decision-making with AI models incorporates the use of AI technology to make investment decisions on behalf of investors. AI models are used to analyze data, identify profitable opportunities, and automatically execute trades in multiple asset classes. AI models are trained on historical data and generate predictions using algorithms trained on historical patterns and market trends [20]. AI models can automate investment decisions such as asset allocation, portfolio selection, and risk management, allowing for a more dynamic approach to investing. Automated investment decisions can also reduce the costs associated with managing a portfolio as well as providing investors with a more objective view on investment decisions. With AI models, investors can minimize risk and maximize returns while focusing on their long-term goals.
12.6.3 Fraud detection and prevention using AI algorithms Fraud detection and prevention using AI algorithms involves using sophisticated ML algorithms to identify patterns in large datasets. The algorithms help to uncover trends and relationships that may not be visible with the human eye, and can detect fraudulent activity more quickly and accurately [21]. AI algorithms can also use previous data to predict future fraudulent activity and alert organizations of potential problems. AI algorithms can be used in a variety of applications, from detecting fraud in online payments or credit card transactions to fraud prevention in banking systems or health insurance companies.
12 The intelligent implications of artificial
261
12.7 AI-driven decision-making in human resources AI-driven decision-making in human resources involves the application of AI technologies to automate or enhance labor-intensive HR functions like onboarding, recruiting, employee engagement, and performance management. AI-driven technology can empower real-time insights and predictive analytics to improve the effectiveness of HR decisions. In the recruitment process, AI-driven decision-making can support the accurate matching of key candidate skills with job roles across a wide array of search criteria. In addition, automated applications can scan vast databases of resumes and rank them to reduce the time spent by recruiters manually screening each document [22]. AI-driven decision-making in employee engagement can support feedback loops to enable real-time feedback for employees. AI-driven engagement data can also be used to monitor employee engagement across teams, departments, and the whole organization. AI-driven performance management systems are also becoming popular and rely on supervised learning techniques to analyze big data on job performance metrics over time. AI-driven performance systems can be used to identify key performance indicators to measure performance and identify high-performance areas or gaps. AI-driven decision-making in human resources (HR) can help to improve the accuracy, speed, cost, and transparency of HR processes. It can allow organizations to create insightful decisions and help to save time and money in the process [23].
12.7.1 AI-assisted candidate screening and recruitment processes AI-assisted candidate screening and recruitment processes are the use of AI and ML technologies to assist with the searching, vetting, and selection of job applicants. AIbased systems are being used to sort through resumes, identify candidates with skills and experience that match the job requirements, and provide shortlisted candidates for interview [24]. AI can also be used to predict the potential success or failure of candidates through automated scoring methods and to assist with the scheduling of interviews. It can also be used to retrieve data from potential applicants, such as references, education records, and other relevant information. AI-assisted candidate screening and recruitment processes can be used to save time and money, while improving the accuracy and reliability of the results.
12.7.2 Performance evaluation and talent management with AI Performance evaluation and talent management with AI is a system that uses AI algorithms to automate the process of evaluating employees’ performance and managing talent. This system uses analytical solutions to determine employee potential, skills, opportunities for improvement, and best way to allocate resources to achieve organi-
262
S. Revathy et al.
zational goals. The system typically includes an AI-driven 360-degree feedback assessment which, unlike traditional feedback methods, is designed to anonymously evaluate the employee’s performance on multiple levels. AI can also be used to identify skills gaps and match employees to roles or development opportunities. Additionally, AI can be used to create custom development plans for each employee based on their current skills and future goals [25]. AI can provide data-driven insights into organizational culture and overall performance, allowing organizations to ensure alignment with their values, and employees to better understand organizational expectations.
12.7.3 Employee engagement and sentiment analysis using AI Employee engagement and sentiment analysis using AI refers to the use of AI techniques to measure and analyze employee engagement and sentiment. AI technologies such as NLP and ML can be used to automatically analyze employee feedback in written and spoken forms and to identify sentiment and engagement levels. These analyses can provide valuable insights for organizations to better understand their employees and optimize their employee experience efforts. AI technologies can also be used to create automated employee sentiment surveys and identify areas within an organization that can be improved to increase employee engagement and satisfaction.
12.8 AI-driven decision-making in customer relationship management AI-driven decision-making in customer relationship management (CRM) is the process of utilizing AI technology to make decisions and take actions from customer data in order to optimize customer relationships. AI-driven decision-making can help by analyzing customer data in order to identify customer trends, recognize customer needs and wants, and recommend personalized products and services to customers [26]. Additionally, AI can be used to measure customer satisfaction and adjust customer service efforts according to customer feedback. This type of decision-making allows for more precise and timely analysis of customer data and can ultimately lead to improved customer relationships.
12.8.1 AI-driven customer analytics and predictive modeling AI-driven customer analytics and predictive modeling is the use of AI ML to analyze customer data and create accurate predictions about customer behavior. AI-driven customer analytics is the process of gathering, organizing, interpreting and managing
12 The intelligent implications of artificial
263
data about customers to better understand, engage with, and target them. Predictive modeling is a type of ML that uses historical customer data to build models to predict future customer behavior. These models can be used to identify the most promising customer segments, understand customer preferences, and target and personalize marketing messages to improve conversion and retention [27]. By leveraging AI and ML to analyze customer data, businesses can more effectively target potential customers, deliver effective campaigns, and improve customer service. This can result in increased sales, higher customer satisfaction, and better customer retention.
12.8.2 AI-powered recommendation systems and personalized customer experiences AI-powered recommendation systems and personalized customer experiences is a technology solution that uses powerful AI algorithms, such as NLP and ML to provide customers with personalized product recommendations or experiences that are tailored to their individual needs. This technology solution can be used to create more personalized shopping experiences, reduce cart abandonment rates, and increase customer loyalty and engagement. Additionally, it can be used to help marketers better understand customer behavior and preferences so that they can tailor product offerings to meet customer needs [28]. By leveraging AI-powered recommendation systems and personalized customer experiences, marketers are able to increase sales and satisfaction levels.
12.8.3 Sentiment analysis and customer feedback analysis with AI Sentiment analysis and customer feedback analysis with AI is a process of using AI and NLP to analyze and interpret the sentiment of customer comments and feedback in real time. It is a powerful tool in understanding customer sentiment, which can be used to build better customer experiences and improve customer loyalty. The AI system can identify the contextual key words, phrases and sentiment of customer feedback so that businesses can accurately identify customer problems and adjust their product, service, and marketing strategies accordingly [29]. In addition, AI can also be used to predict customer sentiment trends over time, helping businesses to adjust their strategies and tactics according to current market conditions. By using AI to power customer feedback analysis, businesses can improve their customer relationships and better understand customer needs.
264
S. Revathy et al.
12.9 Implications and challenges of AI-driven decision-making The implications and challenges of AI-driven decision-making are numerous. AI is being used to automate many decision-making processes, providing greater precision than traditional human decision-makers. However, this automation also presents a number of potential pitfalls and ethical issues. AI-driven decisions can often lack transparency and be difficult to explain due to the complexity of the algorithms being used. There can also be issues of bias in the data that is used to train the algorithms as well as the potential to cause unintended harm due to automated decisions. Additionally, there are concerns about the ethical implications of automation and whether AI has the capacity to be held accountable for its actions [30]. There is also the challenge of ensuring that decisions are fair and equitable for everyone and that AIdriven decisions are trusted by customers. Finally, companies must ensure that there are appropriate investments in security and privacy protection to protect against malicious use of AI-driven decision-making.
12.9.1 Impact on job roles and workforce dynamics Job roles and workforce dynamics are constantly changing and evolving due to the impact of technology. Many industries are seeing increased automation and technological advancements, leading to new and different roles that may not have even existed before. As technology advances, traditional roles become obsolete and people may have to learn new skills and take up new roles in order to remain relevant in their industry. This shift from manual labor-based roles to more technical ones can have a huge impact on a workforce’s dynamics and should be approached with considerable thought and planning [31]. A workforce must stay up-to-date on the latest trends and be willing to learn new skills if they want to remain competitive in the modern economy. Employers must be aware of the implications of technology on job roles and workforce dynamics and become proactive in anticipating and training for the changes. It is also important to address the issue of displacement of existing workers who may not be able to develop new skills or who may become obsolete in their current roles [32]. Employers can take measures to ensure job security for existing workers through retraining initiatives and creating new roles that may not have been possible before. Technological advancements are having a huge impact on job roles and workforce dynamics, and the changes are evolving rapidly. Companies must be aware of the changes and be prepared to adjust their strategies accordingly in order to remain successful.
12 The intelligent implications of artificial
265
12.9.2 Transparency and explainability in AI decision-making Transparency and explainability in AI decision-making refer to the process by which the AI system conveys information about the decisions it is making and how those decisions are obtained. This is important for providing humans with understanding and insights into the AI system’s decision-making process. Transparency and explainability allow humans to understand the decisions made by AI systems, identify bias, and draw conclusions about why certain decisions were made [33]. Transparency in AI decision-making allows stakeholders and users to understand why decisions are made by AI systems by providing information about the data inputs, features, values, and weights influencing the decisions. Explainability is the process of providing explanations as to why certain decisions were made by the AI system. This is crucial for ensuring that AI-generated decisions are fair and based on accurate data. Explainability can be provided through visualization, textual descriptions, and counterfactual reasoning, all of which help to ensure that AI decisions are interpretable and can be trusted.
12.9.3 Regulatory and legal considerations in AI adoption Regulatory and legal considerations are important when adopting AI, and can still vary depending on the industry, country, and business. The most common legal considerations include compliance with laws and regulations applicable to the development, use, and deployment of AI, intellectual property protection, data privacy and security, contracts to regulate the relationship between the parties involved in the AI initiative, and an assessment of potential personal injury and liability claims should something unexpectedly go wrong due to the AI system. Regulatory considerations involve the potential implications of regulatory requirements on the development, use, and deployment of AI. This involves understanding how the existing laws, regulations, and initiatives may overlap with each other in relation to AI technology, including sector-specific regulatory requirements across borders [34]. When implementing AI technology, organizations should ensure that privacy concerns and other applicable laws and regulations are taken into account. This includes compliance with consumer protection laws, data privacy laws, and labor and employment regulations. Organizations should also engage in risk management to identify potential risks associated with implementing AI technology. This should involve considering compliance risks, the potential for bias, data security and privacy risks, and the potential for legal action related to the AI system.
266
S. Revathy et al.
12.9.4 Ethical implications and societal impact of AI-driven decision-making Ethical implications and societal impact of AI-driven decision-making are of vital importance. AI algorithms are increasingly being used to make decisions within and across organizations, particularly in the form of automated decision systems, such as those used to provide predictive analytics. However, incorporating explicit ethical principles and norms into AI decision-making is still a challenge. There are concerns over potential discrimination, lack of transparency, lack of accountability, and a lack of user engagement in the decision-making process [35]. Additionally, when AI-driven decisions are implemented, there can be significant consequences for individuals and society. For example, automated decisions about healthcare treatment and criminal justice have the potential to amplify existing inequalities and decisions about the allocation of resources could exacerbate existing disparities. As such, when incorporating AI in decision-making, it is important to take into account ethical implications in order to ensure that AI is used to support the common good.
References [1]
[2]
[3]
[4]
[5] [6]
[7]
[8]
Novak, A., Bennett, D., & Kliestik, T. (2021). Product decision-making information systems, real-time sensor networks, and artificial intelligence-driven big data analytics in sustainable industry 4.0. Economics, Management and Financial Markets, 16(2), 62–72. Rajagopal, N. K., Qureshi, N. I., Durga, S., Ramirez Asis, E. H., Huerta Soto, R. M., Gupta, S. K., & Deepak, S. (2022). Future of business culture: An artificial intelligence-driven digital framework for organization decision-making process. Complexity, 2022, 14. Article ID 7796507. Andronie, M., Lăzăroiu, G., Iatagan, M., Uță, C., Ștefănescu, R., & Cocoșatu, M. (2021). Artificial intelligence-based decision-making algorithms, internet of things sensing networks, and deep learning-assisted smart process management in cyber-physical production systems. Electronics, 10(20), 2497. Adams, D., & Krulicky, T. (2021). Artificial intelligence-driven big data analytics, real-time sensor networks, and product decision-making information systems in sustainable manufacturing internet of things. Economics, Management and Financial Markets, 16(3), 81–93. Schrettenbrunnner, M. B. (2020). Artificial-intelligence-driven management. IEEE Engineering Management Review, 48(2), 15–19. Dubey, R., Bryde, D. J., Dwivedi, Y. K., Graham, G., & Foropon, C. (2022). Impact of artificial intelligence-driven big data analytics culture on agility and resilience in humanitarian supply chain: A practice-based view. International Journal of Production Economics, 250, 108618. Wong, L. W., Tan, G. W. H., Ooi, K. B., Lin, B., & Dwivedi, Y. K. (2022). Artificial intelligence-driven risk management for enhancing supply chain agility: A deep-learning-based dual-stage PLS-SEM-ANN analysis. International Journal of Production Research, 2022, 1–21. Belhadi, A., Mani, V., Kamble, S. S., Khan, S. A. R., & Verma, S. (2021). Artificial intelligence-driven innovation for enhancing supply chain resilience and performance under the effect of supply chain dynamism: An empirical investigation. Annals of Operations Research, 2021, 1–26.
12 The intelligent implications of artificial
[9] [10]
[11]
[12]
[13]
[14]
[15]
[16]
[17]
[18]
[19]
[20]
[21] [22] [23]
[24] [25]
267
Zohuri, B., & Rahmani, F. M. (2023). Artificial intelligence driven resiliency with machine learning and deep learning components. Japan Journal of Research, 1(1), 1–7. Dey, P. K., Chowdhury, S., Abadie, A., Vann Yaroson, E., & Sarkar, S. (2023). Artificial intelligencedriven supply chain resilience in Vietnamese manufacturing small-and medium-sized enterprises. International Journal of Production Research, 2023, 1–40. Nica, E., & Stehel, V. (2021). Internet of things sensing networks, artificial intelligence-based decision-making algorithms, and real-time process monitoring in sustainable industry 4.0. Journal of Self-Governance and Management Economics, 9(3), 35–47. Pietronudo, M. C., Croidieu, G., & Schiavone, F. (2022). A solution looking for problems? A systematic literature review of the rationalizing influence of artificial intelligence on decision-making in innovation management. Technological Forecasting and Social Change, 182, 121828. Faqihi, A., & Miah, S. J. (2023). Artificial intelligence-driven talent management system: Exploring the risks and options for constructing a theoretical foundation. Journal of Risk and Financial Management, 16(1), 31. Nica, E., Stan, C. I., Luțan, A. G., & Oașa, R. Ș. (2021). Internet of things-based real-time production logistics, sustainable industrial value creation, and artificial intelligence-driven big data analytics in cyber-physical smart manufacturing systems. Economics, Management, and Financial Markets, 16(1), 52–63. Lawrence, J., & Durana, P. (2021). Artificial intelligence-driven big data analytics, predictive maintenance systems, and internet of things based real-time production logistics in sustainable industry 4.0 wireless networks. Journal of Self-Governance and Management Economics, 9(4), 1–12. Pujahari, R. M., Yadav, S. P., & Khan, R. (2022). Intelligent farming system through weather forecast support and crop production. In: Mohammad Ayoub Khan, Rijwan Khan & Mohammad Aslam Ansari Application of Machine Learning in Agriculture (pp. 113–130). United States: Academic Press. Singh, J., Deepika, J., Sathyendra Bhat, J., Kumararaja, V., Vikram, R., Jegathesh Amalraj, J., . . . & Sakthivel, S. (2022). Energy-efficient clustering and routing algorithm using hybrid fuzzy with grey wolf optimization in wireless sensor networks. Security and Communication Networks, 2022, 1–13. Yuvaraj, N., Srihari, K., Chandragandhi, S., Raja, R. A., Dhiman, G., & Kaur, A. (2021). Analysis of protein-ligand interactions of sars-cov-2 against selective drug using deep neural networks. Big Data Mining and Analytics, 4(2), 76–83. Logeshwaran, J., Kiruthiga, T., Kannadasan, R., Vijayaraja, L., Alqahtani, A., Alqahtani, N., & Alsulami, A. A. (2023). Smart load-based resource optimization model to enhance the performance of deviceto-device communication in 5G-PAN. Electronics, 12(8), 1821. Singh, P. K., Chauhan, S. S., Sharma, A., Prakash, S., & Singh, Y. (2023). Prediction of higher heating values based on imminent analysis by using regression analysis and artificial neural network for bioenergy resources. In Proceedings of the Institution of Mechanical Engineers, Part E: Journal of Process Mechanical Engineering, 09544089231175046. Saravanan, V., & Mohan Raj, V. (2016). Maximizing QoS by cooperative vertical and horizontal handoff for tightly coupled WIMAX/WLAN overlay networks. Cluster Computing, 19(3), 1619–1633. Mojjada, R. K., Yadav, A., Prabhu, A. V., & Natarajan, Y. (2020). Machine learning models for covid-19 future forecasting. Materials Today: Proceedings 2020, 1–5. Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). A novel architecture of intelligent decision model for efficient resource allocation in 5G broadband communication networks. ICTACT Journal on Soft Computing, 13(3), 2986–2994. Yadav, S. P. (2020). Vision-based detection, tracking, and classification of vehicles. IEIE Transactions on Smart Processing and Computing, 9(6), 427–434. Sumathi, A., & Saravanan, V. (2015). Bandwidth based vertical handoff for tightly coupled WIMAX/ WLAN overlay networks 74(10), 560–566.
268
S. Revathy et al.
[26] Manikandan, V., & Yuvaraj, N. (2019, May). Advanced expert system using particle swarm optimization based adaptive network based fuzzy inference system to diagnose the physical constitution of human body. In Emerging Technologies in Computer Engineering: Microservices in Big Data Analytics: Second International Conference, ICETCE 2019, Jaipur, India, February 1–2, 2019, Revised Selected Papers (Vol. 985, pp. 349). Springer. [27] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). Energy‐efficient resource allocation model for device‐to‐device communication in 5G wireless personal area networks. International Journal of Communication Systems, 2023,1–14. [28] Yadav, S. P., Mahato, D. P., & Linh, N. T. D. (Eds.). (2020). Distributed artificial intelligence: A modern approach. Boca Raton: CRC Press. [29] Rajalakshmi, M., Saravanan, V., Arunprasad, V., Romero, C. T., Khalaf, O. I., & Karthik, C. (2022). Machine learning for modeling and control of industrial clarifier process. Intelligent Automation and Soft Computing, 32(1) 339–359. [30] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. L-RUBI (2023). An efficient load-based resource utilization algorithm for bi-partite scatternet in wireless personal area networks. International Journal of Communication Systems, 36(4), 1–17. [31] Yuvaraj, N., Raja, R. A., Kousik, N. V., Johri, P., & Diván, M. J. (2020). Analysis on the prediction of central line-associated bloodstream infections (CLABSI) using deep neural network classification. In Computational Intelligence and its Applications in Healthcare (pp. 229–244). India: Academic Press. [32] Belhadi, A., Kamble, S., Fosso Wamba, S., & Queiroz, M. M. (2022). Building supply-chain resilience: An artificial intelligence-based technique and decision-making framework. International Journal of Production Research, 60(14), 4487–4507. [33] Ganapathy, A., Redwanuzzaman, M., Rahaman, M. M., & Khan, W. (2020). Artificial intelligence driven crypto currencies. Global Disclosure of Economics and Business, 9(2), 107–118. [34] Rodgers, W., Yeung, F., Odindo, C., & Degbey, W. Y. (2021). Artificial intelligence-driven music biometrics influencing customers’ retail buying behavior. Journal of Business Research, 126, 401–414. [35] Zohuri, B., & Moghaddam, M. (2020). From business intelligence to artificial intelligence. Journal of Material Sciences and Manufacturing Research. SRC/JMSMR/102 Page, 3.
P. Krishna Priya, Karthik Rudra, Deepak Sai Kandula, G. Charan Teja, K. Venkata Koushik Reddy
13 An innovative analysis of AI-powered automation techniques for business management Abstract: Artificial intelligence (AI) is rapidly changing the way businesses manage their operations. AI-powered automation techniques are becoming an increasingly important tool for enterprise success. Automation helps businesses save time and money by eliminating or reducing manual tasks to streamline processes. Automation also enables decision-making with greater accuracy and speed. AI-powered automation can improve efficiency in business operations by automating repetitive tasks and saving time. Automated systems can also make personal connections with customers by recognizing their preferences and responding accordingly. AI tools can even predict trends and recommend solutions. Automation also helps improve customer experience by providing improved customer service and increased accuracy when interacting with customers. Keywords: Artificial, intelligence, business, automation, streamline, speed, accuracy
13.1 Introduction to AI-powered automation Artificial intelligence (AI)-powered automation techniques can have a big impact on the customer experience and customer loyalty. Automated systems can anticipate customer questions or needs and provide personalized and timely responses. Automated systems also help eliminate human errors, creating a more error-free customer experience. Automation can also increase productivity in workflows by helping employees to make decisions faster and more accurately [1]. AI-powered automation techniques are also beneficial in terms of cost savings. Automated systems help save companies money by eliminating the need for manual labor, reducing labor costs, and eliminating errors. Automated systems also provide scalability for businesses that need to increase or decrease the size of their workforce. By leveraging the power of AI-powered automation techniques, businesses can significantly improve operational efficiency and customer experience. Automation also offers the potential to reduce costs, streamline processes and improve decision-making capabilities. As the power of AI-driven automation continues to grow, businesses can learn ways to unlock the value of automation to gain lasting competitive advantages in the market.
https://doi.org/10.1515/9783111323749-013
270
P. Krishna Priya et al.
13.1.1 Overview of the chapter’s scope and objectives AI-powered automation techniques aim to improve efficiency by reducing manual labor and increasing accuracy in various applications. Automation has the potential to improve efficiency in industries ranging from manufacturing to healthcare and financial services [2]. AI-powered automation techniques can help to improve process accuracy and reduce the time taken to complete complicated tasks. Examples of AIpowered automation techniques include natural language processing (NLP), computer vision, and machine learning. These techniques can be used to automate tasks such as data entry, searching for information in documents, and customer service chatbots. The scope and objectives of AI-powered automation techniques is to increase efficiency and improve accuracy through the use of AI and automated systems. AIpowered automation techniques can be used to automate any task that requires human intelligence, reduce manual labor by automating mundane tasks, and leverage big data to produce actionable insights that can be used to improve operational processes and customer experiences. Additionally, AI-powered automation can help reduce costs and increase customer satisfaction levels.
13.1.2 Importance of AI-powered automation in business management AI-powered automation plays an important role in modern business management. Automation technologies and AI applications can help businesses reduce operational costs, automate tedious or dangerous tasks, and benefit from real-time insights. AI-powered automation can help businesses manage workloads more efficiently and achieve a greater level of process scale and sophistication. Automation also relieves the pressure on human resources, freeing up team members to focus on the most important aspects of a project. AI-powered automation can be used to improve customer service, increase product quality, and optimize a range of business operations [3]. It can help businesses identify new opportunities, find more efficient ways to manage resources, reduce costs, and streamline processes. AI can be used to increase the accuracy and speed of data analysis, enabling businesses to stay ahead of their competitors. Automation helps businesses achieve cost savings, reduce manual labor, increase throughput, and improve productivity. AI-powered automation provides businesses with the insight they need to make data-driven decisions. By automating routine tasks and processes, businesses can reduce the workload on employees and free up their time to focus on more meaningful work. AI-powered automation helps businesses stay up-to-date on the latest insights and trends in a changing world, as well as better understand customer needs [4].
13 An innovative analysis of AI-powered automation techniques
271
13.2 Understanding AI-powered automation AI-powered automation is the process of using AI and machine learning in order to automate certain tasks or processes. This type of automation relies on algorithms that are trained to complete specific tasks without the need for manual input. By leveraging AI and machine learning, companies are able to automate tasks such as data entry, customer service, and more [5]. AI-powered automation also allows companies to focus on their core competency and use their resources more efficiently. AIpowered automation has the potential to make businesses more efficient and costeffective, allowing them to focus on more productive activities.
13.2.1 Definition and benefits of AI-powered automation AI-powered automation refers to automated processes that use AI technology to increase the efficiency, effectiveness, accuracy, and speed of operations. AI-powered automation leverages AI technology and machine learning algorithms to optimize everyday processes, automating mundane tasks and reducing the burden on both customers and employees. This automation provides benefits like decreased labor costs, increased customer satisfaction, more reliable results, and improved efficiency. Additionally, AIpowered automation also offers predictive analytics capabilities, which can help businesses make more informed decisions.
13.2.2 Role of AI in streamlining business processes Artificial Intelligence is playing an increasingly crucial role in streamlining business processes and making them more efficient, cost-effective, and optimized. AI can be used to automate mundane tasks, discover trends and insights, and help businesses make data-driven decisions. AI technologies, such as intelligent agents, NLP, robotic process automation (RPA), and machine learning, are being applied to automate manual processes, speed up customer service, and improve operations. AI can be used to manage and analyze customer data, uncover patterns in customer behavior, and automate conversations with customers. AI can also monitor employee performance, provide tailored training, or help managers with task management [6]. The potential advantages of using AI in business processes to make them more efficient and optimized are immense. AI can help reduce the burden of managing mundane tasks, allowing businesses to focus on more high-value activities such as strategizing and innovating. AI-driven insights can also expose areas for improvement, allowing organizations to make more informed decisions that can ensure a faster return on investment (ROI).
272
P. Krishna Priya et al.
13.2.3 Challenges and considerations in implementing AI-powered automation The main challenge in implementing AI-powered automation is getting the technology to accurately recognize patterns and data. AI-based automation systems need to be trained with massive amounts of data in order to effectively recognize patterns. In order for AI-powered automation to be effective, the system must be able to properly identify objects, text, or images, and act upon them accordingly. Another challenge is keeping up to date with new technologies and industry standards, as AI-powered automation is constantly evolving and improving. In order for an AI-powered automation system to remain effective, it must be regularly updated to reflect changes in the environment [7]. In addition, the technology must be able to scale easily to handle an increasing level of complexity and data. This can be difficult to accomplish, especially when the system is being used in a number of different applications. There is the challenge of ethical considerations. AI-powered automation has the potential to cause problems if it is not used responsibly. In order to ensure the safety and security of users and data, all AI-powered automation systems must follow strict ethical guidelines. This includes considerations such as user privacy, data security, and data accuracy.
13.2.4 Integration of AI with existing automation techniques Integration of AI with existing automation techniques is the process of combining advanced AI algorithms and existing automation strategies to enhance the efficiency and accuracy of automation. By integrating traditional automation with AI techniques, production processes can be accelerated, improved, and become more cost effective. AI solutions can analyze large data sets to identify areas that may need additional human input. For example, AI can be used to make sure that the entire automation process is running efficiently and correctly. AI algorithms can also detect changes in production that may require human attention and take appropriate action. This integration provides a more reliable and efficient production process, which can lead to improved customer experience and satisfaction.
13.3 AI-powered automation in operations management AI-powered automation in operations management is the use of AI and automation to improve the efficiency and productivity of operations management processes [8]. AIpowered automation solutions are used to automate repetitive tasks in areas such as
13 An innovative analysis of AI-powered automation techniques
273
workflow and processes, IT operations and maintenance, document automation, and customer support. AI-powered automation solutions use powerful algorithms to detect patterns, identify inefficiencies, and automate tasks such as the configuration of IT systems, the classification of data, and the automation of manual processes. AIpowered automation can enable organizations to reduce costs, reduce errors, increase scalability, and enable faster decision-making.
13.3.1 Automated data collection and analysis using AI algorithms Automated data collection and analysis using AI algorithms is the process of using AI techniques to gather, process, and analyze data. It is used to identify patterns and draw insights from large and complex data sets. The algorithm-driven approach is based on self-learning, data-driven models, which can be trained to process massive amounts of data with higher accuracy and in lesser time than traditional methods. This allows businesses to quickly identify opportunities, trends, and anomalies in data, leading to improved decision-making and business outcome. AI techniques such as machine learning, NLP, computer vision, and deep learning are used to collect, sort, and analyze data on a large scale, discovering patterns and providing insights that can inform decisions and lead to operational improvements.
13.3.2 Intelligent process automation and workflow optimization Intelligent process automation (IPA) is a term used to refer to any combination of technologies, ranging from workflow optimization and AI to RPA [9]. It is designed to automate the tedious, manual processes that can take up a lot of time and resources in many organizations. IPA is also intended to increase efficiency and productivity, by using algorithmic processes and rules to automate certain business tasks. This type of technology can be applied in any domain right from business process automation, medical automation, construction automation, communication automation, and more. IPA is used to provide intelligent analysis of data and make intelligent decisions, automate complex processes, and provide “human-in-the-loop” capabilities in order to manage exceptions. IPA combines technologies such as RPA, smart analytics, cognitive computing, NLP, and machine learning. This type of solution is used to minimize errors, reduce costs, and increase efficiency. Lastly, IPA can also be used to increase customer satisfaction by ensuring that customer experience is optimized.
274
P. Krishna Priya et al.
13.3.3 AI-driven decision support systems for operations management An AI-driven decision support system is a computerized decision support system that uses AI techniques such as machine learning to analyze data and offer guidance or support in decision-making processes [10]. AI-driven decision support systems are used in operations management to identify potential problems, assess their likelihood of occurring, and develop plans to address them. The systems can also recommend actions based on the analyzed data. AI-driven decision support systems use data from different sources, such as financial records and customer surveys, to provide useful insights and information. This data is then analyzed by a variety of AI algorithms to recognize significant patterns and correlations. This information can then be used to produce better decisions and plans. For example, AI-driven decision support systems can help to optimize inventory levels, balance production capacity and workload, and improve customer service levels.
13.4 AI-powered automation in customer service AI-powered automation in customer service is a technology-based approach that uses AI to automate parts of customer service experiences. AI-powered automation provides companies with new ways to offer improved end-to-end service experiences for customers. Automation helps deliver faster and more accurate customer service when done correctly, freeing up customer service agents to attend to more difficult cases [11]. AI-powered automation technologies include virtual customer service assistants, automated chat bots, text message bots, self-service portals, and other forms of automated customer service that help improve customer satisfaction. AI-powered automation enables customer service agents to work smarter not harder and improves customer loyalty. Automation also helps companies reduce costs, improve speed and accuracy of customer service, and gain valuable customer insights through data analytics.
13.4.1 AI chatbots and virtual assistants for customer support AI chatbots and virtual assistants provide automated customer support. They are essentially computer programs that simulate human conversations, assisting customers with questions, giving advice, and offering helpful solutions. Customers can interact with chatbots through text or voice commands. Chatbots and virtual assistants are great tools for providing customers with quick, safe, and cost-effective service. They can reduce wait times and decrease the cost of live customer support. Additionally, as
13 An innovative analysis of AI-powered automation techniques
275
AI solutions become more sophisticated, they can identify customer needs and proactively provide individualized customer service [12]. AI chatbots and virtual assistants are becoming increasingly popular in customer service, and businesses across various industries are already implementing them to improve customer experiences.
13.4.2 Natural language processing and sentiment analysis in customer interactions NLP is a type of AI technology that is used to process raw data that is in a natural language such as English, Spanish, French, or any other language. It can also be used to identify and interpret the sentiment behind the words that are being used by customers, as well as understand the context and meaning behind a customer’s message. This enables companies to better comprehend and engage with customers by providing more personal, real-time, and meaningful customer interactions [13]. Sentiment analysis is the process of computing the sentiment of customer interactions. It uses NLP algorithms to analyze customer interactions, detect sentiment, and categorize them as positive, negative, or neutral. This allows companies to quickly identify customer sentiment and take action based on that sentiment. For example, a customer service representative may determine that a customer is feeling frustrated and offer personalized assistance to resolve their issue. Ultimately, sentiment analysis allows companies to provide timely and personalized customer service.
13.4.3 Automated ticket routing and issue resolution using AI Automated ticket routing and issue resolution using AI is a new approach to customer service that uses AI to read customer messages and route them to the most appropriate agent. AI is also used to identify the root cause of customer issues and present a suggested resolution to the agent. This helps customer service agents provide more consistent and accurate customer support. AI also helps support teams by bringing in contextual information to agents, such as customer history, custom field values, and knowledge from third-party sources [14]. This allows agents to provide faster and more contextual support, reducing customer frustration. Ultimately, this will help teams reduce ticket resolution time, improve customer satisfaction, and reduce customer support costs.
276
P. Krishna Priya et al.
13.5 AI-powered automation in financial management AI-powered automation in financial management is a form of automation that uses AI technologies to automate various financial processes. AI-powered automation can be used to streamline and improve a variety of financial tasks such as risk assessment, forecasting, compliance, fraud detection, taxation compliance, customer segmentation, and more. AI-powered automation helps reduce manual labor and cost associated with financial management tasks and can lead to 10–30% cost reduction in the financial management process. AI-powered automation is highly accurate and reliable and can save a great deal of time and effort. Additionally, it improves the accuracy of decisions, that is, financial managers can make better informed decisions that help to improve the overall performance of the organization.
13.5.1 Automated data extraction and processing for financial analysis Automated data extraction and processing for financial analysis is a digital process that collects, analyzes, and synthesizes financial data from multiple sources. This process uses sophisticated software programs to automate the extraction of financial data from various sources, such as market data, news feeds, public filings, and financial databases [15]. This data is then processed and analyzed according to predefined parameters to generate useful insights into an organization’s financial performance. Automated data extraction and processing systems can provide a means of quickly and accurately obtaining critical financial information on a wide range of companies and industries. This type of system can be integrated with other enterprise systems to provide an extensive overview of financial performance.
13.5.2 AI-based fraud detection and risk assessment AI-based fraud detection and risk assessment is a technology that uses AI to detect and manage potential fraud and risk situations. It can scan large volumes of data and quickly identify suspicious activity or transactions that could be indicative of fraud. By applying machine learning algorithms such as supervised learning, unsupervised learning, and neural networks, they can establish patterns in the data, identify anomalies, and develop new rules for detecting fraud. AI-driven technology allows companies to gain more insights from their data and to more efficiently respond to fraud threats as they appear [16]. AI-based fraud detection and risk assessment also enables
13 An innovative analysis of AI-powered automation techniques
277
financial institutions to quickly respond to threats and protects their customers from potential losses due to fraudulent activity.
13.5.3 Robotic process automation in financial transactions RPA is a rapidly growing technology that enables organizations to automate mundane, repetitive tasks in financial transactions. RPA allows organizations to automate processes quickly and with minimal intervention from human personnel. RPA can be used to automate highly complex processes in finance, such as accounts payable or receivable, account reconciliation, and payroll processing. Additionally, the automation of manual, data entry-heavy tasks, such as inputting transactions into different accounting software platforms, can help reduce human errors and processing time [17]. Ultimately, RPA can help create more efficient and accurate accounting processes, saving both time and money.
13.6 AI-powered automation in sales and marketing AI-powered automation in sales and marketing is technology that uses AI and algorithms to automate sales and marketing tasks, such as lead qualification, customer segmentation, content optimization, A/B (A and B variant) testing, and more. AI-powered automation helps businesses increase efficiency, optimize campaigns, and increase profitability. It allows businesses to maximize their opportunities and accelerate their growth by automating mundane tasks and freeing up employees’ time for strategic initiatives [18]. AI-powered automation can also help businesses provide personalized experiences to their customers, leading to higher customer satisfaction and loyalty.
13.6.1 AI-driven lead generation and prospect targeting AI-driven lead generation and prospect targeting is a process in which AI and machine learning algorithms are used to locate and identify potential customers and opportunities for marketing, sales, and other business pursuits. AI-driven lead generation and prospect targeting technology helps marketers and sales teams understand customer behavior and preferences, thus offering more accurate insights when it comes to targeting the right customer with the right product or service. It can collect and analyze data from various sources –website visits, social media, and email marketing – in order to identify potential leads as well as send out tailored communications that are more likely to lead to conversions.
278
P. Krishna Priya et al.
13.6.2 Personalized marketing automation using AI algorithms Personalized marketing automation using AI algorithms is a process of leveraging customer data and machine learning algorithms to create tailored deeply personalized experiences in marketing automation. AI algorithms can help marketers identify customer intent as they interact with different channels such as websites, emails, and social media [19]. AI algorithms help marketers quickly respond to customer inquiries, detect trends and patterns in consumer behavior, and optimize marketing campaigns to ensure the best possible results. AI algorithms help marketers create experiences that are relevant and engaging for customers, while boosting their ROI.
13.6.3 Sales forecasting and pipeline management with AI Sales forecasting and pipeline management with AI is an advanced technology that uses AI technologies to predict consumer purchasing behavior and manage the sales funnel accordingly. AI can be used to analyze large amounts of customer data to make informed predictions about future purchases. This process enables marketers to create targeted campaigns that will lead to increased conversion rates. AI can also be used to build a sales pipeline, which is a tool used to track the progress of leads and prospects. By using AI, businesses can better understand each prospect or customer’s needs and interests and adjust their sales strategies accordingly [20]. AI can also be used to automate and simplify the sales process, which can further increase conversion rates.
13.7 AI-powered automation in supply chain management AI-powered automation in supply chain management is a technology that uses AI to automate and optimize supply chain operations. This technology improves speed and accuracy in a variety of tasks, from inventory and procurement, to manufacturing, transportation, scheduling, and more. By leveraging data from across the organization and analyzing it across multiple dimensions, AI-powered automation can identify patterns and inefficiencies, while guiding decisions in every step of the supply chain process [21]. AI-powered automation can also automate the entire logistics process, including predictive route optimization, predicting demand, and managing shipping delays. AI-powered automation in supply chain management can save businesses millions in labor and enable more efficient use of capital.
13 An innovative analysis of AI-powered automation techniques
279
13.7.1 AI-based demand forecasting and inventory optimization AI-based demand forecasting and inventory optimization are two techniques that businesses use to improve their bottom lines. Demand forecasting uses historical data and predictive analytics to anticipate customer demand, allowing companies to better plan for future supply needs. Inventory optimization uses AI-based algorithms and AI-driven data analysis to determine which items in inventory should be kept in stock and which ones should be sold off. Both techniques help businesses adjust their ordering and stocking decisions to ensure that they have the right amount of product available to meet customers’ needs. By doing so, businesses can reduce waste and save money by avoiding excess inventory or running out of stock. Having an accurate demand forecast and inventory optimization strategy also helps businesses better serve their customers, keeping satisfaction levels high.
13.7.2 Intelligent routing and logistics planning using AI algorithms Intelligent routing and logistics planning using AI algorithms is an emerging technology that has increased the efficiency of routing and planning for logistics operations [22]. AI algorithms allow businesses to create intelligent routing plans that can take into account real-time data such as traffic, weather, and customer preferences. AI algorithms can also analyze historical data to help logistics operators predict and plan for potential delays and issues that could slow delivery times. By using AI algorithms, businesses can also ensure that they are routing and planning for optimal routes that will save them time and money. Additionally, businesses can use AI algorithms to monitor customer satisfaction with their deliveries and quickly adjust their routing if necessary.
13.7.3 Automated supplier management and procurement decisions Automated supplier management and procurement decisions seek to streamline the decision-making process by relying on data from multiple sources. This data is used to create models to best analyze the most viable suppliers and sourcing options available. The technology also acts as a tool to break down barriers between suppliers, customers, and internal stakeholders, allowing for faster pace of negotiations and stronger relationships with vendors [23]. Automated supplier management and procurement decisions can help enterprises save money, time, and resources by quickly identifying the best options to meet their organizational needs. Additionally, the technology can evaluate the correct level of service, delivery, and cost base for a product based on the avail-
280
P. Krishna Priya et al.
ability of the suppliers, pricing, delivery times, and more. The use of automation can also provide visibility into historical supplier data in order to make the best decisions.
13.8 AI-powered automation in strategic decision making AI-powered automation in strategic decision-making is the use of AI systems to enable and support the automated evaluation and selection of best-fit outcomes that are derived from strategic decision making processes. AI-powered automation is beneficial to strategic decision makers as it allows for complex analysis of information, data, and trends. This process can lead to smart and intelligent decision-making practices that are both effective and efficient in terms of time and resources. AI-powered automation can also take into account multiple sources and opinions (such as customer feedback) when it comes to making decisions, thus reducing risk and creating better outcomes [24]. Additionally, AI-powered automation facilitates risk assessments, cost estimations, and resource allocation. This helps strategic decision makers to accurately assess the situation, enabling them to make better decisions and plans.
13.8.1 AI-enabled market analysis and trend identification AI-enabled market analysis and trend identification is the application of AI technologies to gain insights into current market trends and to make predictions about future market conditions. The technology works by collecting large amounts of data from various sources, including news reports, financial documents, customer reviews, surveys, and other types of data, and then using this data to identify trends and uncover valuable insights. AI-enabled market analysis can help identify new opportunities, improve marketing strategies, and make better investment decisions [25]. This technology can also help anticipate customer needs and tailor products and services to fit those needs, as well as to predict customer behavior. By leveraging the power of AI, businesses can gain a real-time understanding of their market and take advantage of opportunities more quickly than ever before.
13.8.2 Automated competitive intelligence and strategy formulation Automated competitive intelligence and strategy formulation are interrelated practices that allow businesses to monitor, develop, and adjust competitive strategies in real time. Competitive intelligence is the term for the process of collecting, analyzing,
13 An innovative analysis of AI-powered automation techniques
281
and interpreting data to understand a competitor’s market strategies. With automated competitive intelligence, businesses can track and evaluate their competitors’ actions in order to plan better tactics for increasing their market share. Strategy formulation, on the other hand, is the process of creating a plan of action to reach predefined goals. Automated tools such as AI software can be used to analyze and interpret large amounts of data quickly and accurately [26]. Doing so allows for the development of strategies that are tailored to both current market conditions and future industry developments. The combination of automated competitive intelligence and strategy formulation creates an efficient system for adapting business strategies in a timely manner and allows businesses to stay ahead of their competitors.
13.8.3 AI-driven scenario planning and decision support AI-driven scenario planning and decision support is a method of providing automated guidance to decision-makers by using algorithms to simulate and analyze complex scenarios. AI can be used to generate and assess different options within a particular context and provide information, insights, and guidance to decision makers. For example, a planning platform in an AI-driven scenario could take inputs about a range of factors including market conditions, customer demand, and competitive forces and use them to suggest strategies for making decisions ranging from pricing and inventory management to product and feature selection [27]. AI can also be used to provide real-time alerts and recommendations to ensure that decision-makers are aware of the most up-to-date information and trends. This type of decision support system can greatly improve the accuracy and efficiency of the decision-making process, giving businesses an edge in an ever-evolving market.
13.9 Implications and challenges of AI-powered automation Implications of AI-powered automation: – Increased productivity: AI-enabled automation can be used to automate mundane tasks, freeing up more time for people to focus on more essential and creative tasks. This increased productivity can lead to more efficient processes, faster turnaround times, and improved customer satisfaction. – Reduced costs: With AI-enabled automation, companies can reduce overhead costs by automating common processes and eliminating the need for manual labor. This can result in significant cost savings and increased ROI.
282
–
P. Krishna Priya et al.
Improved quality: AI-enabled automation can dramatically improve product and service quality as it can analyze data efficiently and identify any potential problems quickly.
Challenges of AI-powered automation: – Cost: Though AI-enabled automation can help reduce costs, it can be expensive to implement and maintain. As such, organizations must carefully assess their needs and budget before investing in a solution. – Security: AI-enabled automation can require vast amounts of data and can be vulnerable to malicious attacks. Organizations must ensure their security systems are secure and properly updated. – Complexity: AI-enabled automation can be complex and difficult to implement. Organizations must ensure they have the capabilities to integrate the solution into their existing systems.
13.9.1 Impact on job roles and workforce dynamics The impact of technological advancements on job roles and workforce dynamics is far-reaching. As technology increases efficiency and automation, traditional laborintensive jobs, such as production and manufacturing, are becoming obsolete. In their place, employers are now looking to fill roles that require greater cross-functional skills, such as data processing and analytics. This shift in job roles has also resulted in a more flexible and adaptable workforce, as employers are increasingly relying on technology-based solutions to perform core tasks [28]. At the same time, the evolution of technology is also making a major impact on the dynamics of the workforce. Technology has enabled a more interconnected workplace, where physical distance no longer impedes collaboration. This has allowed for businesses to tap into a broad global talent pool by setting up remote teams, reducing their reliance on a single geographical labor force. Additionally, it has enabled companies to more easily leverage data and analytics to gain insights into their workforce behavior and more effectively align them to business objectives. In doing so, businesses are now leveraging the power of technology to optimize overall workforce talent, resulting in a more productive, efficient, and agile working environment [29].
13.9.2 Ethical considerations and responsible AI automation Ethical considerations and responsible AI automation refer to the ethical implications associated with using AI systems to automate certain tasks and decisions. Ethical considerations encompass how AI systems are developed and deployed, and who gets to control them. Responsible AI automation includes consideration of how the AI system
13 An innovative analysis of AI-powered automation techniques
283
will be used in the real world and what potential implications this may have for individuals, groups, and society at large [30]. This means thinking about potential biases, making sure datasets are representative and accurate, and considering privacy and security issues. Additionally, responsible AI automation involves taking time to think through and clearly articulate the rules that define how the AI system should operate, as well as providing ways for users to provide feedback or request changes in the event that unforeseen issues arise.
13.9.3 Addressing biases and ensuring fairness in AI-powered automation Addressing biases and ensuring fairness in AI-powered automation is a process of removing any potential bias from AI-powered automation systems. This is done by building algorithms that do not rely on any form of bias or unsubstantiated data. Additionally, this process includes creating transparent models and evaluation metrics that can be reviewed by both developers and users, and using ethical principles to guide decision-making [31, 32]. The goal is to create conditions that ensure any automated decisions are based on an unbiased set of criteria and are not founded in any particular agenda. This helps ensure that AI-powered automation provides accurate and fair decisions.
13.9.4 Regulatory and legal aspects of AI automation in business management Regulatory and legal aspects of AI automation in business management are the laws and regulations applicable to the use of AI-powered automation in business operations [33]. This includes everything from the privacy of customer data to the safety standards for the technology. They include the impacts of AI on businesses, such as the effects of automation on the labor market, data handling and security, access and use of consumer data, competition with other businesses, and the potential for AIdriven technological bias in decisions impacting customers. Compliance with applicable laws and regulations with respect to these issues is essential for businesses in any industry [34].
284
P. Krishna Priya et al.
References [1]
Dam, H. K., Tran, T., Grundy, J., Ghose, A., & Kamei, Y. (2019, May). Towards effective AI-powered agile project management. In 2019 IEEE/ACM 41st International Conference on Software Engineering: New Ideas and Emerging Results (ICSE-NIER) (pp. 41–44). IEEE. [2] Militano, L., Zafeiropoulos, A., Fotopoulou, E., Bruschi, R., Lombardo, C., Edmonds, A., & Papavassiliou, S. (2021, December). AI-powered infrastructures for intelligence and automation in beyond-5g systems. In 2021 IEEE Globecom Workshops (GC Wkshps) (pp. 1–6). IEEE. [3] Kumar, M., Raut, R. D., Mangla, S. K., Ferraris, A., & Choubey, V. K. (2022). The adoption of artificial intelligence powered workforce management for effective revenue growth of micro, small, and medium scale enterprises (MSMEs). Production Planning and Control, 2022, 1–17. [4] Beer Mohamed, M. I., & Hassan, M. F. (2022, November). Adaptive security for cognitive robotic process automation in enterprise computing using AI-powered analytical engine. In International Conference on Artificial Intelligence for Smart Community: AISC 2020, 17–18 December, Universiti Teknologi Petronas, Malaysia (pp. 825–835). Singapore: Springer Nature Singapore. [5] Martínez-Rojas, A., Sánchez-Oliva, J., López-Carnicer, J. M., & Jiménez-Ramírez, A. (2021, August). AIRPA: An architecture to support the execution and maintenance of AI-powered RPA robots. In Business Process Management: Blockchain and Robotic Process Automation Forum: BPM 2021 Blockchain and RPA Forum, Rome, Italy, September 6–10, 2021, Proceedings (pp. 38–48). Cham:: Springer International Publishing. [6] Younis, R., & Adel, H. M. (2020, September). Artificial intelligence strategy, creativity-oriented HRM and knowledge-sharing quality: Empirical analysis of individual and organisational performance of AI-powered businesses. In The Annual International Conference of The British Academy of Management (BAM). [7] Pillai, R., Sivathanu, B., & Dwivedi, Y. K. (2020). Shopping intention at AI-powered automated retail stores (AIPARS). Journal of Retailing and Consumer Services, 57, 102207. [8] Kulkarni, P., Bhoj, R., Rajapure, S., Gujar, S., Verma, D., & Ranjan, R. (2021, October). Camelot: AIPowered Campaign Management System. In 2021 2nd Global Conference for Advancement in Technology (GCAT) (pp. 1–6). IEEE. [9] Recto, R. (2022). AI-powered tools at work; zoned in or zoned out?: A correlational study between AI and flow experiences in the workplace (Doctoral dissertation, Azusa Pacific University). [10] Ashri, R. (2019). The AI-powered workplace: How artificial intelligence, data, and messaging platforms are defining the future of work. India: Apress. [11] Vijayakumar, H. (2023). Business value impact of AI-Powered Service Operations (AIServiceOps). Available at SSRN 4396170. [12] He, A. Z., & Zhang, Y. (2022). AI-powered touch points in the customer journey: A systematic literature review and research agenda. Journal of Research in Interactive Marketing, ahead-of-print, 1–20. [13] Brown, N., Brown, A., Degupta, A., Quinn, B., Stringer, D., & Yankov, B. (2021, September). Industry first AI-powered fully automated safety observation system deployed to global offshore fleet. In SPE Offshore Europe Conference & Exhibition. OnePetro. [14] Rosado, L., Faria, P., Gonçalves, J., Silva, E., Vasconcelos, A., Braga, C., . . . & Carlos, C. (2022). EyesOnTraps: AI-powered mobile-based solution for pest monitoring in viticulture. Sustainability, 14(15), 9729. [15] Ma, Y., Ping, K., Wu, C., Chen, L., Shi, H., & Chong, D. (2020). Artificial intelligence powered internet of things and smart public service. Library Hi Tech, 38(1), 165–179.
13 An innovative analysis of AI-powered automation techniques
285
[16] Xin, Y., Yang, K., Chih-Lin, I., Shamsunder, S., Lin, X., & Lai, L. (2023). Guest editorial: AI-powered Telco Network Automation: 5G evolution and 6G. IEEE Wireless Communications, 30(1), 68–69. [17] Yadav, S. P., & Yadav, S. (2019). Mathematical implementation of fusion of medical images in continuous wavelet domain. Journal of Advanced Research in Dynamical and Control System, 10(10), 45–54. [18] Saravanan, V., & Sumathi, A. (2012). Handoff mobiles with low latency in heterogeneous networks for seamless mobility: A survey and future directions. European Journal of Scientific Research, 81(3), 417–424. [19] Yuvaraj, N., Praghash, K., Raja, R. A., & Karthikeyan, T. (2022). An investigation of garbage disposal electric vehicles (GDEVs) integrated with deep neural networking (DNN) and intelligent transportation system (ITS) in smart city management system (SCMS). Wireless Personal Communications, 123(2), 1733–1752. [20] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). Energy‐efficient resource allocation model for device‐to‐device communication in 5G wireless personal area networks. International Journal of Communication Systems, 2023, 1–14. [21] Yadav, S. P., & Yadav, S. (2018). Fusion of medical images in wavelet domain: A discrete mathematical model. Ingeniería Solidaria, 14(25), 1–11. [22] Sumathi, A., & Saravanan, V. (2015). Bandwidth based vertical handoff for tightly coupled WIMAX/ WLAN overlay networks, Journal of Scientific & Industrial Research, 74(10), 560–566. [23] Natarajan, Y., Kannan, S., & Mohanty, S. N. (2021). Survey of various statistical numerical and machine learning ontological models on infectious disease ontology. Data Analytics in Bioinformatics: A Machine Learning Perspective, 13(1), 431–442. [24] Logeshwaran, J., Kiruthiga, T., Kannadasan, R., Vijayaraja, L., Alqahtani, A., Alqahtani, N., & Alsulami, A. A. (2023). Smart load-based resource optimization model to enhance the performance of deviceto-device communication in 5G-WPAN. Electronics, 12(8), 1821. [25] Yadav, S. P., & Yadav, S. (2019). Fusion of medical images using a wavelet methodology: A survey. IEIE Transactions on Smart Processing and Computing, 8(4), 265–271. [26] Subramanian, B., Saravanan, V., Nayak, R. K., Gunasekaran, T., & Hariprasath, S. (2019). Diabetic retinopathy-feature extraction and classification using adaptive super pixel algorithm. International Journal of Engineering and Advanced Technology, 9, 618–627. [27] Shukla, A., Kalnoor, G., Kumar, A., Yuvaraj, N., Manikandan, R., & Ramkumar, M. (2021). Improved recognition rate of different material category using convolutional neural networks. Materials Today: Proceedings. 81(2), 947–950. [28] Sangeetha, R., Logeshwaran, J., Rocher, J., & Lloret, J. (2023). An improved agro deep learning model for detection of Panama Wilts disease in banana leaves. AgriEngineering, 5(2), 660–679. [29] Vashisht, V., Pandey, A. K., & Yadav, S. P. (2021). Speech recognition using machine learning. IEIE Transactions on Smart Processing and Computing, 10(3), 233–239. [30] Sekaran, S. C., Saravanan, V., RudraKalyanNayak, R., & Shankar, S. S. (2019). Human health and velocity aware network selection scheme for WLAN/WiMAX integrated networks with QoS. International Journal of Innovative Technology and Exploring Engineering (IJITEE), 8(12), 198–205. [31] Yuvaraj, N., Kousik, N. V., Raja, R. A., & Saravanan, M. (2020). Automatic skull-face overlay and mandible articulation in data science by AIRS-Genetic algorithm. International Journal of Intelligent Networks, 1, 9–16. [32] Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). A novel architecture of intelligent decision model for efficient resource allocation in 5G broadband communication networks. ICTACT Journal On Soft Computing, 13(3), 2986–2994.
286
P. Krishna Priya et al.
[33] Jiménez-Ramırez, A. (2021, August). AIRPA: An architecture to support the execution and maintenance of AI-powered RPA Robots. In Business Process Management: Blockchain and Robotic Process Automation Forum: BPM 2021 Blockchain and RPA Forum, Rome, Italy, September 6–10, 2021, Proceedings (Vol. 428, pp. 38). Springer Nature. [34] Doshi, M., & Varghese, A. (2022). Smart agriculture using renewable energy and AI-powered IoT. In: Ajith Abraham & Sujata Dash. AI, Edge and IoT-based Smart Agriculture (pp. 205–225). United States: Academic Press.
B. Girimurugan, S. Rajeshwari, S. P. Sreekala, S. Revathy
14 The smart and secured AI-powered strategies for optimizing processes in multi-vendor business applications Abstract: The development of artificial intelligence (AI)-powered processes in multivendor business applications has created significant opportunities to optimize resources, streamline operations, and reduce costs. However, deploying AI in multi-vendor business applications presents a unique set of challenges, including the potential for data privacy and security issues. To ensure the success of AI-enabled processes, organizations should implement best practices for proper data protection and secure AI-powered strategies. With the proper implementation of data security protocols and usage strategies, organizations can harness the potential of AI-powered processes in multivendor business applications to drive greater efficiency and cost savings. This includes creating greater process efficiency, reducing manual labor costs, and improving customer experiences. This will also enable organizations to remain competitive in the rapidly changing business landscape. Keywords: Artificial, intelligence, resource, optimization, business, process
14.1 Introduction to AI-powered strategies for process optimization Data security should be a top priority when implementing artificial intelligence (AI)powered processes for multi-vendor business applications. Organizations should establish secure data access protocols for internal and external users and require encryption of all data in transit and at rest. They should also implement a secure data integrity system that monitors and records all user activities, and implement robust authentication and authorization controls for user access. Finally, the organization should implement data protection methods such as data obfuscation, anonymization, and masking to protect sensitive data and limit the sharing of sensitive information across vendors. In addition to proper data security, organizations should also develop strategies to ensure the effective utilization of AI-powered processes in multi-vendor business applications. This includes developing a clear roadmap for how AI will enhance existing processes, create new ones, and automate workflows [1]. The organization should also identify the most appropriate AI-powered applications for each business process and devise ways to monitor and optimize the process for improved outcomes. Additionally,
https://doi.org/10.1515/9783111323749-014
288
B. Girimurugan et al.
the organization should establish KPIs to measure the performance of AI-enabled processes, and invest in continuing training and education to ensure the AI-powered systems remain effective and adaptive.
14.1.1 Overview of the chapter’s focus and objectives The focus of smart and secured AI-powered strategies is to develop strategies that ensure the security, safe development, and responsible use of AI solutions and services. This includes strategies to protect critical infrastructure, reduce the risk of cyberattacks, ensure ethical practices, protect personal data, and foster trust and accountability. It also aims to create strategies that help to address the economic, regulatory, and ethical challenges that accompany the use of AI and its related technologies [2]. Through these strategies, organizations and businesses can ensure that their AI solutions and services are secure, accountable, and responsible. The ultimate objective is to develop effective policy frameworks that support trust and innovation and ensure that AI is used responsibly to benefit society as a whole.
14.1.2 Importance of process optimization in multi-vendor business applications Process optimization is critical for multi-vendor business applications because it has a direct impact on the speed with which transactions are completed, customer satisfaction, cost savings, and ensuring efficiency in operations. Process optimization can help reduce lead times, provide greater accuracy and visibility, increase scalability, and eliminate the need for manual intervention. It is essential for maximizing process efficiency, meeting customer expectations, and improving the competitiveness of the business. Additionally, process optimization can ensure that processes are performed according to best practices and regulations, errors are minimized, and customer experience is improved. It can also reduce cycle times, optimize cost and resources, and ensure compliance with regulatory requirements [3].
14.1.3 Role of AI in enhancing process efficiency and effectiveness AI is an important tool for improving process efficiency and effectiveness. By leveraging algorithms and machine learning (ML), AI can make a company’s software smarter, its processes more efficient, and its customer service more effective. AI can be used to automate entire processes or just a few key steps that are prone to human
14 The smart and secured AI-powered strategies for optimizing processes
289
error or require manual entry. This automation can save time, improve accuracy, increase customer satisfaction, and reduce costs. AI technologies such as NLP, predictive analytics, and ML can also be used to improve forecasting and scheduling, identify potential process improvements, and recommend process solutions. Additionally, AI-driven process optimization can enable faster decision-making, analyze big data for trends and patterns, improve customer segmentation, and even enable predictive customer service. By using AI to optimize processes, companies can reduce costs, improve efficiency, and remain competitive.
14.2 Challenges in multi-vendor business applications Multi-vendor business applications are applications that allow for multiple vendors to provide goods and services. These applications can greatly increase the variety of goods and services that businesses are offering, but they also come with their own unique set of challenges. One of the biggest challenges that come with multi-vendor business applications is the fact that there are multiple vendors involved. This means that each vendor will have different policies, procedures, and requirements that need to be taken into account. Additionally, each vendor may have their own user interface and data standards that can make it difficult to provide a unified user experience for customers [4]. This can lead to increased customer frustration as well as decreased customer satisfaction. Another challenge with multi-vendor business applications is the complexity of linking data across multiple vendors and ensuring data integrity. With multiple vendors, it can be difficult to ensure that the data collected and shared is accurate, up-to-date, and complete. Additionally, integrating multiple applications into one larger platform can be challenging and require an experienced software development team. Managing vendor relationships can also be a challenge. Establishing and maintaining effective communications between each vendor is essential to the long-term success of the entire system. This includes establishing clear roles and responsibilities for each vendor, defining payment schedules and expectations, and staying up-to-date with the latest industry trends. Multi-vendor business applications can provide a great opportunity for businesses to increase the variety of products and services they offer. However, there are several challenges that need to be taken into consideration when implementing these applications, and developing a successful system requires a competent team of software developers, vendors, and customer service representatives.
290
B. Girimurugan et al.
14.2.1 Complexity and integration issues in multi-vendor environments Complexity and integration issues in multi-vendor environments refer to the difficulties that can arise when trying to integrate multiple vendor systems together. It can be difficult to determine how to integrate different vendor systems as each may use different languages, protocols, and technologies, and each may have custom databases, security requirements, and APIs [5]. Additionally, integration challenges can arise if vendors offer solutions that are incompatible with systems already in use. Furthermore, the cost and time involved in integrating new or existing systems can be a major barrier. Finally, unanticipated or undocumented information exchanges can lead to inconsistencies in data that can be difficult and time-consuming to resolve.
14.2.2 Data inconsistency and interoperability challenges Data inconsistency refers to differences in some way when the same data is stored in two different locations, or across multiple systems. This can cause difficulty in searching and retrieving data and may lead to incorrect comparisons and inaccurate results [6]. Interoperability challenges refer to the ability of two or more different systems or networks to communicate, exchange data, and use the information that has been exchanged to enable mutual tasks. Poor interoperability between systems can cause delays in data transfer, inaccurate data exchange, and incomplete integration of applications. It can also lead to increased transaction costs and a decrease in overall efficiency.
14.2.3 Performance bottlenecks and resource allocation problems Performance bottlenecks occur when a machine or a system reaches its maximum capacity and can no longer process data efficiently. This may result in decreased performance, slower loading times, and instability due to system overload. Resource allocation problems occur when resources such as memory, processing power, and storage capacity become limited. This may lead to higher costs, wasted time, and hindered productivity. As resources become less available over time, systems must be designed more efficiently in order to maximize their use. This can be achieved through careful planning and optimization.
14 The smart and secured AI-powered strategies for optimizing processes
291
14.3 AI-powered process optimization techniques The term “AI-powered process optimization techniques” refers to techniques that use AI to automate and improve the process of performing repetitive tasks and operations within an organization. AI-powered process optimization techniques can help businesses reduce costs, increase efficiency, reduce the need for human labor, and improve productivity and quality. By automating processes, businesses can reduce the amount of time required to complete the process and can ensure that the process is performed correctly and consistently. Some examples of such techniques include ML, deep learning (DL), natural language processing, and robotic process automation [7]. Utilizing these techniques can help organizations not only automate and improve processes but also gain greater insights into their operations and customer interactions. With the right knowledge and implementation of AI-powered solutions, businesses will be able to reduce costs and increase overall revenue.
14.3.1 Intelligent process automation using AI algorithms Intelligent process automation (IPA) is a type of automation technology that combines AI algorithms and robotic process automation (RPA) to automate repetitive, rulebased tasks such as data entry, document processing, and routing transactions. IPA can help businesses reduce costs, enhance customer service, and increase their operational efficiency [8]. IPA works by executing tasks that are initially performed by an individual, but can be offloaded to a computer system. The AI algorithms enable the computer system to learn how to complete these functions with high accuracy and at a much faster rate. Additionally, the use of RPA enables the computer system to interact with other systems and applications autonomously, saving valuable time and resources.
14.3.2 Predictive analytics for demand forecasting and resource planning Predictive analytics is a powerful tool for demand forecasting and resource planning. It utilizes data mining and advanced methods such as AI and ML to make predictions about future sales, demand, and resources. Predictive analytics can help businesses identify customer trends, anticipate customer needs, and optimize strategies for forecasting and resource planning. It is used in a variety of ways, from forecasting sales of upcoming products to finding the most cost-effective way to source resources. By understanding customer needs and market trends, businesses can better manage resources and optimize budget allocations. With the ever-increasing amount of data,
292
B. Girimurugan et al.
predictive analytics offers an efficient and effective way to utilize this data for planning and resource optimization.
14.3.3 Optimization algorithms for scheduling and routing optimization Optimization algorithms for scheduling and routing optimization are used to optimize the scheduling and routing of resources or services to ensure the most efficient use of resources. These algorithms may also optimize scheduling to meet customer needs [9]. The goal of optimization is to either minimize cost or maximize profits while taking other considerations into account. Optimization algorithms may incorporate various techniques such as linear programming, genetic algorithms, and simulated annealing. Depending on the problem at hand, the algorithms may use different techniques for solving the problem. The algorithms may also use data on supply, demand, and other constraints to determine an optimal solution.
14.3.4 Real-time monitoring and adaptive process control with AI Real-time monitoring and adaptive process control with AI and machine reading (MR) allow for automation and optimization of an entire process operation. The entire process can be monitored in real time using AI-driven analytics, and dynamic process control algorithms can be implemented to quickly respond to changes in the process environment and avoid costly production stoppages and downtime. AI-driven analytics can also contribute to predictive maintenance alerts, preventive maintenance plans, and improved equipment availability [10]. MR techniques such as Optical Character Recognition (OCR) can be used to extract useful information from process data, and to make predictive process control decisions based on this information. This information can be used to improve product quality, shorten cycles, increase throughput, and reduce energy consumption. With the help of AI, adaptive process control can maintain optimal conditions in the process parameters, while allowing the process to quickly adjust to new changes or demands.
14.4 Secured AI-powered strategies Secured AI-powered strategies involve AI, ML, and other technologies to create and apply security measures. AI-powered security measures can detect and identify malicious activities with trained patterns and algorithms, and create specialized defense systems targeting the attack. AI can also be used to proactively monitor and flag suspi-
14 The smart and secured AI-powered strategies for optimizing processes
293
cious user activity or anomalous behavior and report any threats or anomalies in near-real time. Security teams can use AI-powered automation to intelligently respond to these threats quickly and appropriately. AI-powered strategies can also be used to secure enterprise systems, networks, applications, and data – across both onpremise and cloud-based infrastructures. This results in improved detection power and better protection against malicious actors. Additionally, AI-driven security systems can conduct an analysis of data and automate routine tasks such as patching and vulnerability scanning to improve security and avoid human errors.
14.4.1 Data security and privacy considerations in AI-powered optimization Data security and privacy are important considerations in AI-powered optimization in healthcare. AI models used for optimization are fundamentally dependent on data, and access to data is necessary for the development and validation of such models. In addition, access to health data is often restricted due to legal implications and regulations. AI algorithms can easily be manipulated to access sensitive information, and lack of proper security can result in data breaches and other incidents. Thus, it is essential to ensure proper privacy and security protocols are in place to protect the personal data that is collected and used in healthcare AI applications [11]. Such protocols need to consider aspects such as data governance, secure data storage, user authorization and authentication, system access control, and secure communication protocols. It is critical to protect patient data from being shared with third-party organizations that lack appropriate security protocols, or without permission from the patient. Data needs to be protected and secured at all times during the life cycle of the data, from collection to storage and use, and should be carefully disposed off after use. Any data used must not have any form of identifying characteristics such as name, address, personal identification number, etc. to ensure patient anonymity. Failure to comply with such laws and regulations can result in hefty penalties and fines, including legal liability.
14.4.2 Authentication and access control mechanisms for multi-vendor systems Authentication and access control mechanisms for multi-vendor systems are designed to ensure that only authorized users can access the resources of a multi-vendor system. Authentication ensures that the user is who they say they are, while access control restricts what users can do, such as what data they can access and modify. Authentication techniques often involve two-factor authentication, passwords, biometrics, or other methods of identification and verification, while access control can be implemented through role-based access control, permission-based controls, audit-
294
B. Girimurugan et al.
ing, and other mechanisms. To ensure a secure multi-vendor system, these authentication and access control mechanisms must be regularly maintained and updated as new users, vendors, and resources are added to the system.
14.4.3 Threat detection and anomaly detection using AI-based techniques Threat detection and anomaly detection using AI-based techniques allow organizations to identify and prevent cyberattacks with a greater degree of accuracy than traditional security solutions. AI-based systems leverage data to learn and relearn the normal pattern of user activity, and can detect deviations from that pattern, alerting security teams to possible attacks in progress. AI-based systems are powered by large datasets of past security incidents, known as “ground truth.” This data can be used to help ML algorithms detect anomalies faster and more efficiently than other methods, such as signature-based detection [12]. Furthermore, AI-based systems can be tailored to the particular needs of an organization, allowing it to take into account the unique characteristics of the organization and its security profile.
14.4.4 Secure communication protocols for data exchange in multi-vendor environments Secure communication protocols for data exchange in multi-vendor environments are protocols designed to provide secure communication and data exchange between different vendors in a multi-vendor environment. These protocols provide a secure communication channel between two or more entities and prevent any unauthorized access to data. Examples of such protocols include Secure Sockets Layer (SSL), Transport Layer Security (TLS), IPSec, and Secure FTP (SFTP). These protocols help in authentication of the sender and receiver, encryption of data exchanges, and integrity of data transfers, thus ensuring secure data exchange among different entities [13]. Additionally, these protocols also provide a uniform platform for secure data exchange, allowing different vendors to use the same protocol without the need for custom implementation.
14.5 Case studies of AI-powered process optimization AI-powered process optimization is the use of AI to optimize a process such as in the manufacturing, logistics, healthcare, and finance industries. Processes can be optimized in almost every field, from monitoring of dynamic conditions in production
14 The smart and secured AI-powered strategies for optimizing processes
295
plants to customer support services. AI-powered process optimization can lead to faster, more efficient operations that are less prone to errors and are more responsive to market trends. Case studies help illustrate how AI-powered process optimization is being used in practice [14]. One example is in retail. Retailers can use AI to optimize their supply chain for faster shipping and better customer service. AI technology can be used to identify and track essential customer data, such as the most popular items sold and when customers make repeat purchases. AI can also be used to automate processes, such as predicting demand for specific items and ordering the appropriate products and stock levels accordingly. AI-powered process optimization also helps retailers gather insights on customer behavior, allowing them to monitor trends and adjust their operations accordingly for maximum efficiency. In healthcare, AI-powered process optimization enables automated workflow management and improved patient outcomes. AI algorithms can be used to monitor and improve patient care processes, such as prescription management, records maintenance, and appointment scheduling. AI can scan records in real time to identify potential risks, and suggest interventions that help improve overall patient health. The banking industry is another field where AI-powered process optimization is being used with great success. Banks can use AI to improve their fraud detection capabilities with more accurate and faster detection capabilities. AI also helps streamline loan approvals, automate decision-making, and build customer loyalty by offering personalized services and targeted marketing. The application of AI-powered process optimization is growing as the technology matures. By leveraging AI-powered technologies, businesses can optimize their processes to be faster, more accurate, and more cost-effective, leading to improved customer service, increased efficiency, and better overall outcomes.
14.5.1 Application of AI in supply chain optimization across multiple vendors AI is increasingly being used to improve the operations of supply chain networks across multiple vendors. It enables supply chain companies to automate processes such as route optimization, inventory management, and pricing optimization. AI-driven supply chain solutions can recognize patterns and trends in customer data, allowing suppliers to better meet customers’ needs and preferences. AI can also be used to predict demand and optimize inventory levels and prices. Additionally, AI can be utilized to power predictive analytics that can increase customer service and reduce costs [15]. Through AI, businesses can make intelligent decisions that help them be agile and respond to customer demand more accurately and efficiently. AI can also be used for automated supplier discovery, allowing companies to find the most cost-effective and reliable suppliers.
296
B. Girimurugan et al.
14.5.2 AI-driven inventory management and demand forecasting in multi-vendor setups AI-driven inventory management and demand forecasting are becoming increasingly popular in multi-vendor setups. This is because it allows the vendors to better manage stock levels, optimize shipping and distribution, and ensure just-in-time delivery. AI can be used to predict demand for specific items based on past sales data, ecommerce trends, weather, seasonality, and market conditions. AI algorithms can also be used to minimize inventory costs, reduce process times, and improve quality control. AI-driven inventory management can also automate supply chain processes, such as order routing, shipping, and tracking. With AI, vendors can better manage inventory across their multiple locations and have access to real-time visibility of inventory levels. This helps them to respond more quickly to price fluctuations, product demand, and changes in the market. In addition, AI can be used to reduce errors in demand forecasting and optimize stock rotation so that vendors can better plan ahead for future supply and demand.
14.5.3 Automated procurement and vendor selection using AI algorithms Automated procurement and vendor selection using AI algorithms is a process that uses ML and DL to optimize the selection and procurement process and identify the most appropriate vendor for a company’s product or services. The algorithm automatically identifies the best vendors in terms of costs, delivery times, and other criteria [16]. AI-powered methods such as natural language processing, scoring-based algorithms, and predictive analytics enable vendors to be selected faster and smarter than traditional manual methods. The use of these algorithms eliminates many of the manual steps that are used in vendor selection and makes the entire process more efficient, accurate, and cost-effective. Additionally, these algorithms can evaluate historical data to accurately assess vendor performance over time and adjust the selection criteria to identify the most suitable vendor.
14.6 Benefits and impact of AI-powered strategies AI-powered strategies have the potential to revolutionize the way business is conducted. AI-powered strategies can provide access to data and insights that have previously been unattainable with manual or traditional methods. For businesses, AI-powered strategies can help improve decision-making and operational efficiency.
14 The smart and secured AI-powered strategies for optimizing processes
297
One of the primary benefits of AI-powered strategies is the ability to automate decision-making. This helps to reduce the amount of manual labor required to analyze and assess information, freeing up employees to focus on more value-adding tasks. AI-powered strategies can also quickly identify patterns in large data sets and use the information to accurately forecast performance. With this information, businesses can make better-informed decisions more quickly. AI-powered strategies can also provide insights that help businesses better understand their customer base. Through AI-powered strategies, businesses can draw conclusions on customer preferences, behaviors and purchase propensities. AI-powered strategies can also provide predictive analytics that help businesses anticipate customer needs and create personalized experiences [17]. By leveraging AI-powered strategies, businesses can see a tangible impact on their bottom line. AI-powered strategies help drive operational efficiency by reducing manual labor and refining processes. AI-powered strategies can also form pricing tactics that maximize profits while keeping customers satisfied. Finally, AI-powered strategies enable businesses to remain competitive by providing access to data and insights that may not be available with traditional methods.
14.6.1 Improved process efficiency and cost savings through AI optimization Improved process efficiency and cost savings through AI optimization refers to the use of AI and ML to improve existing business processes. The goal of this type of optimization is to reduce time, cost, and resources, while increasing the efficiency of the process. AI optimization can be used to analyze datasets or perform predictive analytics to identify weak points in the process, suggest solutions, and continuously monitor and adjust its suggestions. AI optimization can be used to identify inefficiencies, automate manual processes, increase system flexibility, and adjust processes in real time [18]. AI optimization can be used for many different applications, such as customer service, supply chain management, manufacturing, marketing, and finance. Ultimately, AI optimization can save businesses both time and money while improving the accuracy and efficiency of their processes.
14.6.2 Enhanced decision-making capabilities and resource utilization Enhanced decision-making capabilities and resource utilization refer to the ability of an organization to utilize its resources to make more informed decisions, faster, and with a higher degree of accuracy. This includes the use of data-driven insights to make better decisions, leverage technology to process and analyze data quickly and
298
B. Girimurugan et al.
accurately, and optimize the utilization of resources such as time, energy, and money. Enhanced decision-making capabilities and resource utilization can help organizations make more informed and strategic decisions that can potentially lead to greater gains in productivity, efficiency, and profitability. Additionally, this approach can lead to better customer service and satisfaction as well as improved morale among employees.
14.6.3 Reduction in errors and increased customer satisfaction Reducing errors and increasing customer satisfaction is an important part of improving customer service and providing an efficient, effective business. By reducing errors and improving customer satisfaction, businesses can develop their reputation among customers, build trust, and improve customer retention. Reducing errors also leads to increased efficiency, as time and resources are not wasted by having to constantly correct mistakes [19]. Additionally, improved customer satisfaction provides an opportunity for businesses to expand their customer base and gain an edge over the competition.
14.6.4 Long-term strategic advantages in competitive multi-vendor environments Long-term strategic advantages in competitive multi-vendor environments are the tangible advantages that companies gain from forming relationships with multiple vendors and suppliers. Benefits of a multi-vendor strategy provide long-term sustainability and profitability advantages as follows: – Reduced costs: When purchasing from multiple suppliers, companies can gain cost savings from quantity discounts, lower administrative costs, and better terms and conditions for strategic vendor transactions. – Increased variety: Working with multiple vendors gives companies access to more varied product ranges, which can provide more choice for their customers. This can lead to increased sales and customer loyalty. – Improved customer service: Working with multiple vendors can help companies provide quicker order fulfilment and make it easier to meet customer needs. – Solved supply chain issues: Working with multiple vendors can also help companies meet unexpected demand or win lost orders due to supplier issues. This helps companies remain competitive and ensures customer satisfaction. – Risk management: Working with multiple vendors can help companies spread out risks, such as single supplier risks, and ensure continuity in the event of an interruption. This helps companies maintain key relationships with customers.
14 The smart and secured AI-powered strategies for optimizing processes
299
14.7 Implementation considerations and best practices Implementation considerations are the steps used to plan and execute the software, hardware, network, and database components of an IT project. Best practices are techniques, procedures, and processes that are found to produce good results in a particular situation. When it comes to implementation considerations, there are several key components to consider. Project planning and scoping are essential to ensure that the project is achievable and appropriately designed. Specific hardware, software, and network requirements must also be identified and organized to ensure that all elements of the system are compatible. Once the project is planned and the resources have been identified, it is important to test the system thoroughly before moving to production [20]. A backup and recovery plan should also be established to ensure that the system is secure and recoverable in the event of an emergency. When it comes to best practices, there are a number of areas to consider. Processes should be developed to ensure that tasks are completed effectively and efficiently. Documentation should be created and maintained to provide clarity and transparency on how the system functions. Security protocols should also be established to protect the system from attacks. Additionally, proper monitoring and logging should be employed to ensure that any potential issues are identified quickly. Finally, architecture should be designed to not only conform to best practices but to also ensure that the system is scalable, maintainable, and recoverable.
14.7.1 Data preparation and quality assurance for AI-powered optimization Data preparation and quality assurance are essential components of any AI-powered optimization program. The quality of data and its proper preparation are essential for ensuring that the AI-driven optimization performs as desired. Data preparation involves organizing, cleansing, and normalizing the data so that the AI algorithm has all the necessary information to provide optimal results. Data quality assurance involves validating the data to ensure accuracy, completeness, and overall quality before it is fed into the AI algorithm [21]. Quality assurance also includes eliminating duplicate data points, which can lead to false positives or inaccurate results. Quality assurance also requires testing the AI optimization algorithm and reviewing the data before and after optimization. These steps are necessary to ensure that the AI-driven optimization system is working effectively and is providing accurate results.
300
B. Girimurugan et al.
14.7.2 Integration challenges and strategies for multi-vendor systems Integration challenges for multi-vendor systems are those arising when integrating systems provided by multiple vendors. These challenges tend to be multifaceted due to the need to bring together different product versions, software coding languages, protocols, and technologies. Strategies for multi-vendor system integration involve clearly defined requirements, early planning, establishment of a good communication protocol among all vendors, a phased approach, compatibility testing, and the proper use of security measures. Developing open standards and leveraging cloud computing can also ease the process of integration. Additionally, it is important to ensure any integration approach used meet legal and regulatory requirements [22].
14.7.3 Collaboration and partnership models for successful implementation Collaboration and partnerships are models for successful implementation of any project or initiative. The aim of these models is to ensure integrated and effective delivery of services, resources, and strategies that result in increased satisfaction for multiple stakeholders, including customers, employees, and communities. Collaboration refers to an intentional relationship between two or more organizations that work together to identify, develop, and deliver projects and initiatives through a shared vision, strategy, and goals. Partnerships are similar but not the same as collaborations in that they involve a specific agreement involving defined roles of both parties to achieve a specific-defined outcome. Both models involve more than just planning, coordination, and implementation. Successful collaboration and partnership models must also include shared data and resources; compatible organizational cultures; agreed-upon goals and objectives; a planned budget and timeline; and a shared commitment to the project’s success. The collaboration and partnership models for successful implementation should be built on trust, respect, communication, and shared resources [23]. Such partnerships must also be managed by a process that encourages transparency and open communication, and which provides accountability and metrics that measure success. These models should also strive to be continuously improved, as needed, to ensure successful outcomes.
14.7.4 Continuous monitoring and evaluation of AI-powered strategies Continuous monitoring and evaluation of AI-powered strategies is the process used to track the results of the AI-powered strategy that is implemented. This is important as
14 The smart and secured AI-powered strategies for optimizing processes
301
it ensures that the strategy is working and that it will deliver the desired outcomes. It also helps to identify areas where improvements can be made and updates can be made to the strategy, if necessary. Continuous monitoring and evaluation of AIpowered strategies provides the visibility and transparency required to assess the effectiveness of deployed AI implementations. It also helps detect any anomalies and potential risks that could be affecting the success of the implementation. In short, continuous monitoring and evaluation of AI-powered strategies allows businesses to ensure that their strategies are working correctly and make any adjustments that may be needed.
14.8 Ethical and regulatory considerations Ethical and regulatory considerations refer to the moral standards and codes of behavior as well as the legal requirements that organizations must consider when conducting business. Ethical considerations typically address topics such as transparency, privacy, fairness, rights and responsibilities, and safety, while regulatory considerations normally involve a range of laws, regulations, and policies that business must adhere to in order to remain compliant [24]. Ethical and regulatory considerations are necessary to protect the rights and welfare of customers, employees, and other stakeholders. In addition, such considerations can promote industry stability, promote public trust, and protect the environment.
14.8.1 Ensuring transparency and accountability in AI-powered optimization Ensuring transparency and accountability in AI-powered optimization is an important step in mitigating potential biases and ethical dilemmas created by the increasing use of AI technologies. Transparency and accountability should be considered when developing and deploying AI for decision-making. This entails the creation of welldocumented processes, cybersecurity protocols, and data collection processes that are open and transparent. Additionally, organizations should regularly evaluate and monitor their AI optimization processes for any changes that may lead to potential biases or unethical outcomes that might arise from the deployment of AI. Organizations should also ensure that the data used to train the AI model is appropriate and accurately reflects the objectives and goals of the organization [25]. Furthermore, organizations should develop measures that monitor the efficacy of the AI model and can be used to evaluate the implications of the decisions made by AI. This oversight and monitoring of AI, coupled with transparency and accountability, will help ensure that AI-powered optimization processes are effective and ethical.
302
B. Girimurugan et al.
14.8.2 Ethical implications of AI algorithms in multi-vendor applications AI algorithms used in multi-vendor applications can have serious ethical implications, as they can make decisions that impact the autonomy of an individual or organization. For example, an AI algorithm may select the best vendor for a particular task, but this decision could be based on factors that are difficult to identify or control, such as socio-economic status, gender, race, or other demographic data. This could lead to bias and discrimination, which may have serious ethical implications. Additionally, because AI algorithms can automate decision-making, it is difficult to ensure accountability and transparency of the decision-making process [26]. Finally, due to the complexity of AI algorithms, the outcomes of these algorithms can be hard to anticipate, and can lead to unexpected and unintended consequences.
14.8.3 Compliance with data protection and privacy regulations Compliance with data protection and privacy regulations is a critical issue for any organization that stores, processes, or transmits personal data. It refers to the laws, regulations, and other requirements that organizations must adhere to in order to ensure that data security and privacy are properly safeguarded. This includes laws such as the EU General Data Protection Regulation (GDPR), the International Organization for Standardization (ISO) 27001, and any government regulations applicable to data privacy. Data protection and privacy compliance involves crafting and following clear processes for collecting, storing, modifying, encrypting, securing, and disposing of data – and ensuring that employees understand their data protection responsibilities. Organizations must also create processes to regularly monitor how data privacy is being maintained. Adhering to data protection and privacy regulations helps organizations protect the data of their customers, partners, and employees as well as their own business reputation.
14.8.4 Responsible AI usage and mitigating biases in decision-making Responsible AI usage requires a holistic approach that considers accuracy, fairness, accountability, transparency, data quality, and a variety of other dimensions. It is important to avoid introducing biases in AI models, such as those related to race, gender, cultural backgrounds, or personal preferences, as these could result in biased or unfair decisions from the decision-making process [27]. To mitigate this, organizations must take a more comprehensive look at the data used to train models, understand the factors that may influence the model’s accuracy, and pay extra attention to poten-
14 The smart and secured AI-powered strategies for optimizing processes
303
tial bias. In addition, organizations should ensure that their data privacy and security policies are up-to-date and that they are compliant with applicable laws. Finally, organizations should ensure that they have a process and mechanism for monitoring and detecting bias in their models. This includes using a suite of data-quality tools as well as an AI ethics framework to help guide data collection and usage decisions.
14.9 Future directions and emerging trends The future of AI is a rapidly evolving area. Research and development into AI has shown tremendous potential, with experts expecting AI to become smarter and more advanced in the coming years. AI is becoming a key technology for tackling many challenging societal problems in areas such as healthcare, education, and energy. AI emerging trends include DL and reinforcement learning, natural language processing, robotics, facial recognition, and predictive analytics. DL is the process of teaching machines to learn from data, enabling them to recognize patterns, accurately interpret data, and react accordingly. Reinforcement learning allows machines to learn from their successes and failures by giving them rewards or punishments for their behaviors. Natural language processing allows machines to understand the meaning of words, and apply it to tasks like having basic conversations, text summarization, and natural language search. Robotics is allowing machines to learn from visual data, such as images and videos, and interact with the environment. Facial recognition systems use algorithms to detect faces in images and videos, and can be used in applications such as facial authentication for security purposes. Predictive analytics allows computers to predict future outcomes or identify potential problems by drawing from historical data. All of these AI and robotics trends are expected to continue driving AI development in the coming years.
14.9.1 Advancements in AI technologies and their impact on process optimization Advancements in AI technologies have led to significant improvements in process optimization. AI can automate routine tasks and processes, allowing organizations to focus more of their time and resources on higher-value activities. AI-based optimization algorithms can also help organizations discover new insights about their operations by performing complex analysis and identifying inefficiencies. AI-based automation can significantly reduce the need for manual input and help streamline processes. AI-based algorithms can also analyze large volumes of data quickly to identify patterns that may be indicative of inefficiencies or unwanted results. Once identified, the algorithms can recommend solutions or changes to help streamline processes and optimize outcomes
304
B. Girimurugan et al.
[28]. AI-based predictive analytics can be used to anticipate outcomes or trends, enabling organizations to adjust their operations in real time to stay ahead of changes or challenges. This proactive approach tends to improve operational efficiency, as organizations can adjust in advance to changes in the market or environment. AI technologies and their applications have had a positive impact on process optimization. Organizations now have the potential to gain greater insights into their operations and streamline their processes more effectively than ever before.
14.9.2 Integration of AI with emerging technologies like blockchain and IoT The integration of AI with emerging technologies like blockchain and IoT has tremendous potential to revolutionize processes and empower businesses. Integrating AI with blockchain can facilitate secure storage and transfer of data, as blockchain is an inherently secure and distributed ledger technology. This makes it easier to monitor and store data while reducing the risk of data breaches. In addition, integrating AI with IoT increases the possibilities for businesses to use data collected from their IoT devices from multiple remote and diverse sources. This data can also be used to identify patterns and correlations in order to generate insights that can help businesses improve their operations and services. The integration of AI with emerging technologies like blockchain and IoT can also be used to automate various tasks. This could include anything from monitoring changes in the market to providing more tailored solutions for customers. These technologies can also be used to develop tools and applications that automate mundane tasks in order to free up resources and boost productivity.
14.9.3 Evolution of AI governance frameworks for multi-vendor applications AI governance frameworks are a set of tools, processes, and decisions designed to help organizations manage the deployment and use of AI technology. These frameworks provide structure and guidelines to ensure that AI applications are used responsibly and efficiently. This includes setting parameters for data protection, decision management, and ethical application of AI technologies. As the use of AI technology has grown, so has the need for more sophisticated governance frameworks. Multi-vendor applications, such as those used in multi-cloud environments, often require more complex governance models. For example, each cloud vendor has its own unique security and compliance requirements, and the combination of multiple vendors can create additional complexities [29]. In order to address these complexities, a number of AI governance frameworks have emerged that are designed to provide guidance and structure to
14 The smart and secured AI-powered strategies for optimizing processes
305
multi-vendor AI deployments. These frameworks generally encompass four key areas: data privacy, decision management, risk management, and compliance [30]. For example, the Open Data Governance Framework (ODGF) provides a comprehensive set of tools and guidelines for organizations to use in order to establish and maintain responsible and secure management of AI-powered applications. The evolution of these AI governance frameworks reflects the growing recognition of the need for standardized and secure management of AI-powered applications, particularly in multi-vendor environments. By providing organizations with a set of tools and guidelines for responsible AI deployment, these frameworks empower organizations to achieve their goals in a responsible and secure manner [31].
14.9.4 Potential challenges and opportunities in the future of AI-powered optimization Potential challenges in the future of AI-powered optimization include complexity in implementing the technology, data and privacy regulations, and limited resources. There is also a risk of unintentional bias being injected in the algorithms used for optimization, risking the accuracy and reliability of the outcomes received [32]. Additionally, as AIpowered optimization technology continues to develop, it is important that business leaders remain aware of the potential legal implications of leveraging this technology in order to protect their businesses from potential liabilities [33, 34]. On the other hand, AIpowered optimization presents a number of opportunities. It can be used to automate tedious processes, increase decision-making accuracy, reduce costs, and optimize business operations for increased efficiency. Additionally, AI-powered optimization makes it easier to detect subtle patterns from large amounts of data, enabling businesses to gain deep insights into the performance of their operations and make personalized decisions with greater accuracy and confidence. Lastly, the technology can help to reduce human errors, resulting in improved customer service and overall customer experience [35, 36].
References [1] [2]
[3]
Xin, Y., Yang, K., Chih-Lin, I., Shamsunder, S., Lin, X., & Lai, L. (2023). Guest Editorial: AI-Powered Telco Network Automation: 5G Evolution and 6G. IEEE Wireless Communications, 30(1), 68–69. Militano, L., Zafeiropoulos, A., Fotopoulou, E., Bruschi, R., Lombardo, C., Edmonds, A., & Papavassiliou, S. (2021, December). Ai-powered infrastructures for intelligence and automation in beyond-5g systems. In 2021 IEEE Globecom Workshops (GC Wkshps) (pp. 1–6). New York City: IEEE. Asif, R., & Ghanem, K. (2021, January). Ai secured SD-WAN architecture as a latency critical IOT enabler for 5G and beyond communications. In 2021 IEEE 18th Annual Consumer Communications & Networking Conference (CCNC) (pp. 1–6). IEEE.
306
[4] [5] [6]
[7] [8]
[9]
[10] [11]
[12] [13]
[14]
[15]
[16]
[17]
[18]
[19] [20] [21]
B. Girimurugan et al.
Nguyen, H. X., Trestian, R., To, D., & Tatipamula, M. (2021). Digital twin for 5G and beyond. IEEE Communications Magazine, 59(2), 10–15. Gallego-Madrid, J., Sanchez-Iborra, R., Ortiz, J., & Santa, J. (2023). The role of vehicular applications in the design of future 6G infrastructures. ICT Express. Zeb, S., Mahmood, A., Khowaja, S. A., Dev, K., Hassan, S. A., Qureshi, N. M. F., . . . & Bellavista, P. (2022). Industry 5.0 is coming: A survey on intelligent nextG wireless networks as technological enablers. In 2022 IEEE International Conference on Networking and Internet Architecture (pp. 1–8). Singh, R., & Gill, S. S. (2023). Edge AI: A survey. Internet of Things and Cyber-Physical Systems, 3, 71– 92. Al-Mekhlal, M., AlYahya, A., Aldhamin, A., & Khan, A. (2022, August). Network automation pythonbased application: The performance of a Multi-Layer Cloud Based Solution. In 2022 IEEE International Conference on Omni-layer Intelligent Systems (COINS) (pp. 1–8). IEEE. Peltonen, E., Bennis, M., Capobianco, M., Debbah, M., Ding, A., Gil-Castiñeira, F., . . . & Yang, T. (2020). 6G white paper on edge intelligence. In 2020 IEEE International Conference on Distributed, Parallel, and Cluster Computing (pp. 1–6). Ordonez-Lucena, J., Ameigeiras, P., Contreras, L. M., Folgueira, J., & López, D. R. (2021). On the rollout of network slicing in carrier networks: A technology radar. Sensors, 21(23), 8094. Pham, X. Q., Huynh-The, T., & Kim, D. S. (2022, October). A vision of distributed cloud computing. In 2022 13th International Conference on Information and Communication Technology Convergence (ICTC) (pp. 1963–1968). IEEE. Ekudden, E. (2021). Five network trends: Towards the 6G Era. Ericsson Technology Review, 2021(9), 2–10. Thiruvasagam, P. K., Venkataram, V., Ilangovan, V. R., Perapalla, M., Payyanur, R., & Kumar, V. (2023). Open RAN: Evolution of architecture, deployment aspects, and future directions. In 2023 IEEE International Conference on Networking and Internet Architecture (pp. 1–8). Pham, X. Q., Nguyen, T. D., Huynh-The, T., Huh, E. N., & Kim, D. S. (2022). Distributed cloud computing: Architecture, enabling technologies, and open challenges. IEEE Consumer Electronics Magazine, 12(3), 98–106. Wu, Y., Dai, H. N., Wang, H., Xiong, Z., & Guo, S. (2022). A survey of intelligent network slicing management for industrial IoT: Integrated approaches for smart transportation, smart energy, and smart factory. IEEE Communications Surveys and Tutorials, 24(2), 1175–1211. Singh, S., Karimipour, H., HaddadPajouh, H., & Dehghantanha, A. (2020). Artificial intelligence and security of industrial control systems. In Handbook of Big Data Privacy, United States: Springer International Publishing, 2020, pp. 121–164. Henttinen, T. (2021). AI SUPPORTED PURCHASE INVOICE POSTING: Lessons learned from piloting project and future perspectives Master Thesis, Department of Business Management, Turku University of Applied Science, pp.1–59. Firouzi, F., Jiang, S., Chakrabarty, K., Farahani, B., Daneshmand, M., Song, J., & Mankodiya, K. (2022). Fusion of IoT, AI, Edge-Fog-Cloud, and Blockchain: Challenges, solutions, and a case study in healthcare and medicine. IEEE Internet of Things Journal, 10(5), 3686–3705. Yadav, S. P., Agrawal, K. K., Bhati, B. S., Al-Turjman, F., & Mostarda, L. (2022). Blockchain-based cryptocurrency regulation: An overview. Computational Economics, 59(4), 1659–1675. Irfan, D., Tang, X., Narayan, V., Mall, P. K., Srivastava, S., & Saravanan, V. (2022). Prediction of Quality Food Sale in Mart Using the AI-Based TOR Method. Journal of Food Quality, 2022, 1–13. Sivasankari, S. S., Surendiran, J., Yuvaraj, N., Ramkumar, M., Ravi, C. N., & Vidhya, R. G. (2022, April). Classification of diabetes using multilayer perceptron. In 2022 IEEE International Conference on Distributed Computing and Electrical Circuits and Electronics (ICDCECE) (pp. 1–5). IEEE.
14 The smart and secured AI-powered strategies for optimizing processes
307
[22] Avinash Khatri, K. C., Shah, K. B., Logeshwaran, J., & Shrestha, A. (2023). Genetic algorithm based techno-economic optimization of an isolated hybrid energy system. ICTACT Journal on Microelectronics, 8(4), 1447–1450. [23] Yadav, S. P., & Yadav, S. (2020). Fusion of medical images in wavelet domain: A hybrid implementation. Computer Modeling in Engineering and Sciences, 122(1), 303–321. [24] Rajalakshmi, M., Saravanan, V., Arunprasad, V., Romero, C. T., Khalaf, O. I., & Karthik, C. (2022). Machine learning for modeling and control of industrial clarifier process. Intelligent Automation and Soft Computing, 32(1), 339–359. [25] Veerappan Kousik, N. G., Natarajan, Y., Suresh, K., Patan, R., & Gandomi, A. H. (2020). Improving power and resource management in heterogeneous downlink OFDMA networks. Information, 11(4), 203. [26] Ramesh, G., Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). Prediction of energy production level in large PV plants through AUTO-Encoder Based Neural-Network (AUTO-NN) with restricted Boltzmann feature extraction. Future Internet, 15(2), 46. [27] Yadav, S. P., & Yadav, S. (2018). Fusion of medical images in wavelet domain: A Discrete Mathematical Model/Fusion de imagenes medicas en el ambito de la transformada de ondicula: modelo matematico discreto/Fusao de imagens medicas no ambito da transformada de wavelet: Modelo matematico discreto. Revista Ingenieria Solidaria, 14(25), NA–NA. [28] Saravanan, V., & Neeraja, A. (2013, January). Security issues in computer networks and stegnography. In 2013 7th International Conference on Intelligent Systems and Control (ISCO) (pp. 363–366). IEEE. [29] Brilly Sangeetha, S., Wilfred Blessing, N. R., Yuvaraj, N., & Adeline Sneha, J. (2020). Improving the training pattern in back-propagation neural networks using holt-winters’ seasonal method and gradient boosting model. Applications of Machine Learning, 2020, 189–198. [30] Adhikari, N., Logeshwaran, J., & Kiruthiga, T. (2022). The artificially intelligent switching framework for terminal access provides smart routing in modern computer networks. BOHR International Journal of Smart Computing and Information Technology, 3(1), 45–50. [31] Pujahari, R. M., Yadav, S. P., & Khan, R. (2022). Intelligent farming system through weather forecast support and crop production. In: Mohammad Ayoub Khan, Rijwan Khan and Mohammad Aslam Ansari (Eds.). Application of Machine Learning in Agriculture (pp. 113–130). United States: Academic Press. [32] Khan, S., Saravanan, V. N. G. C., Lakshmi, T. J., Deb, N., & Othman, N. A. (2022). Privacy protection of healthcare data over social networks using machine learning algorithms. Computational Intelligence and Neuroscience, 2022, 1–12. [33] Yuvaraj, N., Praghash, K., & Karthikeyan, T. (2022). Data privacy preservation and trade-off balance between privacy and utility using deep adaptive clustering and elliptic curve digital signature algorithm. Wireless Personal Communications, 124(1), 655–670. [34] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. L.-R. U. B. I. (2023). An efficient load-based resource utilization algorithm for bi-partite scatternet in wireless personal area networks. International Journal of Communication Systems, 36(4), 1–17. [35] Vermesan, O., Coppola, M., Nava, M. D., Capra, A., Kornaros, G., Bahr, R., . . . & McGrath, S. (2020). New waves of IoT technologies research–transcending intelligence and senses at the edge to create multi experience environments. Internet of Things–The Call of the Edge. Everything Intelligent Everywhere, Europe: River Publisher, 1–168. [36] Sandeepa, C., Siniarski, B., Kourtellis, N., Wang, S., & Liyanage, M. (2022). A survey on privacy for B5G/6G: New privacy challenges, and research directions. Journal of Industrial Information Integration, In 2022 IEEE International Conference on Cryptography and Security (pp. 1–49).
S. Rajeshwari, D. Praveenadevi, S. Revathy, S. P. Sreekala
15 Utilizing AI technologies to enhance e-commerce business operations Abstract: E-commerce has become an increasingly important sector for businesses, offering customers convenient and fast access to goods and services. With the proliferation of technological advancement, particularly Artificial intelligence (AI) technology, e-commerce continues to develop and evolve. AI technologies such as machine and deep learning can be utilized to enhance e-commerce operations. Companies can employ AI-based solutions for personalized product recommendations, automated customer service, optimized search and user interfaces, and improved security. By utilizing AI technologies, e-commerce companies can improve customer experiences, reduce operational costs, and make business operations more efficient. Keywords: E-commerce, business, customers, goods, technology, machine learning, deep learning
15.1 Introduction to AI in e-commerce Artificial intelligence (AI) technologies can also be used to develop predictive models to improve product recommendations, predict market trends, and automate pricing strategies. AI-based solutions, such as chatbots, can help automate customer service functions. AI can also help enhance user experience by optimizing the search experience and providing personalized recommendations based on customer preferences. Furthermore, AI-based technologies can provide improved security, helping to detect fraud, protect data, and optimize inventory [1]. By utilizing AI technologies, e-commerce companies can enhance customer service and expand operational capabilities.
15.1.1 Overview of the chapter’s focus and objectives The focus and objectives of AI in e-commerce is to provide a more personalized and efficient customer experience. AI can be used to tailor product recommendation, suggest promotions and discounts, and automate various customer service related tasks. AI can also be used to process purchases and payments more quickly and securely, thereby boosting customer loyalty [2]. AI can even be used to analyze customer data to create personalization strategies that improve conversion rates. By utilizing AI, merchants can improve their overall customer experience and generate more sales.
https://doi.org/10.1515/9783111323749-015
310
S. Rajeshwari et al.
15.1.2 Importance of AI technologies in enhancing e-commerce operations AI technologies are playing an increasingly important role in the e-commerce industry. These technologies can provide businesses with data-driven insights that enable them to improve efficiency, increase sales, and deliver a better customer experience. AIdriven automation, personalization, and analytics can be used to reduce operational costs, develop automated workflows, and offer tailored experiences to customers. AI can also be used to power predictive analytics, enabling businesses to anticipate customer behavior and optimize operations accordingly. Additionally, AI’s conversational capabilities can be used to power virtual assistants, providing customers with a more convenient and personalized shopping experience. Finally, by using AI-driven insights, businesses can fine-tune their marketing strategies to reach more customers online. In short, AI technologies are playing a major role in improving the efficiency and effectiveness of businesses’ e-commerce operations.
15.1.3 Potential benefits and challenges of implementing AI in the e-commerce industry Potential benefits: – Improved user experience: AI can be used to improve the user experience. It can recognize customer buying patterns and offer personalized recommendations, providing customers with more targeted and relevant information. AI can also help reduce shopping cart abandonment by reminding customers of items they’ve left in their carts and sending personalized notifications during the shopping process. – Enhanced security and fraud prevention: AI can be used to detect suspicious behavior and spot fraudulent activities, reducing the risk of online payments and frauds. AI can also be used to automate security measures such as face and voice recognition, two-factor authentication and more. – Increased productivity: AI can help automate mundane tasks, freeing up employees to focus on more strategic and meaningful tasks. AI can also be used to help automate order processing and fulfillment, resulting in increased productivity and fewer delays. Potential challenges: – Cost: AI can be expensive to implement and maintain. Companies need to invest in hardware, software, and specialized teams to properly use and manage AI. – Data quality: AI relies on data to produce accurate results. Poor quality data can lead to inaccurate results, which can be costly for businesses.
15 Utilizing AI technologies to enhance e-commerce business operations
–
311
Privacy and security: AI systems collect and process sensitive customer data, creating security risks and potential privacy violations. Companies must set policies and procedures in place to protect customer data and ensure it is used responsibly.
15.2 AI-powered product recommendation systems AI-powered product recommendation systems are computer-based algorithms which use machine learning and AI technologies to provide personalized product recommendations to end users. These recommendation systems are based on user behavior, demographics, and past product sales. They enable companies to deliver personalized product recommendations to their customers for improved customer engagement and conversions [3]. AI-Powered Product Recommendation systems are used in e-commerce stores for an optimized shopping experience, such as Amazon and Netflix. These systems enable them to offer personalized recommendations based on customer preferences and usage history. AI-powered recommendation systems also help reduce the effort customers need to go through in finding items that match their preferences.
15.2.1 Personalized product recommendations using collaborative filtering and content-based filtering Personalized product recommendations using collaborative filtering and contentbased filtering are two popular recommendation algorithms that businesses employ to provide personalized product or service recommendations. Collaborative filtering is based on extracting data from the user’s past interactions (e.g. ratings, purchases). Then, the algorithm determines patterns of user-product relationships and makes a prediction of which items the user would prefer. Content-based filtering, on the other hand, uses the attributes of the products and user’s profiles to determine the similarity between items and users; thus, it tends to recommend products that share similar attributes [4]. Companies are increasingly leveraging these two methods to offer more personalized and accurate product recommendations, improve user experience, and increase sales.
15.2.2 Recommendation algorithms such as collaborative filtering, association rules, and deep learning models –
Collaborative filtering is a method of making recommendations by finding patterns in users’ past behaviors. It works by predicting what a user is likely to find interesting based on their past behavior and the behavior of other similar users.
312
–
–
S. Rajeshwari et al.
Association rules are a type of machine learning algorithm used to find relationships between items in large datasets. This type of algorithm can be used for recommendation systems, market basket analysis, and other areas. Deep learning models are a type of machine learning algorithm used to model complex relationships between input and output data. Deep learning models are used in a wide variety of applications, such as object detection, image recognition, natural language processing (NLP), and recommendation systems. These models can analyze large datasets and learn patterns and trends that are difficult for humans to identify.
15.2.3 Implementing recommendation systems for cross-selling and upselling opportunities Implementing recommendation systems for cross-selling and upselling opportunities allows companies to make personalized product and service recommendations to their customers based on their past purchase history, interest, preferences, and more. Cross-selling is recommending similar products to existing customers. Upselling is about offering customers better quality products and services for a higher price. Recommendation systems help companies improve customer experience by making more accurate and relevant recommendations that can be personalized to each customer, thereby increasing the chances of customers buying these recommended items [5]. Companies can also use recommendation systems to identify uprecommendations and cross-sell opportunities to existing customers. They can also leverage these systems to identify opportunities for customer segmentation and targeting. By utilizing recommendation systems, companies can provide customers with more relevant and personalized recommendations, which not only drives more sales but also enhances customer satisfaction and loyalty.
15.3 Intelligent search and natural language processing Intelligent search and NLP are two closely related concepts when it comes to understanding how computers can understand text and retrieve relevant information. 1. Intelligent search uses technologies such as AI, NLP, data mining, and machine learning to enable computers to understand and query written language and identify meaningful information from it. 2. NLP is the field of computer science which is dedicated to allowing computers to interpret and understand natural language. It helps machines, such as computers and mobile devices, to understand, process, and respond to human language. NLP
15 Utilizing AI technologies to enhance e-commerce business operations
313
involves techniques such as entity extraction, sentiment analysis, natural language understanding, and semantic search. These techniques help machines to interpret the meaning of words and phrases and determine the context in which they are used. For example, NLP can be used to create custom web search tools that understand a range of search queries in a more sophisticated way than a typical web search engine. NLP can also be used to create chatbots that can interact with users in a natural, humanlike manner. Together, intelligent search and NLP can enable computers to process natural language in a more meaningful way and generate more intelligent search results. This can help to provide users with more accurate and relevant information, as well as create an improved user experience for interface designers.
15.3.1 AI-driven search engines and intelligent search algorithms AI-driven search engines are search engines that use AI to provide more accurate and relevant search results. AI-driven search engines can incorporate machine learning algorithms, NLP algorithms, and use deep learning techniques to understand user intent better and return more comprehensive results for a given query [6]. These techniques are used to improve the accuracy of the search results and minimize the amount of time required to sift through irrelevant information. Additionally, AIdriven search engines are leveraging intelligent search algorithms to return the most up-to-date results so that users always have the most current information. Intelligent search algorithms dynamically adjust search parameters based on the latest data from a variety of sources in order to return the most relevant set of results. This results in more targeted and relevant search results and enables users to find the most accurate and up-to-date answers to their search queries.
15.3.2 NLP techniques for improved search results and customer interactions NLP is a type of AI technology that enables computers to understand and interpret human language in a more natural, efficient, and accurate way. NLP is used in a variety of applications, from dialogue systems that power customer service automation, to spell checkers, to natural language search engines. NLP can be used to improve search results by allowing for more sophisticated search queries, such as natural language queries, rather than just exact keywords [7]. NLP can also be used to improve customer interactions by using natural language understanding to interpret customer queries more accurately, as well as by incorporating personalized information into responses and conversations.
314
S. Rajeshwari et al.
15.3.3 Voice search and chatbot applications in e-commerce Voice search and chatbot applications are becoming increasingly important for e-commerce businesses as they allow customers to quickly access information and easily make transactions. Voice search allows customers to search for products and services in natural language, making it easier and faster for customers to find what they are looking for. Chatbot applications provide customers with an interactive experience that is personalized to their needs and can help to provide quick answers to customer queries. Chatbots can also help to provide customers with promotional offers, further increasing customer engagement [8]. These applications provide a more immersive and personalized shopping experience for customers and can lead to customer loyalty and increased sales. In addition, these applications can help businesses to reduce costs associated with providing customer service, as they can take over many of the duties performed by employees.
15.3.4 Enhancing customer experience and conversion rates through AI-powered search capabilities AI-powered search capabilities are becoming increasingly important for enhancing customer experience and boosting conversion rates. AI-powered search capabilities enable businesses to deliver a personalized experience for their customers by providing insight into their preferences and enabling them to find what they’re looking for more quickly and effectively than ever before. AI-powered search capabilities leverage machine learning to analyze customer behavior and search patterns in order to better understand their tastes and preferences. The AI-powered search capabilities also use NLP to understand customer queries and provide recommendations based on results from past searches [9]. This allows customers to find what they’re looking for in fewer clicks and makes it easier to navigate and identify products or services of interest. By customizing the search experience, businesses can provide a more personalized experience for customers, which can lead to better customer satisfaction and higher conversion rates.
15.4 AI for customer segmentation and targeted marketing AI for customer segmentation and targeted marketing (AI-CSTM) is a data-driven marketing approach that uses AI and machine learning algorithms to identify customer segments, target different customer segments with personalized marketing campaigns, and evaluate the effectiveness of these campaigns. AI-CSTM uses data from
15 Utilizing AI technologies to enhance e-commerce business operations
315
customer behaviors, demographics, interests, and actions to identify key customer segments and target them with specific campaigns. By leveraging AI algorithms to analyze customer data, marketers are better able to identify potential customers, create more personalized campaigns, and measure their results [10]. AI-CSTM can also be used to optimize pricing in order to maximize customer lifetime value. AI-CSTM can help marketers quickly and efficiently identify target customer segments, target them with appropriate campaigns, and measure the results in order to continuously refine and improve the campaigns.
15.4.1 AI-based customer segmentation techniques using clustering and classification algorithms AI-based customer segmentation techniques involve the use of computer algorithms that can identify patterns in customer data and divide customers into distinct clusters or segments. Clustering algorithms, such as k-means, can be used to group customers based on their preferences, behaviors, or other characteristics. Classification algorithms, such as decision trees and logistic regression, can be used to create decision criteria and assign unknown customers into a predefined segment [11]. This approach enables companies to tailor their marketing strategies to different customer segments, helping them to more effectively reach their desired target audiences.
15.4.2 Predictive analytics for personalized marketing campaigns and customer retention strategies Predictive analytics for personalized marketing campaigns and customer retention strategies is the process of using data-driven techniques to create and improve marketing campaigns and customer retention strategies. By looking at customer behavior, companies can gain valuable insights into who their target customers are and how best to reach them. This allows for improved personalization and better targeting of campaigns, leading to increased conversion rates and customer lifetime value. Predictive analytics also enables companies to identify customer churn signals, allowing them to improve their customer retention strategies. Companies can use predictive analytics to predict customer trends and identify likely customer needs that can be addressed in marketing campaigns [12]. This helps establish strong relationships with customers that ultimately lead to better customer experiences and loyalty.
316
S. Rajeshwari et al.
15.4.3 Sentiment analysis and social media mining for understanding customer preferences Sentiment analysis and social media mining are two powerful techniques for understanding customer preferences. Sentiment analysis is a method of analyzing sentiment or opinion from text data, usually collected from social media platforms. It is an important tool for learning customer behaviors and preferences and can provide useful insights for marketers and product developers. Social media mining involves mining information from online sources, including forums, blogs, tweets, and other sources, to determine customer sentiment [13]. This type of analysis is especially useful when trying to measure customer satisfaction with a product or service. By understanding what customers are saying, and the sentiment around it, companies can create better experiences for their customers.
15.5 Fraud detection and security in e-commerce Fraud detection and security in E-commerce is a critical component of e-commerce success. E-commerce stores are vulnerable to fraudulent activity from hackers, cybercriminals, and unethical individuals, which can have a major financial and reputational impact on the business. To prevent such activities, it is essential to have an effective fraud detection and security strategy in place. A fraud detection and security system typically uses a combination of technological, procedural, and human methods to detect and prevent fraudulent activity. Technologies such as AI, machine learning, and neural networks are increasingly being used to identify fraudulent trends and patterns [14]. These methods can block fraudulent transactions, identify suspicious accounts, and build a risk profile of customers. Procedural methods such as requiring certain documents for verification, implementing customer verification procedures, and implementing enhanced authentication protocols are also effective means for detecting fraud. Additionally, businesses should use strong passwords and encrypt data to improve security. Finally, human intervention is sometimes the best option for thoroughly screening high-risk customers and catching signs of suspicious activity. Customer service agents and call center representatives should address customer concerns quickly, while also gathering pertinent information. This information can then be used to generate a risk profile of the customer that can be used for further investigation. Ultimately, fraud detection and security in e-commerce is essential for a successful business. It is critical for businesses to develop a comprehensive strategy that includes a mix of technological, procedural, and human techniques to stay ahead of the ever-evolving online threats.
15 Utilizing AI technologies to enhance e-commerce business operations
317
15.5.1 AI-powered fraud detection algorithms for identifying fraudulent activities and transactions AI-powered fraud detection algorithms use machine learning and AI techniques to identify fraudulent activities and transactions. The algorithms profile normal user behavior and detect anomalies that can indicate suspicious activity [15]. The algorithms use advanced statistical techniques to identify abnormal patterns that may indicate fraud, and then use predictive models to predict the likelihood of fraud. These algorithms can also learn with time, so the more data it is fed, the more accurate it can become at identifying fraudulent activities.
15.5.2 Behavioral analytics and anomaly detection techniques Behavioral analytics is the practice of using, collecting and analyzing data pertaining to user behavior to better understand and inform decisions about product, marketing, and other aspects of modern businesses. Anomaly detection techniques are used to detect suspicious behavior or patterns that don’t conform to expected norms. They can be used to detect fraud, malware, and other kinds of malicious activities [16]. Anomaly detection techniques typically involve applying machine learning algorithms to identify suspicious or abnormal patterns. These techniques leverage data mining, statistical methods, text analytics, and other techniques to analyze and identify unusual or outlier events in a dataset.
15.5.3 Addressing security and privacy concerns in e-commerce using AI technologies AI technologies can be used to address security and privacy concerns in e-commerce. AI can be used to detect fraud by analyzing large amounts of data, identifying patterns, and flagging potential suspicious activity. This can be used to detect credit card fraud, identity theft, and other malicious activities. AI can also be used to strengthen authentication methods by recognizing user behavior and verifying user identities. Additionally, AI can be used to improve user privacy by performing automatic data encryption and anonymization [17]. By encrypting data as soon as it is entered into the system, data is safeguarded against hackers and unauthorized personnel. AI can also be used to implement granular access controls, which helps ensure that only authorized users are able to access sensitive data. Finally, AI can be used to detect anomalous activity, which can protect e-commerce platforms from potential threats.
318
S. Rajeshwari et al.
15.5.4 Real-time monitoring and proactive measures for fraud prevention Real-time monitoring is a process of using AI and advanced analytics to detect fraud attempts in (near)-real time. It can be used to identify suspicious or irregular transaction patterns, analyze data across multiple platforms, review alerts for potential fraud attempts and even proactively help prevent fraud before it happens [18]. Proactive measures for fraud prevention are steps taken to avert or prevent fraud before it occurs. These measures include monitoring transactions, conducting enhanced due diligence to identify risks, reviewing customer information and activity, and using authentication protocols to verify customer identity. Other proactive fraud prevention measures include implementing know your customer policies, increasing transparency for customers, implementing data encryption for data, and implementing a comprehensive fraud detection system that monitors customer activity [19].
15.6 AI-enabled supply chain management AI-enabled supply chain management is a process where AI is used to improve the supply chain process and operations. AI solutions are used to automate various supply chain operations such as supply forecasting, optimization, pricing, and inventory management. AI can take into consideration data ranging from individual inventory items to large-scale supply chain trends. This helps to identify the most efficient, cost-effective and timely solution for a supply chain process. AI can even help predict customer demand, reduce costs, and increase customer satisfaction [20]. AI can also be used to prevent theft, fraud, and counterfeiting in the supply chain. AI-enabled supply chain management provides a more efficient, optimized, and automated supply chain solution, thereby improving customer satisfaction, reducing costs, and increasing profits.
15.6.1 AI-driven demand forecasting and inventory optimization AI-driven demand forecasting and inventory optimization is a process that helps businesses to accurately predict customer demand for products and services. It uses AI to determine the best inventory levels to optimize profits while minimizing waste and inefficiencies. AI algorithms collect data from sources such as seasonality, customer feedback, past order rates, environmental factors, and economic changes to create predictions. These predictions are then used to calculate the optimal inventory levels for the company to keep on hand [21]. AI-driven demand forecasting and inventory optimization enable businesses to reduce the cost of indexing, increase customer satisfaction through better service, and improve stock management.
15 Utilizing AI technologies to enhance e-commerce business operations
319
15.6.2 Intelligent logistics and route optimization for efficient supply chain operations Intelligent logistics and route optimization for efficient supply chain operations are processes used to optimize the flow of goods through a supply chain to reduce costs, transit time, and resources. They use sophisticated algorithms and AI to evaluate and re-evaluate supply chain elements, including order fulfillment, routing, inventory, scheduling, distribution, and cost/benefit calculations. This helps to create an efficient supply chain that can adapt to changing customer and business demands. Route optimization uses algorithms to generate the most cost-effective and timely routes for delivery vehicles and shipments [22]. AI algorithms are used to examine multiple route options and consider the impact of traffic, warehouse times, order quantities, and other variables to generate the most efficient route. Advanced technologies like predictive analytics can also be used to detect trends and anticipate demand to better optimize routes. Intelligent logistics and route optimization are essential to modern supply chain operations, ensuring that goods are delivered on time and at the lowest possible cost.
15.6.3 Predictive maintenance and quality control using AI algorithms Predictive maintenance and quality control using AI algorithms are two different aspects in the analytics field. Predictive maintenance is used to identify mechanical problems prior to them occurring, so that they can be repaired or replaced before they become an issue. It is an integral component in improving efficiency and performance of products and reducing costs incurred due to unexpected repairs. Quality control using AI algorithms involves applying AI algorithms to measure and assess the quality of manufactured products. These algorithms help to identify and classify faults in production and to ensure that the manufactured products match the desired specifications [23]. AI algorithms can be applied to detect faulty products and to identify the source of the problem allowing for proactive resolution. Both predictive maintenance and quality control using AI algorithms are effective methods for reducing costs and improving the overall efficiency and performance of products.
15.7 AI for pricing optimization and dynamic pricing AI for pricing optimization and dynamic pricing is a technology that helps automate the price-setting process. It uses sophisticated algorithms to analyze market trends, customer segment data, competitor analysis, and other data points in order to optimize pri-
320
S. Rajeshwari et al.
ces and maximize revenue. This technology offers an efficient and accurate solution to the issue of manual price-setting, which can be labor-intensive and prone to errors. AIdriven dynamic pricing algorithms can quickly respond to fluctuations in demand and competitor prices, facilitated by the use of machine learning models that continuously learn and optimize in real-time. The result is optimal pricing customized to the customer’s needs and continuously adjusted to suit changing market conditions.
15.7.1 AI algorithms for dynamic pricing and competitive pricing analysis AI algorithms for dynamic pricing are designed to optimize the pricing of products and services according to current market conditions. The algorithms look at factors like supply and demand, perceived value, and competitor pricing in order to optimize a product’s pricing. These algorithms enable companies to set and adjust prices in real time, automatically responding to changes in the market. Additionally, AI algorithms can help businesses analyze competitor pricing strategies and identify opportunities to set competitive prices [24]. By taking into account market trends, product availability, and an understanding of consumer behavior, these algorithms can provide businesses with valuable insights on how to set competitive prices. Ultimately, using AI algorithms for dynamic pricing and competitive pricing analysis can help businesses make better pricing decisions and maximize their profits.
15.7.2 Price optimization techniques using machine learning and optimization algorithms Price optimization techniques using machine learning and optimization algorithms refer to the practice of setting prices using the most accurate and profitable pricing strategies. This is accomplished by combining advanced data analysis and predictive modeling with mathematical optimization and algorithms. The purpose of this technique is to maximize profits while providing the optimal pricing for customers. The advanced data analysis and predictive modeling step of the process involves collecting data from various sources including competitors, customers, market demand, and market trends. This data is then used to create predictive models to identify target customer segments, average prices, and other trends. Using mathematical optimization and algorithms, this data is combined to generate optimal pricing techniques for the given market. Depending on the preferences and goals of the company, the optimized prices may be considered with a view to maximizing total revenues, total profits, or both [25]. The optimized prices are then tested against current market conditions and compared with competitors before being implemented. Price optimization can be used to monitor customer buying patterns over time and ensure that pricing strategies remain accurate
15 Utilizing AI technologies to enhance e-commerce business operations
321
and profitable. It can also be used to create highly customized and individualized pricing models for each customer. Additionally, price optimization techniques are becoming increasingly important for businesses in volatile markets, as they allow companies to quickly adjust prices in response to market trends.
15.7.3 Real-time pricing adjustments based on market trends and customer behavior Real-time pricing adjustments based on market trends and customer behavior are strategies used by sellers to maximize their profits by adjusting their prices in response to market trends and customer behavior. This approach allows sellers to adjust their prices in response to market signals, meaning they can take advantage of changing circumstances and capitalize on short-term opportunities in the marketplace. This strategy also allows sellers to compete more effectively with other sellers, as they can adjust their prices to match or beat those of their competitors [26]. Additionally, by analyzing customer behavior, sellers can identify customer segments and adjust their pricing accordingly, targeting different segments more effectively. Finally, this approach can help to build brand loyalty, as customers can view the price adjustments as a sign of the seller’s understanding of their needs and willingness to meet them.
15.7.4 Benefits and challenges of implementing AI in pricing strategies Benefits: – AI can help businesses better understand market trends and customer needs by analyzing data and allowing for the optimization of pricing strategies. This can lead to greater profits and customer satisfaction. – AI can provide businesses with more timely insight into customer behavior, which can help them create pricing strategies that are tailored to their specific audiences. – AI-driven pricing strategies can also help businesses reduce costs, as they can more accurately predict what customers are willing to pay. – Furthermore, AI can reduce human error and improve forecasting accuracy. Challenges: – AI requires a considerable amount of data in order to produce effective insights. Without adequate data, AI solutions may be unable to optimize pricing strategies. – Additionally, AI solutions are not foolproof and may produce inaccurate results. Businesses need to regularly audit and adjust their pricing strategies in order to ensure accuracy.
322
– –
S. Rajeshwari et al.
AI can also be expensive to implement and may require long-term investments. Finally, AI solutions can be challenging to integrate into existing systems. Businesses need to be sure to have a comprehensive plan in place when implementing AI solutions.
15.8 Enhanced customer support with AI Enhanced customer support with AI is a service that uses AI to provide customers with a personalized, omnichannel, and tailored service. It enables customers to interact with a customer service representative in a more conversational way. Because AI has the ability to understand natural language, customers can ask questions and receive tailored responses that are based on their preferences [27]. Additionally, AI can automate certain customer service processes, such as data entry or checkout, which can significantly reduce wait times and improve customer loyalty.
15.8.1 AI-powered chatbots and virtual assistants for customer support AI-powered chatbots and virtual assistants for customer support are online tools used to provide customers with an automated self-service experience and to interact with customers in real-time. They provide customers with the convenience of a simple interface to help them resolve their customer service needs quickly and efficiently. AIpowered chatbots and virtual assistants use AI technologies such as NLP and machine learning to respond to customer inquiries and requests. With the help of these technologies, a virtual assistant can understand customer inputs and accurately respond to customer questions, all without any human intervention [28]. Additionally, AIpowered chatbots and virtual assistants can leverage customer data analysis to better evaluate customers’ needs and requests in order to recommend the most suitable service.
15.8.2 NLP and sentiment analysis for customer interactions NLP is the technology and discipline of extracting meaningful information from natural language text. NLP can help automate customer interactions by allowing computers to understand natural language. Sentiment analysis is a type of NLP technique that uses algorithms to interpret a customer’s text to understand the sentiment the language conveys. It can be used to identify customer sentiment in emails, online reviews, surveys, social media posts, and other customer interactions.
15 Utilizing AI technologies to enhance e-commerce business operations
323
Sentiment analysis can help automate customer interactions by assessing customer sentiment in real time. It can be used to understand customer’s complaints and can help improve customer service by allowing companies to take action based on the customer’s sentiment. For example, sentiment analysis can suggest personalized responses for customers and allow companies to proactively reach out to customers with offers or assistance. It can also be used to identify issues in customer interaction that can help inform governance and compliance.
15.8.3 Automating customer inquiries, order tracking, and issue resolution Automating customer inquiries, order tracking, and issue resolution is an effective means of streamlining the customer service process and improving customer satisfaction. Automation allows customer service departments to quickly and accurately respond to customer queries and orders, as well as swiftly resolving any issues that may arise. By automating the customer service process, customer service departments are able to respond to customers in a timely manner while also reducing the amount of administrative time associated with customer interactions [29]. Additionally, automated customer service systems are designed to help automate and streamline the customer service process, which can result in a more efficient and efficient customer service experience for customers. Automation also allows customer service departments to quickly resolve any issues that may arise, allowing them to avoid time-consuming and costly processes such as manual order tracking and dispute resolution.
15.8.4 Improving customer satisfaction and reducing support costs through AI-driven solutions AI-driven solutions are being used to improve customer satisfaction and reduce support costs for businesses. This is done by automating customer service tasks and providing intelligent solutions that can better respond to customer inquiries. AI-based solutions use predictive analytics, machine learning, and NLP to provide customers with better and more personalized solutions. AI-driven solutions can be used to provide customers with live chats, automated daily reports, automated responses to customer inquiries, and self-service options for customers. Additionally, these solutions can integrate with existing customer service systems to provide quicker and more reliable support. By utilizing AI-driven solutions, businesses can improve customer satisfaction and reduce support costs.
324
S. Rajeshwari et al.
15.9 Ethical and legal considerations in AI-powered e-commerce Ethical and legal considerations in AI-powered e-commerce are becoming increasingly important as AI technology is adopted by businesses of all types. AI is playing an increasingly important role in the e-commerce industry, driving innovation in product discovery, customer experience, and pricing. With this increased power comes the potential for misuse and misunderstanding of the technology, which is why it is important for e-commerce businesses to take into account the ethical and legal implications of AI-powered decision-making [30]. An ethical consideration is the possibility of algorithmic bias when making decisions, particularly when those decisions could unfairly affect certain groups. This is especially important in e-commerce, as decisions made on pricing and product recommendations could expose customers to price discrimination or reveal personal information or characteristics about customers. Companies must ensure that their AI-powered decisions are fair and not biased against any groups. Legally, companies need to be aware of the implications of their AIpowered decisions. This includes being aware of applicable consumer protection laws, as well as privacy laws applicable to the collection and use of customer data [31]. Additionally, companies must be sure to comply with any applicable industry regulations and accepted industry practices. By addressing these ethical and legal considerations when using AI-powered decision-making in e-commerce, businesses can help ensure that their use of the technology is both ethical and legal.
15.9.1 Ensuring transparency and fairness in AI algorithms and decision-making Ensuring transparency and fairness in AI algorithms and decision-making is about guaranteeing that AI algorithms don’t produce biased results and ensuring that decisionmaking based on these algorithms is consistent and follows ethical principles. In order to achieve this transparency and fairness, it is required to establish audit mechanisms for the algorithm and predictive models. This should be followed up by an evaluation of performance and processes which should include audit logs, source code validity checks, sensitivity analysis, data quality checks, model accuracy checks, impact assessments, and anecdotal analysis. Additionally, it is important to take into consideration variables such as demographic, background, and context in order to reduce bias in decision-making [32]. Creating transparent AI requires three key factors: human-centered design, data governance, and algorithmic fairness. Human-centered design requires companies to consider the stakeholders in the product and their needs in order to create a product that caters to the diverse range of people, based on different backgrounds and opinions. Data governance is important in order to guarantee the quality and integrity of the data
15 Utilizing AI technologies to enhance e-commerce business operations
325
being used, as well as to ensure that sensitive data is protected and collected ethically. Algorithmic fairness is also essential in order to avoid bias or discrimination, such as intentional or unintentional errors and inaccuracies in functional AI tasks. By adhering to these three factors, companies can create more transparent and fair AI solutions.
15.9.2 Privacy and data protection considerations in e-commerce Privacy and data protection considerations in e-commerce refer to the steps taken to ensure that customer data is kept secure and protected from potential abuse or misuse. This includes measures such as encryption, strong authentication, and secure transmission protocols. Additionally, when collecting personal data, businesses should make sure to collect only the data necessary for their operations, and make sure customers are made aware of how their data is used or shared. Furthermore, businesses should have processes in place for overseeing how collected data is used, stored, and secured. Ultimately, businesses should take all reasonable steps to ensure they are compliant with applicable data protection regulation.
15.9.3 Compliance with regulations and guidelines related to AI and e-commerce Compliance with regulations and guidelines related to AI and e-commerce is essential for all organizations that are using or considering using AI and e-commerce technologies. AI technologies are complex and require organizations to be aware of the various legal and regulatory frameworks that may be applicable to their use. Key components include relevant data protection laws, data security and privacy standards, anti-money laundering laws, and consumer protection laws [33]. Furthermore, e-commerce businesses should also take into consideration the restrictions on certain items or activities that are imposed by different jurisdictions. The use of AI and e-commerce requires organizations to develop policies and procedures regarding data, identity, and privacy that are compliant with laws and industry regulations such as GDPR. Organizations must also consider the potential risks of using AI and e-commerce and develop appropriate mitigation strategies to address these issues.
15.9.4 Responsible AI practices and mitigating biases in e-commerce operations Responsible AI practices focus on creating and maintaining ethically and ethically responsible AI solutions to improve business processes and operations. This includes ensuring that AI algorithms are designed and developed with fairness in mind and with
326
S. Rajeshwari et al.
accountability regarding how decisions are made. Responsible AI also entails taking steps to meet privacy standards and ensure data protection for consumers and the business. Mitigating bias in e-commerce operations start with understanding how personal data is collected and used by e-commerce platforms. It’s important to also understand the algorithms used to make recommendations and decisions. Companies should strive to create properly functioning and accountable AI systems that make decisions in a way that takes customer needs and preferences into account [34]. To do this, it is necessary to establish AI models that are regularly checked and maintained to ensure accuracy and consistency with set rules. Companies should also monitor AI models to identify any areas of bias and take corrective action. Finally, companies should consider diversity among their data scientists and other team members to enable a range of unique perspectives to guide the development and implementation of AI models.
15.10 Future trends and implications of AI in e-commerce AI has already established itself as a critical part of the e-commerce experience and is becoming ever more important as online stores become increasingly competitive. AIdriven automation, personalization, and analytics are becoming commonplace and are expected to become even more so in the future. AI technology is helping to power e-commerce stores by providing more intelligent search capabilities, intelligent pricing algorithms, and advanced analytics. AI-enabled automation is making it easier for merchants to handle more orders efficiently and with improved accuracy and at lower costs. Automation technology is used to produce personalized product recommendations for customers, to handle customer service inquiries, and to deliver better product search results [35]. In addition, AI is allowing merchants to better understand customer behaviors, buying patterns, and preferences. This information can then be used to optimize product placements, pricing, discounts, and promotional campaigns. As AI technology continues to evolve and become smarter, more e-commerce stores will likely implement it in order to improve their services. In the future, machine learning and NLP technologies are expected to help merchants better understand their customers and refine their services based on user behavior. Automated bots and virtual customer service agents may become commonplace, providing 24/7 support to customers and helping to reduce customer service costs. AI will also be used to improve supply chain and logistics operations. E-commerce stores have already started using predictive analytics and machine learning technology to anticipate demand and supply chain disruptions, forecast customer demand, and optimize inventory levels. In the future, AI-powered robots may be used to manage warehouses and logistics operations, thereby further improving efficiency and cost reduction. In the coming years, it is
15 Utilizing AI technologies to enhance e-commerce business operations
327
likely that AI technology in e-commerce will become even smarter and more integrated into the user experience. As machine learning and NLP become more pervasive, more platforms and applications will be able to understand customer requests and provide them with personalized and automated assistance. This will result in improved customer experience, greater customer loyalty, and increased sales.
15.10.1 Emerging AI technologies and their potential impact on e-commerce Emerging AI technologies such as machine learning, NLP, and computer vision are transforming the e-commerce industry. Machine learning algorithms allow businesses to personalize customer experiences, optimize consumer behavior predictions, and accurately forecast demand. NLP has enabled the development of chatbots that make it easier for customers to find exactly what they’re looking for quickly when interacting with a website or app. Computer vision is being used to streamline product management and accelerate product recommendations, which means customers can find items more quickly. The potential impact of these technologies on the e-commerce industry is far-reaching, from streamlining sales and marketing strategies to improving customer service. Companies have the potential to increase sales by providing better customer experiences through the use of AI technologies.
15.10.2 AI-driven personalization and hyper-targeted marketing AI-driven personalization and hyper-targeted marketing are two different tools used to enhance marketing campaigns. AI-driven personalization is the use of advanced algorithms to create personalized content and experiences for customers that are tailored to each individual. This involves using AI to automatically analyze customer data, make predictions about what the customer wants, and then provide content that is tailored to the customer’s preferences [36]. Hyper-targeted marketing is the use of customer data and other data sources to target an audience with a specific message or offer. This involves segmenting the audience into groups based on their characteristics or interests and then creating marketing messages tailored to each individual group. By using hyper-targeted marketing, companies can ensure that their message is being delivered to the right people.
328
S. Rajeshwari et al.
15.10.3 Integration of AI with emerging technologies like augmented reality and blockchain Integrating AI with emerging technologies such as augmented reality (AR) and Blockchain is seen as a way to enable the development of artificial general intelligence (AGI). Augmented reality (AR) can enhance AI by providing realistic visualizations that enable agents to more accurately perceive their environment. Meanwhile blockchain can provide secure data transfer and storage, allowing for the secure transfer of information across multiple agents within a distributed network. Together, these new technologies can be leveraged to enable AI to better understand and respond to complex real-world situations while also increasing its efficiency and security.
15.10.4 The future of AI in transforming the e-commerce industry AI is rapidly transforming the e-commerce industry. AI can help e-commerce businesses improve their customer experience by automating customer service, enabling facial recognition, making personalized product recommendations, and other features. It can also be used to automate backend operations such as inventory management, payment processing, and more. AI can analyze customer purchasing patterns to predict future trends and make forecasts about new products, services, and promotional activities. AI can also be used to automate digital marketing activities such as email campaigns and social media promotions. By leveraging the power of AI, e-commerce companies can quickly identify and address customer needs, and create a more personal and enjoyable shopping experience.
References [1] [2] [3] [4] [5]
[6]
Song, X., Yang, S., Huang, Z., & Huang, T. (2019, August). The application of artificial intelligence in electronic commerce. Journal of Physics: Conference Series, 1302(3), 032030. IOP Publishing. Khrais, L. T. (2020). Role of artificial intelligence in shaping consumer demand in E-commerce. Future Internet, 12(12), 226. Xue, M., Xiu, G., Saravanan, V., & Montenegro-Marin, C. E. (2021). Cloud computing with AI for banking and e-commerce applications. The Electronic Library, 39(4), 539–552. Bawack, R. E., Wamba, S. F., Carillo, K. D. A., & Akter, S. (2022). Artificial intelligence in E-Commerce: A bibliometric study and literature review. Electronic Markets, 32(1), 297–338. Zhang, D., Pee, L. G., & Cui, L. (2021). Artificial intelligence in E-commerce fulfillment: A case study of resource orchestration at Alibaba’s Smart Warehouse. International Journal of Information Management, 57, 102304. Yang, X. (2019, June). Satisfaction evaluation and optimization of tourism e-commerce users based on artificial intelligence technology. In 2019 International Conference on Robots & Intelligent System (ICRIS) (pp. 373–375). IEEE.
15 Utilizing AI technologies to enhance e-commerce business operations
[7] [8]
[9] [10] [11] [12]
[13] [14]
[15] [16]
[17]
[18]
[19]
[20]
[21]
[22] [23]
329
George, A. S., & George, A. H. (2023). A review of ChatGPT AI’s impact on several business sectors. Partners Universal International Innovation Journal, 1(1), 9–23. Ikumoro, A. O., & Jawad, M. S. (2019). Intention to use intelligent conversational agents in e-commerce among Malaysian SMEs: an integrated conceptual framework based on tri-theories including unified theory of acceptance, use of technology (UTAUT), and TOE. International Journal of Academic Research in Business and Social Sciences, 9(11), 205–235. Toleuuly, A., Yessengeldin, B., Jumabaeva, S., & Zhanseitov, A. (2019). Contemporary problems and prospects of e-commerce development in modern conditions. Revista Espacios, 40(06), 1–6. Mnyakin, M. (2020). Investigating the impacts of AR, AI, and Website Optimization on Ecommerce Sales Growth. ResearchBerg Review of Science and Technology, 3(1), 116–130. Micu, A., Micu, A. E., Geru, M., Căpățînă, A., & Muntean, M. C. (2021). The Impact of artificial intelligence use on E-Commerce in Romania. Amfiteatru Economic, 23(56), 137–154. Pallathadka, H., Ramirez-Asis, E. H., Loli-Poma, T. P., Kaliyaperumal, K., Ventayen, R. J. M., & Naved, M. (2023). Applications of artificial intelligence in business management, e-commerce and finance. Materials Today: Proceedings, 80, 2610–2613. Li, S. (2020). Structure optimization of e-commerce platform based on artificial intelligence and blockchain technology. Wireless Communications and Mobile Computing, 2020, 1–8. Maslak, O. I., Maslak, M. V., Grishko, N. Y., Hlazunova, O. O., Pererva, P. G., & Yakovenko, Y. Y. (2021, September). Artificial intelligence as a key driver of business operations transformation in the conditions of the digital economy. In 2021 IEEE International Conference on Modern Electrical and Energy Systems (MEES) (pp. 1–5). IEEE. Panigrahi, D., & Karuna, M. (2021). A review on leveraging artificial intelligence to enhance business engagement in ecommerce. Journal Homepage: www.ijrpr.com ISSN, 2582(7421), 2. Moriuchi, E., Landers, V. M., Colton, D., & Hair, N. (2021). Engagement with chatbots versus augmented reality interactive technology in e-commerce. Journal of Strategic Marketing, 29(5), 375–389. Rani, P., Verma, S., Yadav, S. P., Rai, B. K., Naruka, M. S., & Kumar, D. (2022). Simulation of the lightweight blockchain technique based on privacy and security for healthcare data for the cloud system. International Journal of E-Health and Medical Communications (IJEHMC), 13(4), 1–15. Khan, S., Saravanan, V. N. G. C., Lakshmi, T. J., Deb, N., & Othman, N. A. (2022). Privacy protection of healthcare data over social networks using machine learning algorithms. Computational Intelligence and Neuroscience, 2022, 1–9. Yuvaraj, N., & Vivekanandan, P. (2013, February). An efficient SVM based tumor classification with symmetry non–negative matrix factorization using gene expression data. In 2013 International Conference on Information Communication and Embedded Systems (Icices) (pp. 761–768). New York: IEEE. Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). A novel architecture of intelligent decision model for efficient resource allocation in 5G broadband communication networks. ICTACT Journal on Soft Computing, 13(3), 2986–2994. Yadav, S. P., Zaidi, S., Mishra, A., & Yadav, V. (2022). Survey on machine learning in speech emotion recognition and vision systems using a recurrent neural network (RNN). Archives of Computational Methods in Engineering, 29(3), 1753–1770. Saravanan, V., & Mohan Raj, V. (2016). Maximizing QoS by cooperative vertical and horizontal handoff for tightly coupled WiMAX/WLAN overlay networks. Cluster Computing, 19(3), 1619–1633. Yuvaraj, N., Chang, V., Gobinathan, B., Pinagapani, A., Kannan, S., Dhiman, G., & Rajan, A. R. (2021). Automatic detection of cyberbullying using multi-feature based artificial intelligence with deep decision tree classification. Computers and Electrical Engineering, 92, 107186.
330
S. Rajeshwari et al.
[24] Ramesh, G., Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). Prediction of energy production level in large PV plants through AUTO-Encoder Based Neural-Network (AUTO–NN) with restricted Boltzmann feature extraction. Future Internet, 15(2), 46. [25] Yadav, S. P., Mahato, D. P., & Linh, N. T. D. (Eds.). (2020). Distributed artificial intelligence: A modern approach. Boca Raton: CRC Press. [26] Munagala, N. V. L., Saravanan, V., Almukhtar, F. H., Jhamat, N., Kafi, N., & Khan, S. (2022). Supervised approach to identify autism spectrum neurological disorder via label distribution learning. Computational Intelligence and Neuroscience, 2022, 1–11. [27] Gobinathan, B., Mukunthan, M. A., Surendran, S., Somasundaram, K., Moeed, S. A., Niranjan, P., . . . & Sundramurthy, V. P. (2021). A novel method to solve real time security issues in software industry using advanced cryptographic techniques. Scientific Programming, 2021, 1–9. [28] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). L-RUBI: An efficient load-based resource utilization algorithm for bi-partite scatternet in wireless personal area networks. International Journal of Communication Systems, 36(4), 1–17. [29] Yadav, S. P., & Yadav, S. (2020). Fusion of medical images in wavelet domain: A hybrid implementation. Computer Modeling in Engineering and Sciences, 122(1), 303–321. [30] Saravanan, V., & Sumathi, A. (2012, December). Dynamic handoff decision based on current traffic level and neighbor information in wireless data networks. In 2012 Fourth International Conference on Advanced Computing (ICoAC) (pp. 1–5). IEEE. [31] Raja, S., Logeshwaran, J., Venkatasubramanian, S., Jayalakshmi, M., Rajeswari, N., Olaiya, N. G., & Mammo, W. D. (2022). OCHSA: Designing energy-efficient lifetime-aware leisure degree adaptive routing protocol with optimal cluster head selection for 5G communication network disaster management. Scientific Programming, 2022, 1–12. [32] Sivaram, M., Yuvaraj, D., Mohammed, A. S., Manikandan, V., Porkodi, V., & Yuvaraj, N. (2019). Improved enhanced DBTMA with contention-aware admission control to improve the network performance in MANETS. CMC-Computers Materials and Continua, 60(2), 435–454. [33] Sandu, N., & Gide, E. (2019, September). Adoption of AI–Chatbots to enhance student learning experience in higher education in India. In 2019 18th International Conference on Information Technology Based Higher Education and Training (ITHET) (pp. 1–5). IEEE. [34] Ping, N. L. (2019, December). Constructs for artificial intelligence customer service in E-commerce. In 2019 6th International Conference on Research and Innovation in Information Systems (ICRIIS) (pp. 1–6). IEEE. [35] Agnihotri, R. (2021). From sales force automation to digital transformation: How social media, social CRM, and artificial intelligence technologies are influencing the sales process. A Research Agenda for Sales, 2021, 21–47. [36] Zoroja, J., Klopotan, I., & Ana-Marija, S. (2020). Quality of e-commerce practices in European enterprises: Cluster analysis approach. Interdisciplinary Description of Complex Systems: INDECS, 18(2–B), 312–326.
Vijaya Gunturu, Charanjeet Singh, Nikhil S. Patankar, S. Praveena
16 Exploring the potential of artificial intelligence in wireless sensor networks Abstract: In various monitoring, tracking, and control applications, wireless sensor networks (WSNs) are gaining popularity. To fully utilize the promise of WSNs in various applications, two significant challenges – power efficiency and scalability – must be overcome. Recent developments in artificial intelligence (AI) methods, such as deep learning, machine learning (ML), and reinforcement learning, present fresh perspectives and development opportunities for WSNs to make wise decisions and effectively use resources. For example, AI-based techniques can enable WSNs to learn the underlying patterns and trends in the data streams exchanged among different sensor nodes, and choose optimal parameters for efficient data collection and analysis. Keywords: Wireless sensor networks, monitoring, tracking, control, power
16.1 Introduction The utilization of AI techniques can lead to smart, adaptive strategies in WSNs which are capable of self-organization and to quickly react to external dynamic environments. Therefore, AI-based methods could greatly improve the power efficiency, scalability, flexibility, and robustness of WSNs [1]. This paper provides an overview of recent advancements in AI techniques and explores potential applications in WSNs. Furthermore, potential research opportunities are discussed, and open challenges are identified in order to further leverage the potential of AI for WSNs.
16.1.1 Background and motivation A subset of distributed computing known as Wireless Sensor Networks (WSNs) is composed of several sensor nodes that are placed across a physical space to gather sensory data, act to monitor the environment, and provide essential communication links between nodes [2]. Artificial intelligence (AI) is an increasingly important technology in WSNs due to its ability to allow efficient sensing and data processing, in a single device. AI can enable WSNs to address difficult tasks such as data analysis, problem solving, and pattern recognition which have traditionally been outside the scope of conventional networks. Applications for AI-based WSNs include surveillance, industrial automation, monitoring of the environment, and monitoring of human health. AI can also contribute to lower energy costs and longer network lifespans by https://doi.org/10.1515/9783111323749-016
332
Vijaya Gunturu et al.
enabling efficient resource consumption [3]. Furthermore, AI can ease the stress on a network’s communication by enabling automated decision-making.
16.1.2 Objectives and scope The objective of AI in WSNs is to develop novel algorithms, techniques, and tools to enable WSNs to autonomously learn, diagnose, and respond to their environment. This includes leveraging AI to enable self-healing of the network, self-configuration, energy optimization, and network security [4]. The scope of AI in WSNs is broad and includes progress in both low-level hardware-oriented algorithms, such as distributed data fusion, and high-level network and application-oriented algorithms, such as deep learning techniques. AI techniques that will be useful for WSNs include clustering, decision tree–based feature selection, and automated reasoning. Research into AI for WSNs will also examine solutions for recognizing temporal behavior patterns (such as traffic inhibition characteristics) and developing new algorithms that can exploit the spatiotemporal correlation of sensor data [5]. The AI-based WSNs should be able to learn from their surroundings over time and adapt, which will improve security, detection performance, and energy management. Research work in this sphere could focus on modeling the relationship among environmental and network factors and tuning system parameters for mobility and power optimization [6].
16.2 Wireless sensor networks: an overview WSNs are collections of independent sensor nodes that communicate with one another via wireless links in order to gather and process data about a specific physical event. WSNs can be installed in locations where installing wired networks is difficult or costly, such as in remote locations, inaccessible or dangerous areas, or in areas without access to power or communication lines. Sensing environmental variables like temperature, humidity, light level, air quality, motion, and noise are WSN’s most popular uses. Additionally, WSNs can be utilized for motion detection, medical monitoring, surveillance, and other purposes such as asset location tracking and monitoring [7]. The sensor nodes, gateway nodes, and routers are the key elements of a WSN. Sensor nodes are little sensors that perceive their surroundings and gather data. The gateway node, a device that transmits the collected data to the remote monitoring agent, receives it from the sensor nodes. In order to network the nodes and make sure they can communicate with one another, routers are nodes [8].
16 Exploring the potential of artificial intelligence in wireless sensor networks
333
The WSNs are networks of interconnected nodes that allow for data collection and processing with minimal dependence on wired connections. With their growing popularity and affordability, they are becoming more common for applications from industrial sensing to IoT applications [9].
16.2.1 Definition and characteristics A WSN is a group of spatially dispersed autonomous sensors that work together to cooperatively transmit data to a central location while simultaneously monitoring physical or environmental factors like temperature, sound, vibration, pressure, mobility, or pollution. Due to the bidirectional nature of more recent networks, control over sensor activity is also possible [10]. The characteristics of WSN include: – Ad hoc network: WSN typically consists of randomly distributed nodes that are self-organizing and self-managing. – Low power consumption: Nodes are powered by batteries and designed to operate for a long time without maintenance. – Small size and lightweight: The nodes are typically small for easy implementation and short-range communication. – Low cost: WSN nodes are usually low cost compared to other networks. – Fault tolerance and reliability: The network should be able to operate even when sensors fail or damage. – Multi-hop communications: Information is passed through multiple nodes before it reaches the base station. – Security: The information conveyed over WSNs should be able to communicate in a secure environment.
16.2.2 Applications and challenges Deploying inexpensive, low-power wireless sensor nodes throughout a geographical area forms the basis of the emerging network technology referred to as wireless sensor network (WSN). WSN is utilized in a variety of applications, including industrial, environmental, military, and medical ones, to collect data and monitor operations [11]. Applications of WSN: – Scientific exploration: WSN can be used in such fields as geology, biology, and meteorology to measure, monitor, and take discrete data in remote areas. This is beneficial for scientific research and exploration. – Industrial applications: WSN can be used in the industrial sector for tracking and identifying critical assets as well as monitoring environment parameters such as temperature and humidity.
334
– –
–
Vijaya Gunturu et al.
Infrastructure monitoring: WSN can be employed in the monitoring of infrastructure such as bridges, roads, and smart cities. Home automation and security: WSN can be used in home automation and security to control systems such as lighting, heating, and entertainment systems, as well as detect intruders. Healthcare: By noninvasively monitoring patients’ vital signs including heart rate and blood pressure, WSN can be tremendously helpful in the healthcare industry.
Challenges in WSN: – Due to its reliance on radio-based communication that can be intercepted, security continues to be one of the largest difficulties facing WSN. To keep the network and its data safe from hostile actions, security needs to be addressed. – Network management and maintenance: Given the abundance of sensor nodes inside a WSN and the limited resources available, network management and maintenance can frequently be a challenging undertaking. – Power management: Because WSN sensor nodes have a certain amount of power available to them, careful power management is necessary to guarantee the network functions properly. – Data collection: Since real-time data is lacking in WSNs, gathering data might be challenging. – Scalability: The WSN needs to be scalable in order to quickly fulfill the need for increased data and resources.
16.2.3 Role of artificial intelligence in WSNs The role of AI in WSNs is to provide autonomous computing, decision-making, and data analysis capabilities to WSNs. AI enables WSNs to make decisions autonomously, based on obtained data and on-board computing power [12]. This can help WSNs become more reliable, efficient, and accurate, as they have the ability to process data faster and learn from their experience. By spotting irregularities in communication patterns, weak or broken links, aberrant energy usage, and traffic levels, WSNs can use AI algorithms to efficiently lower costs, save energy, and extend the lifetime of the networks [13]. AI-assisted WSNs also make it possible to reduce complexity in wireless communications by compressing data and performing more intelligent routing. AI is also being applied in WSNs to enable tasks to be completed quickly, such as monitoring environments, detecting intrusions, and recognizing patterns. AI has the potential to improve the performance of WSNs by providing them with the capability to detect, identify, and act autonomously on the basis of data collected from the environment [14].
16 Exploring the potential of artificial intelligence in wireless sensor networks
335
16.3 Artificial intelligence techniques for WSNs WSNs can function better when AI techniques are applied. Numerous applications, including data fusion, effective communication, energy optimization, fault detection, and location, can benefit from the usage of AI approaches. A more accurate and thorough description of the data can be obtained by combining heterogeneous data from several sources via data fusion. A higher-level comprehension of the data from virtually unconnected sources can be produced using AI techniques like fuzzy logic and Bayesian networks [15]. For WSN networks, effective communication protocols are essential since they speed up throughput and cut down on data transmission time. Effective routing pathways between nodes can be built using AI-based techniques. The communication protocols of the network can be improved for performance using AI approaches like genetic algorithms and ant colony optimization. WSNs must take energy optimization into account [16]. Finding the best node routing policies and energy levels to reduce energy usage can be done using AI-based approaches like reinforcement learning. An essential function of WSN networks is fault detection, which can be accomplished using AI techniques like pattern recognition and machine learning algorithms. It is also possible to estimate the location of nodes within the network using localization algorithms that employ AI-based methods. Nodes in WSNs can be located more precisely using methods like particle-swarm optimization and Monte Carlo simulation [17].
16.3.1 Machine learning algorithms Algorithms that employ statistical methods to help machines learn from data without explicit programming are known as machine learning algorithms. They are employed to base predictions or choices on information. These techniques include support vector machines, decision trees, random forests, logistic regression, and linear regression as examples. Numerous industries, including finance, healthcare, marketing, and computer vision, use these algorithms. They can also be utilized for tasks like fraud detection, stock market analysis, facial recognition, and text classification [18].
16.3.1.1 Supervised learning One form of machine learning algorithm that extracts knowledge from labeled data is supervised learning. The objective of supervised learning is to create a model that can predict outcomes based on fresh data. Predictive analysis and categorization issues are typically addressed using it. In supervised learning, the algorithm is trained on labeled data in order to produce a function that can be used to predict outcomes from fresh data [19].
336
Vijaya Gunturu et al.
16.3.1.2 Unsupervised learning Unsupervised learning is a kind of machine learning technique that uses data analysis without the need for prior labeling of the data. Without receiving any outside direction, it tries to find data patterns and features that are connected to or resemblant to one another. As a result, it’s frequently used to find clusters in huge datasets and to uncover information that might not have been discovered otherwise. Algorithms for unsupervised learning can be utilized for a variety of applications, including anomaly detection, dimensionality reduction, and grouping [20].
16.3.1.3 Reinforcement learning By receiving feedback in the form of incentives or penalties, an agent can learn behaviors from its environment using a machine learning technique called reinforcement learning. It is based on the concept of learning from mistakes, and it provides the agent with the opportunity to do so by exploring and exploiting new opportunities. The agent will learn behaviors to maximize its expected reward for a given task by making observations and taking actions as it learns how to interact with its environment. Reinforcement learning is used in many applications, such as robotics, selfdriving cars, finance, and control systems [21].
16.3.2 Deep learning techniques A type of AI called deep learning uses numerous layers of neurons to process input and produce patterns. Deep learning algorithms can forecast the data and discover brand-new patterns that were previously undetected by using these patterns. Healthcare, finance, marketing, and other industries can all benefit from the use of this technology. Deep learning’s capacity to find latent properties in data is its key benefit. It can identify data patterns that are too complicated for humans to notice [22]. Because it can swiftly and reliably process massive amounts of data, this technology is growing in popularity. Due to this, many organizations are now utilizing deep learning to make decisions and boost their production and efficiency.
16.3.3 Swarm intelligence algorithms Swarm intelligence algorithms are computer algorithms designed to mimic the behavior of swarms or groups of organisms. These algorithms are used mainly in robotics and AI, and they are inspired by how real-world swarms usually act. They simulate collective behavior and the ability of swarms to learn from their environment. Swarm in-
16 Exploring the potential of artificial intelligence in wireless sensor networks
337
telligence algorithms are used to solve complex problems or make decisions where no single entity has control [23]. These algorithms rely on agents working cooperatively and exchanging information so they can learn from each other.
16.4 AI-enabled applications in wireless sensor networks AI-enabled applications in WSNs are systems that leverage AI algorithms and techniques in order to enable WSNs to sense, monitor, and communicate in real time. AI-enabled WSNs have applications in various industries, such as environmental and health monitoring, industrial automation, smart homes, and security surveillance. They enable intelligent sensing and actuation by utilizing AI methods like supervised machine learning and deep learning to make WSNs smarter. AI-enabled WSNs can detect events at an earlier stage, analyze and identify patterns in real time, and adjust their behavior accordingly. They are reliable, efficient, and energy-saving, and offer longer battery life. AI-enabled WSNs have the potential to revolutionize many industry sectors by providing smart sensing, monitoring, and actuation capabilities [24]. Such capabilities could be used to detect, monitor, and forecast environmental factors including air and soil quality, water levels, and temperatures, as well as to detect and monitor human activities and to deliver alarms and diagnostics in hazardous or dangerous locations. AI-enabled WSNs can also be used to monitor network traffic, detect and identify intrusions, and optimize network performance.
16.4.1 Energy management and optimization Energy optimization and management in AI-based technology known as WSNs enables the network’s energy usage to be optimized. It operates by tracking how much energy is used by the WSN nodes, then calculating each node’s energy using AI algorithms. This allows the WSN to efficiently route data, while consuming the least amount of energy. The AI algorithms also allow for fault tolerance, and accurately predict the energy requirements for future operations, so that the energy consumption can be optimized. This reduces the overall energy costs of the WSN [25]. AI-based WSNs can also be connected to the smart grid for further energy optimization.
16.4.2 Data fusion and information processing Data fusion and information processing in AI-based WSN are both key elements of managing and analyzing data that is collected by WSNs. The process of merging data
338
Vijaya Gunturu et al.
from several sources to produce a comprehensive picture of the environment is known as data fusion. The process of turning data from an external source into information that intelligent algorithms and robots may utilize to make decisions is known as information processing [26]. Because it enables the gathering of extra information from many sources, data fusion is crucial for WSNs. AI may improve its understanding of an environment by merging input from several sources, and the resulting data can be used with machine learning techniques to make better decisions. [27]. Information processing is important for AI-based WSNs because it allows for the accumulation of large amounts of data. This data can then be used for more accurate visions and predictions. It involves taking data from external sources, such as weather sensors, and turning it into useful data that can be used by AI.
16.4.3 Localization and tracking The process of locating sensor nodes (or devices) physically in WSNs involves exploiting the infrastructure that is already in place, such as beacons, GPS, accelerometers, or other wireless communication systems. Numerous WSN applications, such as surveillance, asset tracking, and environmental monitoring, depend on localization. As part of a tracking and localization loop, tracking allows the network to keep track of how the device is moving around the environment and to update the device’s location [28]. Ranging methods (triangulation, trilateration, etc.) and positioning methods are two groups of localization algorithms. Both approaches rely on the ability to deterministically measure distances or angles to beacons (e.g., Wi-Fi APs and stationary nodes) or other peers in the network. Ranging methods, also referred to as viewbased approaches, are typically based on measuring received signal strength or time of arrival. Positioning methods are based on GPS or UWB which provide highly precise coordinates for sensor nodes [29]. Tracking in WSNs can be divided into two categories as well: motion tracking and tracking with respect to absolute location. Motion tracking involves tracking node movements within a given area (e.g., indoor navigation) and is usually combined with position-based localization algorithms. Tracking with respect to absolute location can be used to determine the position of sensor nodes in a larger scale environment (e.g., indoor or outdoor) and requires ranging or positioning methods [30]. Localization and tracking technologies form the backbone of many WSN applications such as robotics, environment monitoring, or asset tracking. By providing accurate location and positioning information, they enable applications to react and adapt to the environment surrounding them.
16 Exploring the potential of artificial intelligence in wireless sensor networks
339
16.4.4 Fault detection and diagnosis A method for detecting and diagnosing flaws in WSNs is known as fault detection and diagnosis in AI in WSNs. This method uses machine learning and AI algorithms to find faults. In order to detect errors and determine their root cause, the AI algorithms used for fault detection examine network data and extract information from it. The AI algorithms then provide a diagnosis to the user about the fault’s origin and the best course of action to remedy it [31]. This process is useful for autonomous systems, such as for example unmanned aerial vehicles (UAVs), as it allows them to selfdiagnose and self-heal in case of a failure. AI-based fault detection and diagnosis can be used in a variety of settings, including medical, industrial, and military systems.
16.5 AI-based security and privacy in WSNs AI-based security and privacy in WSNs refers to the application of AI techniques to improve security and privacy in WSNs. These methods can be applied to spot suspicious activity, spot and stop intrusion, stop unauthorized access, shield the sensor data from attacks, and prevent security breaches. AI-based security and privacy measures in WSNs also provide real-time intelligence for monitoring networks and detecting potential attacks [32]. AI-based security and privacy measures have the ability to adapt and respond to changing conditions in the network and help the network secure itself against evolving threats. Examples of AI-based security and privacy measures in WSNs include machine learning-based anomaly detection, intrusion detection and prevention, and authentication techniques.
16.5.1 Intrusion detection and prevention Intrusion detection and prevention is a technology designed to identify and prevent malicious activity in a computer network. It is a combination of software and hardware products that detect, monitor, and alert system administrators to suspicious activity. Intrusion detection and prevention systems (IDPS) are used to detect unauthorized access to networked systems and applications. They can also provide protection against malware, unauthorized access to system resources, data theft, and denial of service attacks [33]. IDPS can be used to identify malicious activity by monitoring network traffic for signs of malicious behavior and alerting administrators to potential threats. Additionally, an IDPS can be configured to take action to block or deny access to malicious traffic when it is detected.
340
Vijaya Gunturu et al.
16.5.2 Secure data transmission Secure data transmission is an approach for securely exchanging data over a network. In secure data transmission, endpoints encrypt communications with the use of asymptotically secure encryption algorithms such as TLS, IPSec, and SSH. The encryption and authentication can be done for the entire data packet or only the meta-data. Additional security protocols such as PKI (public key infrastructure) and authentication methods like Kerberos are used in order to ensure that data is transferred without being tampered by malicious actors. It also ensures that the data is transferred only to authorized users and that the communication is only between the authorized parties. Secure data transmission is essential for sensitive data such as financial data, medical records, and other confidential information [34].
16.5.3 Privacy preservation In the process of gathering, storing, processing, using, and disseminating personal data, privacy preservation safeguards the privacy of people, businesses, or other entities. Customers’ information and personally identifiable information like a social security number, license plate number, email address, or bank account information might all be included in this data. Data protection and preventing unauthorized access are the two main aims of privacy preservation. Additionally, it’s crucial to guarantee that the information is only utilized for the purposes for which it was acquired [35]. The protection of privacy can be accomplished by the use of access control technologies, secure networks, and encryption methods. Organizations must employ appropriate governance procedures and policies to maintain data privacy, even though these measures offer some degree of security.
16.6 Challenges and future directions The challenges of AI techniques for WSNs include reducing energy consumption, dynamic reconfiguration of nodes, improving data reliability, and developing robust security mechanisms such as intrusion detection. The future directions of AI techniques for WSNs include using deep learning, fuzzy logic, and reinforcement learning for specific application areas like health monitoring and environmental monitoring, while also using machine learning and AI-driven algorithms to enable more reliable and efficient communications in WSNs [36]. Additionally, AI can be used for dynamic reconfiguration of WSNs, and predictive maintenance, allowing for the identification of potential problems before they occur.
16 Exploring the potential of artificial intelligence in wireless sensor networks
341
References [1] [2] [3]
[4]
[5]
[6] [7]
[8]
[9]
[10]
[11] [12]
[13]
[14]
[15]
[16]
Ghazal, T. M. (2022). Data fusion-based machine learning architecture for intrusion detection. Computers, Materials and Continua, 70(2), 3399–3413. Howard, J. (2019). Artificial intelligence: Implications for the future of work. American Journal of Industrial Medicine, 62(11), 917–926. Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). L-RUBI: An efficient load-based resource utilization algorithm for bi-partite scatternet in wireless personal area networks. International Journal of Communication Systems, 36(4), 1–17. Lăzăroiu, G., Kliestik, T., & Novak, A. (2021). Internet of things smart devices, industrial artificial intelligence, and real-time sensor networks in sustainable cyber-physical production systems. Journal of Self-Governance and Management Economics, 9(1), 20–30. Mekonnen, Y., Namuduri, S., Burton, L., Sarwat, A., & Bhansali, S. (2019). Machine learning techniques in wireless sensor network based precision agriculture. Journal of the Electrochemical Society, 167(3), 037522. Ali, S. S., & Choi, B. J. (2020). State-of-the-art artificial intelligence techniques for distributed smart grids: A review. Electronics, 9(6), 1030. Sawhney, R., Malik, A., Sharma, S., & Narayan, V. (2023). A comparative assessment of artificial intelligence models used for early prediction and evaluation of chronic kidney disease. Decision Analytics Journal, 6, 100169. Adeel, A., Gogate, M., Farooq, S., Ieracitano, C., Dashtipour, K., Larijani, H., & Hussain, A. (2019). A survey on the role of wireless sensor networks and IoT in disaster management. Geological Disaster Monitoring Based on Sensor Networks, 10, 57–66. Nica, E., & Stehel, V. (2021). Internet of things sensing networks, artificial intelligence-based decision-making algorithms, and real-time process monitoring in sustainable industry 4.0. Journal of Self-Governance and Management Economics, 9(3), 35–47. Thangaramya, K., Kulothungan, K., Logambigai, R., Selvi, M., Ganapathy, S., & Kannan, A. (2019). Energy aware cluster and neuro-fuzzy based routing algorithm for wireless sensor networks in IoT. Computer Networks, 151, 211–223. Arroyo, P., Herrero, J. L., Suárez, J. I., & Lozano, J. (2019). Wireless sensor network combined with cloud computing for air quality monitoring. Sensors, 19(3), 691. Yuvaraj, N., & Vivekanandan, P. (2013, February). An efficient SVM based tumor classification with symmetry non-negative matrix factorization using gene expression data. In 2013 International Conference on Information Communication and Embedded Systems (Icices) (pp. 761–768). IEEE. Yadav, S. P., & Yadav, S. (2019). Mathematical implementation of fusion of medical images in continuous wavelet domain. Journal of Advanced Research in Dynamical and Control System, 10(10), 45–54. Yuvaraj, N., Chang, V., Gobinathan, B., Pinagapani, A., Kannan, S., Dhiman, G., & Rajan, A. R. (2021). Automatic detection of cyberbullying using multi-feature based artificial intelligence with deep decision tree classification. Computers and Electrical Engineering, 92, 107186. Vaniprabha, A., Logeshwaran, J., Kiruthiga, T., & Shah, K. B. (2022). Examination of the effects of long-term COVID-19 impacts on patients with neurological disabilities using a neuro machine learning model. BOHR International Journal of Neurology and Neuroscience, 1(1), 21–28. Saravanan, V., & Sumathi, A. (2012). Handoff mobiles with low latency in heterogeneous networks for seamless mobility: A survey and future directions. European Journal of Scientific Research, 81(3), 417–424.
342
[17]
[18] [19] [20]
[21]
[22]
[23]
[24]
[25]
[26]
[27]
[28]
[29] [30] [31]
[32]
[33]
Vijaya Gunturu et al.
Pujahari, R. M., Yadav, S. P., & Khan, R. (2022). Intelligent farming system through weather forecast support and crop production. In: Mohammad Ayoub Khan, Rijwan Khan and Mohammad Aslam Ansari (Eds.), Application of Machine Learning in Agriculture (pp. 113–130). United States: Academic Press. Mojjada, R. K., Yadav, A., Prabhu, A. V., & Natarajan, Y. (2020). Machine learning models for covid-19 future forecasting. Materials Today: Proceedings, 2020, 1–5. Yadav, S. P., & Yadav, S. (2018). Fusion of medical images in wavelet domain: A discrete mathematical model. Ingeniería Solidaria, 14(25), 1–11. Tripathy, R., Nayak, R. K., Saravanan, V., Mishra, D., Parasa, G., Das, K., & Das, P. (2021). Spectral clustering based fuzzy C-means algorithm for prediction of membrane cholesterol from ATP-binding cassette transporters. In Intelligent and Cloud Computing: Proceedings of ICICC 2019 (Vol. 2, pp. 439–448). Springer Singapore. Rani, P., Verma, S., Yadav, S. P., Rai, B. K., Naruka, M. S., & Kumar, D. (2022). Simulation of the lightweight blockchain technique based on privacy and security for healthcare data for the cloud system. International Journal of E-Health and Medical Communications (IJEHMC), 13(4), 1–15. Singh, J., Deepika, J., Sathyendra Bhat, J., Kumararaja, V., Vikram, R., Jegathesh Amalraj, J., . . . & Sakthivel, S. (2022). Energy-efficient clustering and routing algorithm using hybrid fuzzy with grey wolf optimization in wireless sensor networks. Security and Communication Networks, 2022, 1–13. Gopi, B., Logeshwaran, J., & Kiruthiga, T. (2022). An innovation in the development of a mobile radio model for a dual-band transceiver in wireless cellular communication. BOHR International Journal of Computational Intelligence and Communication Network, 1(1), 20–25. Saravanan, V., & Sumathi, A. (2012, December). Dynamic handoff decision based on current traffic level and neighbor information in wireless data networks. In 2012 Fourth International Conference on Advanced Computing (ICoAC) (pp. 1–5). IEEE. Gobinathan, B., Mukunthan, M. A., Surendran, S., Somasundaram, K., Moeed, S. A., Niranjan, P., . . . & Sundramurthy, V. P. (2021). A novel method to solve real time security issues in software industry using advanced cryptographic techniques. Scientific Programming, 2021, 1–9. Kakani, V., Nguyen, V. H., Kumar, B. P., Kim, H., & Pasupuleti, V. R. (2020). A critical review on computer vision and artificial intelligence in food industry. Journal of Agriculture and Food Research, 2, 100033. Wang, C. X., Di Renzo, M., Stanczak, S., Wang, S., & Larsson, E. G. (2020). Artificial intelligence enabled wireless networking for 5G and beyond: Recent advances and future challenges. IEEE Wireless Communications, 27(1), 16–23. Adhikari, N., Logeshwaran, J., & Kiruthiga, T. (2022). The artificially intelligent switching framework for terminal access provides smart routing in modern computer networks. BOHR International Journal of Smart Computing and Information Technology, 3(1), 45–50. Savaglio, C., Pace, P., Aloi, G., Liotta, A., & Fortino, G. (2019). Lightweight reinforcement learning for energy efficient communications in wireless sensor networks. IEEE Access, 7, 29355–29364. Otoum, S., Kantarci, B., & Mouftah, H. T. (2019). On the feasibility of deep learning in sensor network intrusion detection. IEEE Networking Letters, 1(2), 68–71. Gruson, D., Helleputte, T., Rousseau, P., & Gruson, D. (2019). Data science, artificial intelligence, and machine learning: Opportunities for laboratory medicine and the value of positive regulation. Clinical Biochemistry, 69, 1–7. Lussier, F., Thibault, V., Charron, B., Wallace, G. Q., & Masson, J. F. (2020). Deep learning and artificial intelligence methods for Raman and surface-enhanced Raman scattering. TRAC: Trends in Analytical Chemistry, 124, 115796. Nallaperuma, D., Nawaratne, R., Bandaragoda, T., Adikari, A., Nguyen, S., Kempitiya, T., . . . & Pothuhera, D. (2019). Online incremental machine learning platform for big data-driven smart traffic management. IEEE Transactions on Intelligent Transportation Systems, 20(12), 4679–4690.
16 Exploring the potential of artificial intelligence in wireless sensor networks
343
[34] Ghazal, T. M., Hasan, M. K., Alshurideh, M. T., Alzoubi, H. M., Ahmad, M., Akbar, S. S., . . . & Akour, I. A. (2021). IoT for smart cities: Machine learning approaches in smart healthcare – a review. Future Internet, 13(8), 218. [35] Pandian, A. P. (2019). Artificial intelligence application in smart warehousing environment for automated logistics. Journal of Artificial Intelligence, 1(02), 63–72. [36] Kumar, D. P., Amgoth, T., & Annavarapu, C. S. R. (2019). Machine learning algorithms for wireless sensor networks: A survey. Information Fusion, 49, 1–25.
M. V. Jagannatha Reddy, J. Somasekar, Kaushalya Thopate, Sanjeev Kumar Angadi
17 Exploring artificial intelligence techniques for enhanced sentiment analysis through data mining Abstract: Sentiment analysis is a subset of data mining that is used to locate, collect, and examine people’s opinions. It makes it possible for data mining tools to comprehend peoples’ feelings and ideas more fully. Natural language processing is one of the most used and established methods used for sentiment analysis tasks. This technique is mainly based on rules, principles, and algorithms which aid in understanding the text making it easier to segment and extract meaningful data. However, this technique is becoming obsolete due to the sheer volume and complexity of data present these days. Keywords: Sentiment analysis, data mining, identify, extract, analyze, understanding
17.1 Introduction A new generation of smart algorithms such as artificial intelligence (AI) have been developed for the improved sentiment analysis through data mining. AI techniques such as deep learning networks, logic programming, and machine learning techniques are proving to be two times more accurate than traditional sentiment analysis. These AI techniques provide greater accuracy, flexibility, and efficiency in sentiment analysis tasks [1]. They understand not only the sentiment of the data but also the context in which the data is used. This helps in making the sentiment analysis more precise as it can also detect the intensity of sentiment expressed in the data. Therefore, AI techniques can be used as an enhanced approach for sentiment analysis through data mining.
17.1.1 Scope and objectives The scope and objectives of the AI techniques for enhanced sentiment analysis through data mining are to utilize intelligent technologies to automate and improve the accuracy of sentiment analysis. Natural language processing (NLP) and supervised learning are examples of AI approaches [2]. The AI-driven systems will be able to better identify the latent associations within the data and, based on this, ascertain the sentiment of the text by using data mining techniques like clustering, association rule mining, or sentiment https://doi.org/10.1515/9783111323749-017
346
M. V. Jagannatha Reddy et al.
analysis. This might potentially be applied to supervised learning models, which train models using labeled data. Moreover, the AI-driven system could also be used to identify new themes that have not yet been identified by traditional sentiment analysis tools. This method aims to offer sentiment analysis results that are more accurate than those from conventional techniques [3]. The AI-driven approach should give a better understanding of the text and the sentiment behind it, making the sentiment results more accurate and specific. This is beneficial for organizations looking to gain more insight into their users’ sentiments about their product or services.
17.2 Overview of sentiment analysis Sentiment analysis is a sort of NLP that locates and extracts subjective data about people’s thoughts, feelings, and emotions. It is used to analyze customer feedback and sentiments to gain insight into customer perceptions and thoughts. The results of sentiment analysis are used to inform decision-making processes and help to develop better customer experiences. It can be used to monitor and analyze emotions in conversations on social media, review websites, customer feedback forms, etc. Sentiment analysis is often referred to as emotion AI, sentiment mining, and opinion mining.
17.2.1 Definition of sentiment analysis Sentiment analysis is a branch of NLP, commonly referred to as opinion mining. It is used to understand the opinion or attitude of a speaker regarding a particular topic. Through sentiment analysis, businesses are able to monitor customer feedback, respond promptly, and explore areas of improvement in products and services [4]. Also, it helps providers gain actionable insights from customer conversations. Sentiment analysis is used to detect customer emotions using algorithms and NLP technology. It enables organizations to track consumer sentiments expressed across channels, quickly identify potential areas of improvement, and develop targeted strategies to retain and win back customers.
17.2.2 Importance of sentiment analysis The technique of examining a customer’s perception of a good, service, or brand is known as sentiment analysis. It has become an increasingly important tool for companies as it helps them to better understand customer opinions and behavior. It allows for better customer experience, improved customer service, greater customer satisfaction, and more targeted marketing campaigns [5–6]. By quantifying the sentiment of
17 Exploring artificial intelligence techniques for enhanced sentiment analysis
347
customer feedback, companies can identify potential issues that they may not have previously been aware of and address them before they become serious problems. Additionally, sentiment analysis can be used to develop a better understanding of customer preferences and develop more effective products and services. Ultimately, sentiment analysis helps companies to create a more positive customer experience.
17.2.3 Challenges in sentiment analysis The study of sentiment conveyed in text is the focus of the field of NLP known as sentiment analysis. The difficulty of the issue stems from the intricacy of language. Challenges in sentiment analysis include: – Interpretation ambiguities: Natural language contains ambiguity. This can lead to difficulty in interpreting the sentiment of a particular phrase. For example, the phrase “I’m not bad” may be interpreted as having a positive sentiment or a negative sentiment, depending on the context. – Context dependency: The sentiment of a phrase may depend on the context. For example, the phrase “It’s going to be a great day” may be interpreted as having a positive sentiment if it is uttered in the morning before the individual has encountered any difficulties, but may be interpreted as a negative sentiment if it is uttered late in the evening after the individual has experienced setbacks. – Temporal dependencies: The sentiment of a phrase may change over time. For example, a phrase uttered during a historical moment or event can carry a different sentiment than a phrase uttered after the event, when the current sentiment surrounding it may have changed. – Lexical ambiguities: The sentiment of a phrase may depend on the word used. For instance, depending on the situation, the word “luck” can be used either positively or negatively. – Grammatical ambiguities: The sentiment of a phrase may depend on the grammar used. For example, the phrase “it isn’t bad” may be interpreted as having a positive sentiment, whereas the phrase “it isn’t good” may be interpreted as having a negative sentiment.
17.2.4 Traditional approaches to sentiment analysis Linear algorithms are used in conventional approaches to sentiment analysis. This often entails using a machine learning technique to classify text as positive or negative, such as a support vector machine or linear regression [7]. These methods frequently depend on the usage of NLP features such part-of-speech tags, word embeddings, and sentiment lexicons and call for a large training data collection of labeled instances. With simple language and more conventional use cases, these methods typically work well.
348
M. V. Jagannatha Reddy et al.
17.3 Data mining techniques for sentiment analysis The use of data mining tools for sentiment analysis aids in understanding client perceptions of various business facets. To analyze customer reviews and determine how customers feel about the items or services, techniques including NLP, text analytics, data clustering, and sentiment analysis can be employed. These methods can be used to evaluate internet data, including reviews, surveys, and comments on social media, and to spot trends and patterns [8]. For example, data clustering techniques can help to categorize comments into different clusters to pinpoint the key drivers of customer opinion about a certain product or service. Text analytics is used to find themes in consumer feedback, and NLP techniques can be used to analyze unstructured text and extract sentiment. Last but not least, sentiment analysis classifies client input as good, neutral, or negative by using a variety of algorithms to determine sentiment.
17.3.1 Introduction to data mining Extraction of usable information from vast volumes of data is known as data mining. To find patterns, trends, and linkages, massive datasets must be sorted through. Techniques like machine learning, AI, and statistics are used in data mining to evaluate data and make inferences from it [9]. Data mining can be used to make predictions about future trends by understanding current patterns. Data mining is used in many areas such as healthcare, finance, marketing, and education. It can be used to gather information about customer preferences, and to analyze customer buying patterns. Data mining is also used to identify fraud and other anomalies in datasets.
17.3.2 Data collection and preprocessing Gathering, cleaning, organizing, and transforming data into a format that can be used by a machine learning algorithm for modeling and analysis are all parts of data collection and preprocessing. Data collection typically involves obtaining information from a variety of sources, including surveys, databases, websites, and social media [10]. Data preprocessing involves cleaning the data to correct errors, fixing inconsistencies, reducing noise and outliers, and transforming it into a form suitable for use by a machine learning algorithm. This process includes normalizing the data, transforming it into a suitable form, and eliminating redundant information. Additionally, data preprocessing often involves reducing the dimensions of the dataset by combining or eliminating features that are not relevant to the problem.
17 Exploring artificial intelligence techniques for enhanced sentiment analysis
349
17.3.3 Feature extraction and selection The technique of separating useful information from unprocessed data is called feature extraction. By lowering the amount of characteristics that describe the data, it is used to reduce the dimensionality of datasets. The process of choosing which features to include in the modeling process is known as feature selection [11]. The method may assist in increasing the model’s correctness or decreasing the computational expense associated with model evaluation. A subset of traits that have the greatest influence on a desired outcome are sought after through feature selection. The type of feature selection will depend largely on the type of data being analyzed. For example, a dataset consisting of text documents would require a different type of feature extraction/ selection than a dataset consisting of images.
17.3.4 Machine learning algorithms for sentiment analysis The process of evaluating a text to ascertain the underlying sentiment of the tone is known as sentiment analysis. In other words, it involves computationally recognizing and classifying viewpoints that people express in written form. In order to learn the sentiment related to a certain text, machine learning algorithms for sentiment analysis often rely on supervised learning approaches [12]. These algorithms often use labeled training data, which teaches the algorithm to associate the sentiment of certain words or phrases with the overall sentiment of the text. The trained model can then be used by the computer to ascertain the sentiment of new, unexplored material. Naive Bayes, Support Vector Machines, and Decision Trees are common algorithms for sentiment analysis.
17.3.5 Evaluation metrics for sentiment analysis models The accuracy of a sentiment analysis model is measured using evaluation metrics for sentiment analysis models. The performance of the model on a particular set of data is often quantified by these metrics, which may be used to compare other models. The area under the curve (AUC), F1, precision, recall, and accuracy are typical measures employed for this purpose. Additional metrics include the less popular Matthews correlation coefficient (MCC), mean absolute error, and mean square error. All of these metrics measure the model’s ability to correctly detect sentiment in texts, and each metric offers a different measure [13].
350
M. V. Jagannatha Reddy et al.
17.4 Artificial intelligence techniques for sentiment analysis Textual emotional and contextual information can be swiftly extracted using AI sentiment analysis techniques. Building a sentiment model that can precisely anticipate the emotion of a given text is helpful. To comprehend the sentiment of text, these models employ a variety of methodologies, including rule-based learning, supervised learning, unsupervised learning, and deep learning. Rule-based learning involves the rules extracted from training data to classify text. Supervised learning involves the use of labelled data to train a model [14]. Unsupervised learning models use clustering algorithms to classify text. Recurrent neural networks (RNNs) are used in deep learning to examine and extract useful characteristics from enormous datasets. All of these techniques help to automatically detect sentiment from text, which can be used for various applications such as customer feedback analysis or predicting the stock price direction.
17.4.1 Introduction to artificial intelligence A technology known as AI enables machines (computers, robots, etc.) to carry out functions including thinking, learning, and problem-solving. The foundation of AI is the notion that robots may be programmed with precise instructions to mimic the behaviors of a human thinker. AI has been applied in a wide array of areas, including finance, healthcare, business, and transportation, where it is used to improve processes, automate tedious tasks, and provide accurate information [15]. AI has been rapidly evolving over the past decade, and it is becoming increasingly popular as businesses and other organizations seek to increase efficiency and reduce costs. With the right programming, AI can also be an invaluable tool for producing new ideas, products, and services.
17.4.2 Natural language processing for sentiment analysis Understanding and processing human language is the focus of the branch of AI known as NLP. It makes it possible for machines to meaningfully analyze, comprehend, and produce natural language. The goal of sentiment analysis, which employs NLP, is to locate and classify the opinions stated in a text, particularly those relating to a good or service. It is used to identify client emotions, such as happiness, rage, or contempt toward a specific subject or product [16]. In order to analyze a text’s sentiment, compare it to established categories, and assign a sentiment score, NLP uses machine learning models like naive Bayes, support vector machine, and RNNs. This score can then be used to determine customer’s opinion about a particular product or service.
17 Exploring artificial intelligence techniques for enhanced sentiment analysis
351
17.4.3 Deep learning models for sentiment analysis Deep learning models create sentiment analysis for text documents using neural networks. These advanced networks are able to learn underlying features and patterns in text documents that classify them as positive or negative. This allows deep learning models to detect even subtle nuances in sentiment that human readers may miss [17]. Deep learning models recognize complex linguistic patterns, such as sarcasm, and can even recognize when an opinion is being expressed by an individual. These models can be trained on large datasets, allowing them to glean more information from a single text sample. As a result, deep learning-based sentiment analysis is known to be more accurate than traditional machine learning methods.
17.4.4 Transfer learning for sentiment analysis Transfer learning is the process of using sentiment analysis models that have already been trained. To solve a new problem, it entails using a pretrained model and finetuning its weights on a fresh dataset. This method can greatly increase a model’s accuracy and performance while reducing the amount of labeled data required to train a new model. It has been used to increase the precision of sentiment analysis models in machine learning and deep learning [18]. In sentiment analysis, transfer learning can be used to train a model on a dataset of labeled reviews from a specific domain and then use that model to analyze reviews in a different domain. This saves time and resources since you are not training a model from scratch for each new domain.
17.4.5 Reinforcement learning for sentiment analysis A subfield of machine learning called reinforcement learning for sentiment analysis is concerned with exploiting past performance to enhance the effectiveness of a sentiment analysis system. It uses rewards or punishment to help the system automatically adjust how it works in order to achieve a desired result. Reinforcement learning uses an online approach, where a model can learn from its own mistakes in order to properly analyze sentiment. Additionally, reinforcement learning can be used to find patterns within unstructured data or to generate sentiment scores on new data [19]. This helps to produce more accurate and reliable sentiment classifications. Ultimately, reinforcement learning for sentiment analysis helps to reduce time and effort needed to get the desired performance from a sentiment analysis system.
352
M. V. Jagannatha Reddy et al.
17.5 Enhancing sentiment analysis through AI techniques The use of AI techniques in sentiment analysis can significantly increase the accuracy of sentiment predicting from text. To recognize and analyze the sentiment of the text, AI techniques including NLP, reinforcement learning, and deep learning are being applied. These methods enable the training of models to comprehend the sentiment of a huge number of text samples. They can also identify sentiment in unseen text by relying on previously acquired knowledge and capturing more subtle differences between different sentiments [20]. Using AI in sentiment analysis also allows for more resources to be developed to help identify the relationship between language and sentiment. This includes understanding the impact of linguistic factors like syntax, semantics, and discourse on sentiment. Additionally, AI can be used to identify relationships between sentiment and contextual elements such as emotions, culture, and political beliefs. The AI can also be utilized to generate sentiment-specific features. These features can be used to develop a sentiment-specific model, allowing for more accurate sentiment predictions in new text [21]. The application of AI to sentiment analysis can result in more precise and better predictions, whether they are made for study, commercial usage, or both.
17.5.1 Leveraging word embeddings for sentiment analysis A potent method for determining the sentiment of text is to use word embeddings for sentiment analysis. Word embeddings are numerical representations of words, capable of capturing semantic relationships between the words. Sentiment analysis is the understanding of a text’s emotional tone, expressed through words or phrases, to determine an overall sentiment. Utilizing word embeddings in sentiment analysis enables the capturing of associations between words; for instance, if a word is linked with a positive sentiment, it may alternatively be connected with a negative sentiment if it appears next another word that is associated with a negative attitude [22]. This contextual relationship between words is what enables word embeddings to capture sentiment. By leveraging word embeddings, sentiment analysis can be achieved faster and more accurately than using traditional methods, such as bag-of-words.
17.5.2 Handling negation and sarcasm in sentiment analysis Negation and sarcasm in sentiment analysis can be extremely tricky to detect as they are highly subjective and sometimes depend on tone, context, and even cultural background. Negation occurs when a negative word is used, or when the opinion expressed using positive words is intended to have a negative connotation. For example,
17 Exploring artificial intelligence techniques for enhanced sentiment analysis
353
the phrase “not bad” could indicate something that is either acceptable or mediocre, depending on how it is said. Sarcasm is a form of irony which often requires an understanding of cultural context and past events to be detected accurately [23]. It often uses positive language or phrases to portray a negative sentiment. Sentiment analysis requires more than just a vocabulary of positive and negative terms to handle negation and sarcasm. Irony and sarcasm can be recognized in language by using NLP techniques. Natural language understanding (NLU) can be used to analyze the context of a statement to infer the intended sentiment. Machine Learning algorithms can be employed to detect sentiment in a large amount of data. This can involve training the models to recognize nuances like irony, sarcasm, or negation. Ultimately, sentiment analysis tools should rely on both human expertise and AI approaches to accurately capture the sentiment of a statement.
17.5.3 Incorporating contextual information for improved sentiment analysis Incorporating contextual information for improved sentiment analysis is a concept which involves using a variety of contextual information sources, such as news, customer reviews, and social media, in order to provide more accurate and meaningful sentiment analysis of a given text. Contextual information helps to identify the sentiment expressed by the speaker, providing an additional layer of understanding. By leveraging this added contextual information, a sentiment analysis system is able to better assess how the sentiment of a given text impacts the overall sentiment of the discussion. For example, a sentence like “I love this product” may hold more weight when considering the context. In the sentence, the speaker might be talking about a product they’ve heard about for the first time and has strong positive sentiments associated with it [24]. However, if they were talking about a product they’ve had a negative experience with in the past, their sentiments would naturally be more negative even if they used the same phrase. Incorporating contextual information can also help to create more balanced sentiment analysis results, as it allows sentiment analysis systems to have a greater reach.
17.5.4 Addressing imbalanced datasets in sentiment analysis Imbalanced data is an issue in sentiment analysis as it affects the accuracy of the classification algorithms. Class imbalances in sentiment datasets arise from a minority of sentiment classes being heavily present in the dataset, while other sentiment classes are minority classes, and often difficult to classify accurately. Addressing class imbalance in sentiment datasets is important to ensure accurate predictions for all sentiment classes [25]. Undersampling or oversampling data points from the minority
354
M. V. Jagannatha Reddy et al.
classes, or combining the two strategies, is the most typical method of addressing class imbalance in emotion datasets. When sampling is too small, some data points from the majority class are discarded. Undersampling is used to increase the amount of data points from the minority class while decreasing the number of data points from the majority class. This helps to create a more balanced dataset that can be used to train the sentiment classifier [26]. Oversampling involves the duplication of minority data points. This technique can be used to create more training data for the minority class. Oversampling is used to provide a more balanced dataset for training the sentiment classifier by increasing the quantity of data points from the minority class. In addition to undersampling and oversampling, another technique that can be used to address class imbalance in sentiment datasets is to introduce weights to the data points in different classes. Weights can be used to increase the importance of minority classes during training and reduce the importance of majority classes. To sum up, the techniques that can be used to address class imbalance in sentiment datasets are undersampling, oversampling, and introducing weights to the data points in different classes. These approaches can help to ensure accurate predictions for all sentiment classes.
17.6 Applications of AI techniques in sentiment analysis Sentiment analysis, commonly referred to as opinion mining, is a type of NLP that uses AI methods to locate, extract, quantify, and analyze affective states and subjective data. This method is employed to acquire and examine user opinions on a specific good, service, or subject in order to make wise decisions [27]. Numerous activities, including sentiment tracking, opinion mining, sentiment classification, automated opinion extraction, sentiment summarization, and sentiment analysis for text and images, can be performed using AI algorithms for sentiment analysis. AI techniques allow for faster and more accurate analysis of large volumes of text data, including sentiment-rich sources such as social media platforms and customer reviews. Sentiment analysis can be used to gain insights into customer behavior, as well as to identify potential risk factors that could result in customer dissatisfaction and turnover. Additionally, sentiment research can assist marketers in better comprehending client demands and preferences, enabling them to more successfully adapt their strategies to the target audience [28]. Finally, AI-based sentiment analysis can be used to detect signs of online fraud and fake reviews, thus protecting the integrity of businesses and products.
17 Exploring artificial intelligence techniques for enhanced sentiment analysis
355
17.6.1 Sentiment analysis in social media Sentiment analysis in social media is the process of identifying and extracting subjective information from user-generated material in social media using NLP and computational linguistic approaches. It is employed to track public opinion about a certain company, good, or service. By categorizing comments made on social media as favorable, negative, or neutral, it aids in determining how the general population feels about a particular subject [29]. For sentiment categorization, topic extraction, opinion mining, sentiment trend analysis, and brand monitoring, sentiment analysis can be employed.
17.6.2 Sentiment analysis in customer reviews Customer reviews about a product, service, or company are analyzed using a process called sentiment analysis. By recognizing any positive, negative, or neutral sentiment that may be present, sentiment analysis seeks to quantify the total sentiment of customer reviews. This can be used to identify trends in customer sentiment and gain an understanding of how customers feel about the product, service, or business in question [30]. The output of such an analysis can be used to inform marketing or product strategy decisions.
17.6.3 Sentiment analysis in political discourse Sentiment analysis in political discourse uses language-based algorithms to assess the sentiment (positive, negative, or neutral) of written or spoken political discourse. Sentiment analysis in political discourse is a form of NLP. This type of analysis can be used to uncover public opinion of political candidates and policies, providing insight into unmet needs and potential areas for further discussion. Additionally, campaigns, social media monitoring, and political polls can all benefit from sentiment analysis [31]. By identifying negative sentiments, campaigns can address those issues in their messaging and work to proactively reach out to those voices. This type of analysis can also allow for effective message targeting, allowing campaigners to craft tailored messages to broad demographics.
17.7 Challenges and future directions To identify the emotions and attitudes expressed in written language, an NLP problem known as sentiment analysis is used. It is a branch of AI that deals with deciphering human sentiment patterns. Sentiment analysis typically involves text classifiers that
356
M. V. Jagannatha Reddy et al.
are trained to categorize text as positive or negative, or some other sentiment-based categories. Challenges: 1. Natural language understanding: Sentiment analysis is based on NLU technology, which is still far from a perfect science. Complex linguistic constructs, slang, acronyms, etc. can all throw off otherwise perfectly accurate sentiment analysis. 2. Context and subjectivity: Sentiment analysis is subjective and depends on the context of a given message. Sentiment analysis is susceptible to context misunderstanding because the context in which a sentence is used might modify its meaning. 3. Sentiment polarity: The polarity of a statement is difficult to detect, given the subjectivity involved. Many sentiment analysis algorithms use a binary classification such as positive vs. negative but there is a need for a more nuanced approach that can better identify sentiment polarity across a wider range of emotions. Future directions: 1. Deep learning: A subclass of AI, deep learning is particularly helpful for challenging tasks like sentiment analysis. Deep learning neural networks are capable of discovering patterns in data that are not accessible to conventional machine learning algorithms. This can result in sentiment analysis that is more precise. 2. Transfer learning: This method can be applied to increase the precision of sentiment analysis. The accuracy of sentiment analysis can be increased by utilizing pretrained models developed on substantial datasets. 3. Social media analysis: As the use of social media keeps expanding, sentiment analysis of social media data is an area of growing interest. Decisions can be made using social media sentiment analysis, which can offer insightful information about public opinion on a variety of subjects.
17.7.1 Limitations of current AI techniques in sentiment analysis – –
– –
The current AI techniques for sentiment analysis have difficulty handling sarcasm, irony, and nuance as sentiment is not always clearly stated. AI sentiment analysis algorithms are not yet able to truly understand the true sentiment of a statement or phrase, as a result of the lack of ability to infer meaning and context. The restricted capacity of AI sentiment analysis to obtain and examine vast volumes of organized and unstructured data is another drawback. Many current AI models have difficulty recognizing and understanding the intent of a statement.
17 Exploring artificial intelligence techniques for enhanced sentiment analysis
–
357
It is also difficult to generate accurate sentiment scores as different people may interpret the same statement differently. This often leads to the underestimation and overestimation of sentiment ratings.
17.7.2 Ethical considerations in sentiment analysis Ethical considerations in sentiment analysis refer to the moral dilemmas that may arise when analyzing the sentiment of text or objects [32]. It includes understanding how people’s ideologies, cultures, and even personal beliefs may influence the analysis. As sentiment analysis becomes increasingly popular, it is important to adhere and consider ethical practices. This includes respecting people’s privacy, protecting confidential data, and avoiding bias when analyzing data. Additionally, it is important to consider how sentiment analysis can affect certain groups, such as those belonging to minority communities or even those belonging to technological literacies [33]. Finally, the ethical considerations include the understanding of the implications of the, making sure they are considered when drawing conclusions from the analysis.
17.7.3 Future trends and research directions In the future, AI will continue to be used in sentiment analysis, but the way it is used will be more refined and specialized. AI-powered sentiment analysis will become more accurate and adept at interpreting the complexities of language and subtle nuances in sentiment [34, 35]. In order to better comprehend emotion, AI will also be utilized to create better datasets and feature sets. NLP, which uses AI to comprehend the meaning of words and phrases, is another increasingly significant method for sentiment analysis. AI techniques such as deep learning and neural networks are being used to identify sentiment in texts more accurately and faster than ever before. In addition, AI-powered sentiment analysis applications are becoming increasingly more context aware. By utilizing NLP, AI can interpret the nuanced emotions associated with a particular sentence or phrase and interpret the sentiment accurately [36]. The AI-driven sentiment analysis will be used to better understand customer behavior. AI can analyze customer reviews and conversations from various platforms and provide businesses with valuable insights on how customers interact with and feel about different products or services. This insight can then be used to create better customer experiences and improve customer service.
358
M. V. Jagannatha Reddy et al.
References [1] [2] [3]
[4] [5] [6] [7] [8] [9]
[10]
[11] [12]
[13] [14]
[15]
[16]
[17]
[18]
Xu, G., Meng, Y., Qiu, X., Yu, Z., & Wu, X. (2019). Sentiment analysis of comment texts based on BiLSTM. IEEE Access, 7, 51522–51532. Mehta, P., & Pandya, S. (2020). A review on sentiment analysis methodologies, practices and applications. International Journal of Scientific and Technology Research, 9(2), 601–609. Zad, S., Heidari, M., Jones, J. H., & Uzuner, O. (2021, May). A survey on concept-level sentiment analysis techniques of textual data. In 2021 IEEE World AI IoT Congress (AIIoT) (pp. 0285–0291). IEEE. Pereira, D. A. (2021). A survey of sentiment analysis in the Portuguese language. Artificial Intelligence Review, 54(2), 1087–1115. Elfaik, H., & Nfaoui, E. H. (2020). Deep bidirectional LSTM network learning-based sentiment analysis for Arabic text. Journal of Intelligent Systems, 30(1), 395–412. Chen, L. C., Lee, C. M., & Chen, M. Y. (2020). Exploration of social media for sentiment analysis using deep learning. Soft Computing, 24, 8187–8197. Saura, J. R., Palos-Sanchez, P., & Grilo, A. (2019). Detecting indicators for startup business success: Sentiment analysis using text data mining. Sustainability, 11(3), 917. Mittal, A., & Patidar, S. (2019, July). Sentiment analysis on twitter data: A survey. In Proceedings of the 7th International Conference on Computer and Communications Management (pp. 91–95). Kurani, A., Doshi, P., Vakharia, A., & Shah, M. (2023). A comprehensive comparative study of artificial neural network (ANN) and support vector machines (SVM) on stock forecasting. Annals of Data Science, 10(1), 183–208. Basiri, M. E., Nemati, S., Abdar, M., Cambria, E., & Acharya, U. R. (2021). ABCDM: An attention-based bidirectional CNN-RNN deep model for sentiment analysis. Future Generation Computer Systems, 115, 279–294. Wankhade, M., Rao, A. C. S., & Kulkarni, C. (2022). A survey on sentiment analysis methods, applications, and challenges. Artificial Intelligence Review, 55(7), 5731–5780. Ali, N. M., Abd El Hamid, M. M., & Youssif, A. (2019). Sentiment analysis for movies reviews dataset using deep learning models. International Journal of Data Mining & Knowledge Management Process (IJDKP), 1–9. Yadav, S. P., & Yadav, S. (2018). Fusion of medical images in wavelet domain: a discrete mathematical model. Ingeniería Solidaria, 14(25), 1–11. Gobinathan, B., Mukunthan, M. A., Surendran, S., Somasundaram, K., Moeed, S. A., Niranjan, P., . . . & Sundramurthy, V. P. (2021). A novel method to solve real time security issues in software industry using advanced cryptographic techniques. Scientific Programming, 2021, 1–9. Yadav, S. P., & Yadav, S. (2019). Mathematical implementation of fusion of medical images in continuous wavelet domain. Journal of Advanced Research in Dynamical and Control System, 10(10), 45–54. Sivaram, M., Yuvaraj, D., Mohammed, A. S., Manikandan, V., Porkodi, V., & Yuvaraj, N. (2019). Improved enhanced DBTMA with contention-aware admission control to improve the network performance in MANETS. CMC-Computers Materials and Continua, 60(2), 435–454. Rani, P., Verma, S., Yadav, S. P., Rai, B. K., Naruka, M. S., & Kumar, D. (2022). Simulation of the lightweight blockchain technique based on privacy and security for healthcare data for the cloud system. International Journal of E-Health and Medical Communications (IJEHMC), 13(4), 1–15. Manikandan, V., & Yuvaraj, N. (2019, May). Advanced expert system using particle swarm optimization based adaptive network based fuzzy inference system to diagnose the physical constitution of human body. In Emerging Technologies in Computer Engineering: Microservices in Big Data Analytics: Second International Conference, ICETCE 2019, Jaipur, India, February 1–2, 2019, Revised Selected Papers (Vol. 985, pp. 349). Springer.
17 Exploring artificial intelligence techniques for enhanced sentiment analysis
359
[19] Pujahari, R. M., Yadav, S. P., & Khan, R. (2022). Intelligent farming system through weather forecast support and crop production. In : Mohammad Ayoub Khan, Rijwan Khan and Mohammad Aslam Ansari (Eds.). Application of Machine Learning in Agriculture (pp. 113–130). United States: Academic Press. [20] Yuvaraj, N., Raja, R. A., Kousik, N. V., Johri, P., & Diván, M. J. (2020). Analysis on the prediction of central line-associated bloodstream infections (CLABSI) using deep neural network classification. In: Jitendra Kumar Verma, Sudip Paul and Prashant Johri (Eds.). Computational Intelligence and Its Applications in Healthcare (pp. 229–244). United States: Academic Press. [21] Irfan, D., Tang, X., Narayan, V., Mall, P. K., Srivastava, S., & Saravanan, V. (2022). Prediction of Quality Food Sale in Mart Using the AI-Based TOR Method. Journal of Food Quality, 2022, 1–13. [22] Ramesh, G., Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). Prediction of energy production level in large PV plants through AUTO-Encoder Based Neural-Network (AUTO-NN) with restricted Boltzmann feature extraction. Future Internet, 15(2), 46. [23] Rajalakshmi, M., Saravanan, V., Arunprasad, V., Romero, C. T., Khalaf, O. I., & Karthik, C. (2022). Machine learning for modeling and control of industrial clarifier process. Intelligent Automation and Soft Computing, 32(1), 339–359. [24] Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). A novel architecture of intelligent decision model for efficient resource allocation in 5G broadband communication networks. ICTACT Journal on Soft Computing, 13(3), 2986–2994. [25] Saravanan, V., & Neeraja, A. (2013, January). Security issues in computer networks and stegnography. In 2013 7th International Conference on Intelligent Systems and Control (ISCO) (pp. 363–366). IEEE. [26] Sangeetha, R., Logeshwaran, J., Rocher, J., & Lloret, J. (2023). An improved agro deep learning model for detection of Panama wilts disease in banana leaves. AgriEngineering, 5(2), 660–679. [27] Singh, J., Deepika, J., Sathyendra Bhat, J., Kumararaja, V., Vikram, R., Jegathesh Amalraj, J., . . . & Sakthivel, S. (2022). Energy-efficient clustering and routing algorithm using hybrid fuzzy with grey wolf optimization in wireless sensor networks. Security and Communication Networks, 2022, 1–13. [28] Logeshwaran, J., Kiruthiga, T., Kannadasan, R., Vijayaraja, L., Alqahtani, A., Alqahtani, N., & Alsulami, A. A. (2023). Smart load-based resource optimization model to enhance the performance of deviceto-device communication in 5G-WPAN. Electronics, 12(8), 1821. [29] Harish, B. S., Kumar, K., & Darshan, H. K. (2019). Sentiment analysis on IMDb movie reviews using hybrid feature extraction method. International Journal of Interactive Multimedia and Artificial Intelligence, 45(2), 1–10. [30] Jain, P. K., Pamula, R., & Srivastava, G. (2021). A systematic literature review on machine learning applications for consumer sentiment analysis using online reviews. Computer Science Review, 41, 100413. [31] Chen, X., Zou, D., Xie, H., Cheng, G., & Liu, C. (2022). Two decades of artificial intelligence in education. Educational Technology and Society, 25(1), 28–47. [32] Abualigah, L., Alfar, H. E., Shehab, M., & Hussein, A. M. A. (2020). Sentiment analyss in healthcare: A brief review. Recent Advances in NLP: The Case of Arabic Language, 67, 129–141. [33] Kumar, A., & Garg, G. (2020). Systematic literature review on context-based sentiment analysis in social multimedia. Multimedia Tools and Applications, 79, 15349–15380. [34] Alsaeedi, A., & Khan, M. Z. (2019). A study on sentiment analysis techniques of Twitter data. International Journal of Advanced Computer Science and Applications, 10(2), 1–18. [35] Xu, G., Yu, Z., Yao, H., Li, F., Meng, Y., & Wu, X. (2019). Chinese text sentiment analysis based on extended sentiment dictionary. IEEE Access, 7, 43749–43762. [36] Nemes, L., & Kiss, A. (2021). Social media sentiment analysis based on COVID-19. Journal of Information and Telecommunication, 5(1), 1–15.
Senthil Kumar C, A. R. Arunarani, Piyush Charan, Sanjeev Kumar Angadi
18 Exploring the potential of artificial intelligence for automated sentiment analysis and opinion mining Abstract: The development of automated sentiment analysis and opinion mining approaches for enhancing the precision and scalability of text analysis has been made possible by the rising integration of artificial intelligence (AI) technology in digital systems. The method of automatically evaluating and extracting sentiment-related data from a text is known as automatic sentiment analysis. Automatically extracting and summarizing opinions from a text is referred to as opinion mining. AI has enormous promise for automated sentiment analysis and opinion mining, which can increase consumer engagement and experience. Keywords: Artificial intelligence, digital systems, automated, sentimental analysis, text
18.1 Introduction Artificial intelligence (AI)-based automated sentiment analysis and opinion mining techniques provide a reliable means of deriving customers’ emotion or opinions from text sources such as social posts, reviews, emails, and surveys. Such technique can help to uncover customer sentiment on a product, service, or company, allowing businesses to better understand their customers in terms of overall satisfaction and provide better customer service [1]. AI-driven techniques further allow for automated processes for deriving the sentiment of large numbers of text sources. This is especially beneficial in industries where customer feedback is essential, such as in customer service, healthcare, retail, and hospitality.
18.1.1 Background and significance Automated sentiment analysis and AI have completely changed how people interact and communicate online. Machines can now comprehend and further extract meaning from text thanks to AI technologies like natural language processing (NLP) [2]. In turn, this has made it possible for automated sentiment analysis and opinion mining techniques to better understand people’s sentiments and ideas as they are conveyed in textual data. https://doi.org/10.1515/9783111323749-018
362
Senthil Kumar C et al.
Sentiment analysis has become increasingly important in today’s competitive business environment. It allows companies to gain insight into their customers’ attitudes and perceptions, allowing them to adjust their offerings to better suit the needs of their clientele. This can be especially useful for brand recognition and reputation management, as well as for crisis management [3]. Furthermore, automated sentiment analysis and opinion mining techniques can aid social networks, content creators, and search engines in better comprehending the sentiment of their users’ interactions, enabling them to provide their users with more pertinent information[4]. The automated sentiment analysis has become an essential tool for businesses in order to gain an in-depth understanding of their customers, aiding in marketing, customer service, and decision-making. As AI and automated sentiment analysis continue to develop, the technology will only become increasingly important for businesses in understanding their customers and creating an even more personalized experience.
18.1.2 Objectives of the study The objectives of AI for automated sentiment analysis and opinion mining are: 1. To understand human emotions in social media and other textual data. 2. To identify the opinions of the author about any product or service by analyzing the text in the review. 3. To classify sentiment into one of the four categories (positive, negative, neutral, or mixed) based on the content of the review. 4. Identifying trends in opinion for further market research, analysis, and decisionmaking. 5. Identifying customer service issues and providing timely solutions. 6. Encouraging customer engagement and loyalty. 7. To generate insights on customer sentiment to make improvements to product and services. 8. To predict customer behavior and analyze the impact of marketing activities.
18.1.3 Scope and limitations The usage of automated systems has a significant impact on the capabilities and restrictions of AI for automated sentiment analysis and opinion mining. Text analysis techniques are typically used to discover the emotional content in texts in order for automated sentiment analysis and opinion mining to be able to determine the overall sentiment communicated in the text. Additionally, it can identify the precise opinions contained in the text, as well as the mood of those ideas (whether good or negative) [5].
18 Exploring the potential of artificial
363
However, automated sentiment analysis and opinion mining have some limitations. They are often reliant upon NLP techniques, which means the accuracy of the outputs are greatly influenced by the complexity of the text. Automated systems are also limited in their ability to understand the sentiment and opinion being conveyed in a text. For example, the systems might look for certain key phrases or words to detect sentiment, but may fail to identify the more nuanced elements of sentiment, such as sarcasm or double-entendres. The quality of the data being used in the analysis will also affect automated sentiment analysis and opinion mining. Erroneous or biased data in a dataset may produce erroneous results. It is crucial to be aware of the potential scope and restrictions when establishing an automated system because automated sentiment analysis and opinion mining have enormous potential to be employed in a variety of applications [6].
18.2 Sentiment analysis and opinion mining Sentiment analysis, commonly referred to as opinion mining, is the act of recognizing and examining emotions in text in order to ascertain the overall attitude of a piece of writing or a sentence. It has many uses, including customer service, market research, and polling the general population. It entails taking important textual elements such significant words, subjects, phrases, and emotions out of the text and giving each one a sentiment score [7]. This rating reveals if the text’s general tone is upbeat, downbeat, or neutral. Sentiment analysis can also be utilized to identify fake or spam reviews.
18.2.1 Definition and concepts Identifying the emotional undertone of a string of words through the method of sentiment analysis allows one to classify a passage of text as positive, negative, or neutral. Finding out about the attitudes, ideas, and feelings conveyed in the text can be accomplished through studying sentiment in the text. This can be used to identify trends in public opinion or understand consumer sentiment towards a particular product, service, or company [8]. It can also be used to detect sarcasm and other subtleties in language that are not able to be detected through traditional textual analysis methods.
18.2.2 Applications and importance The method of sentiment analysis involves employing text analysis technologies to automatically identify the conveyed sentiment in a text. The process, commonly referred
364
Senthil Kumar C et al.
to as opinion mining, seeks to find subjective information in texts. In order to find patterns and trends in text data from multiple sources, text mining techniques are used [9]. Sentiment Analysis is used to gain an understanding of the attitude, opinions and emotions expressed in text data. It may be used to categorize customer feedback into positive and negative categories, in order to measure customer satisfaction [10]. Sentiment analysis applications range from market research to customer service improvement. Businesses can use it to better understand what consumers think and feel about their goods and services. It can also be used as an early warning system to detect customer dissatisfaction, in order to take corrective action and improve customer service. Furthermore, this information can be used to inform product and service improvement plans [11].
18.2.3 Challenges and limitations Challenges: – Interpretability: As opinion mining techniques are mainly based on complex algorithms, it can be difficult to understand their internal behavior. This makes it hard to build trust in the opinion mining systems and interpret the results they generate accurately. – Contextual Issues: Opinion mining systems can be affected by context-specific terms and phrases, which can make it difficult to generalize findings and apply them in similar situations. – Ambiguity: The meanings of words can differ based on the context and author, making it difficult to understand the true sentiment behind a particular phrase or sentence. – Variety of opinions: Opinion mining systems are often trained on a limited dataset, which does not always reflect the full range of opinions that exist in the real world. Limitations: – Limited accuracy: The accuracy of opinion mining systems can still be affected by the quality of the dataset they are trained on and the complexity of the algorithms. – Expensive: Developing and deploying a robust opinion mining system can be costly, as it requires a significant amount of resources in terms of data acquisition, development, and deployment. – Time-consuming: Training an opinion mining system can be time-consuming and requires expertise in machine learning. This can lead to delayed results in a live system.
18 Exploring the potential of artificial
365
18.3 Artificial intelligence for sentiment analysis Sentiment analysis is a procedure that uses text analysis, AI, and NLP to identify and detect the sentiment of a given text. It can be used to forecast how customers will feel about a given good, service, or other subject. Companies can utilize the sentiment analysis results to establish plans for customer interaction, marketing, and customer service and to better understand customer opinion [12]. To evaluate and decipher the underlying feelings or emotions in huge volumes of data, sentiment analysis employs AI. Important sentiment trends and consumer sentiment patterns can be uncovered by AI-based sentiment analysis technologies, which can be used to inform decisions regarding business activities and marketing efforts.
18.3.1 Overview of artificial intelligence techniques One of the most debated and cutting-edge technologies used in enterprises today is AI. Robotic process automation (RPA), deep learning, machine learning (ML), and NLP are just a few of the methods used in AI. AI is designed to allow machines and computer programs to learn from experience, identify trends, and make decisions with human-like intelligence. AI techniques are applied to enterprise applications such as security systems, rules engines, chatbots, and enterprise search to automate processes, provide higher levels of accuracy, and reduce costs across many industries [13]. AI is also becoming more widely used in the healthcare sector to identify and manage patient health conditions, reduce medical errors, and assist with precise drug doses and medication management. Furthermore, AI is used to power autonomous vehicles as well as to power robotic arms in modern factories. AI is constantly advancing and being used in a greater number of applications, and more innovative uses are continually being developed.
18.3.2 Machine learning approaches In ML, a subset of AI, computers may “learn” from data without having to be explicitly taught. In order to produce predictions or choices, ML approaches rely on algorithms that can spot patterns in data. ML approaches can be divided into three categories: supervised learning, unsupervised learning, and reinforcement learning [14]. Algorithms are trained on data and given the desired outcomes as part of supervised learning. The algorithm will look for a pattern that corresponds to the anticipated outcomes as new data is submitted to it. You can utilize supervised learning to find solutions to regression, classification, and ranking issues. Contrarily, unsupervised learning doesn’t rely on the data having labels that have already been established. Instead, the system is given an unlabeled dataset and is
366
Senthil Kumar C et al.
forced to identify patterns on its own. Clustering and anomaly detection are two common unsupervised learning applications. A further ML strategy is the use of reinforcement learning algorithms. Algorithms for reinforcement learning are taught to iterate over challenges to find the optimal answer. Problems requiring the agent to learn from its environment or have complex incentive systems can be resolved with this type of learning.
18.3.3 Supervised learning The machine is taught about the connections between input and output data via supervised learning, a sort of ML method. By giving the algorithm labeled examples to refer to when making judgments and predictions, the system gets taught. A prediction model is created using the input data using the supervised learning algorithm. After that, this model is used to unobserved data to produce predictions and categorizations [15].
18.3.4 Unsupervised learning Unsupervised learning is a type of ML technique that searches unlabeled data for hidden patterns without requiring any prior knowledge or supervision. It is frequently used to uncover significant structure in large data sets. Finding hidden connections between datasets of records or between sets of distinct variables is helpful. It reveals insights from an unexplored dataset and can be applied to find data anomalies. Unsupervised learning methods provide the computer the ability to sleuth out hidden patterns in data. Clustering and association rules are examples of unsupervised learning [16].
18.3.5 Semi-supervised learning A ML method called semi-supervised learning trains models using both labeled and unlabeled input. The main goal of semi-supervised learning is to supplement labeled data with unlabeled data in order to learn from a bigger set of data. This will help the model generalize more effectively and enhance its predictive ability. Semi-supervised learning algorithms are typically used for classification and clustering tasks [17]. Some example scenarios of using semi-supervised learning include image recognition, spam detection, and NLP.
18 Exploring the potential of artificial
367
18.4 Deep learning for sentiment analysis Deep learning for sentiment analysis is the process of using computational tools to analyze texts or documents to extract and identify positive, negative, or neutral sentiment from data sets. It is a type of NLP that leverages automated techniques to understand how words are being used in sentences that have an emotional context. Deep learning refers to a set of methods and algorithms that allow an AI system to “learn” from its environment and perform tasks without relying on explicitly programmed rules and structures [18]. This enables a machine to “learn” patterns by looking at examples of data, rather than relying on the programmer to program rules. In the case of sentiment analysis, deep learning can be used to identify the sentiment of a given text. For example, it can identify the sentiment of a tweet or a review on an online service. Deep learning can accurately identify the sentiment of a given positive or negative sentiment without relying on explicit annotated keywords or labels.
18.4.1 Introduction to deep learning A branch of ML called deep learning uses artificial neural networks – algorithms based on the structure and operation of the human brain – to identify patterns and learn from massive datasets. Deep learning models require substantially more data and computer capacity to find patterns and relationships in data than do classic techniques [19]. Deep learning can grasp and interpret context and meaning in text, audio, and video when paired with methods from the field of NLP, such as natural language understanding (NLU). Computer vision, audio recognition, NLP, and robotics are just a few of the issues that deep learning is being utilized to solve.
18.4.2 Convolutional neural networks A sort of artificial neural network called a convolutional neural network (CNN) is employed in computer vision, NLP, and a number of other applications. A CNN functions by taking a set of input images, running a series of convolutional filters across the images at various sizes, and then extracting a set of features from the images. The tasks that the features from the CNN may subsequently be applied to include sentiment analysis, object recognition, and image segmentation [20]. Typically, CNNs operate by taking a set of input images, running a number of convolutional filters across the images at various scales, pooling the features, and then flattening the input images into a high-dimensional vector.
368
Senthil Kumar C et al.
18.4.3 Recurrent neural networks An artificial neural network known as a “recurrent neural network” (RNN) is made specifically to analyze data having recurrent structure, such as sequences, video streams, text, and audio. RNNs have the ability to process data over several time steps as opposed to a single direction, like standard neural networks, which enables them to retain and utilize information from previous steps to better inform next stages [21]. NLU, music production, and other temporal tasks are particularly well suited for RNNs. They can also be applied to time-series analysis for other purposes, such as forecasting the weather and predicting market prices.
18.4.4 Long short-term memory (LSTM) A unique kind of RNN called long short-term memory (LSTM) is utilized in deep learning. It is used to remember long-term data dependencies and is made for sequence prediction challenges. Using unique “memory cells” that can retain data for lengthy periods of time, it can solve the vanishing gradient problem experienced by conventional RNNs [22]. It can accurately detect patterns in data sequences as well as learn complex connections. LSTM can be applied to a variety of tasks, including time series analysis, text generation, language modeling, and speech recognition.
18.5 Techniques for opinion mining An NLP method for drawing conclusions from texts is called sentiment analysis or opinion mining. It sometimes goes by the name “sentiment analysis,” and its objective is to find and extract subjective information from a text. It is frequently employed by businesses to comprehend the viewpoints of clients, staff members, and other stakeholders. Sentiment analysis, opinion mining, sentiment-driven summarizing, sentiment-driven topic grouping, topic modeling, sentiment polarity identification, and sentiment-based opinion summarization are techniques for extracting opinions. Sentiment analysis looks for a writer’s perspective on a subject or the general polarity of the context in a document [23]. Opinion mining looks for opinions or subjective statements expressed within a text through various techniques such as polarity identification, subjectivity detection, opinion summarization, opinion summarization, and opinion-based topic modeling. Sentiment-driven summarization looks for the opinion bearing events in text and summarizes the facts in the document. Sentiment-driven topic grouping and topic modeling are methods for generating clusters of information about topics in the documents according to sentiments expressed by the writer. Sentiment polarity detection looks for occurrences of certain phrases or words that are as-
18 Exploring the potential of artificial
369
sociated with specific sentiment or emotions. The sentiment-based opinion summarization aims to bring together the reports of an individual opinion from different sources.
18.5.1 Rule-based approaches Rule-based approaches in sentiment analysis use predefined rules like regular expressions, lexicons, and syntactic parsing to classify the sentiment of text. The goal of these methods is to classify the sentiment of text by using predefined rules [24]. These methods rely heavily on prior knowledge of language and can be used out of the box with a set of predefined rules. However, for a particular domain or language, custom setup rules are also necessary for the successful application of the rulebased approach [25]. The main advantages of rule-based approaches in sentiment analysis are that they are easy to understand, easy to implement, and cost-effective. Compared to MLbased approaches, they require less maintenance and data processing as pre-defined rules are used. On the downside, rule-based approaches are limited in their ability to detect complex sentiment and associations between words [26].
18.5.2 Lexicon-based approaches Lexicon-based approaches for sentimental analysis are a type of NLP technique that use dictionary-based lexicons to analyze the sentiment of a given text. This method entails labeling words in a text with sentimental categories like positive, negative, or neutral. After that, the document’s overall sentiment is ascertained using the sentiment labels. In lexicon-based methods, sentiment labels are applied to specific words in accordance with a specified sentiment lexicon. This lexicon is often a set of words and phrases with the labels that correspond to the respective sentiments [27]. With this method, sentiment analysis may be completed for vast amounts of text rapidly and accurately. Due to the fact that the sentiment labels have already been established, little training data is also needed. Last but not least, this strategy may typically be interpreted in terms of the underlying sentiment labels employed in the sentiment lexicon.
18.5.3 Aspect-based sentiment analysis An NLP technique called element-based sentiment analysis (ABSA) uses sentiment analysis to find and extract a text’s views in relation to a particular element. As it refers to a certain feature, such as a product, movie, or service, aspect-based senti-
370
Senthil Kumar C et al.
ment analysis aids in determining the sentiment of a text [28]. For instance, it is possible to evaluate a movie review to ascertain the general mood – whether positive or negative – as well as the reviewer’s opinions on particular elements, such as the cast, direction, and script. Aside from helping with customer support, ABSA can also be used to learn more about how customers feel about a given good or service.
18.6 Challenges and future directions Analysis of people’s views, sentiments, feelings, and attitudes toward a specific subject or event is known as opinion mining. It is a branch of NLP that has recently experienced tremendous growth because of its capacity to examine user-generated material and offer insightful data to businesses and organizations [29]. One of the major challenges of opinion mining is dealing with noisy data. This refers to data that contains errors, inconsistencies, misspellings, and other irregularities that can make it difficult to accurately analyze. The sheer volume of user-generated content also poses a challenge for opinion mining, as it must be manually filtered through to extract meaningful sentiments [30, 31]. Online conversations are often highly dynamic and a single event can spark numerous related conversations, making it difficult to capture the full context of a conversation [32]. There are various areas that need to be researched in order to further increase the accuracy and effectiveness of the opinion mining industry, which is still very young [33]. Research has been conducted in semantic analysis to capture the contextual meaning of language, as well as sentiment analysis to capture the overall sentiment of a comment. However, there is still work to be done for opinion mining to identify more complex emotions such as trust and mistrust, as well as to account for external factors such as personal relationships that may influence the opinions of users [34]. Additionally, algorithms need to be developed that can better discern between genuine opinions and automated bots that are designed to spread negative sentiments. The future of opinion mining lies in intelligent systems that can capture the context of conversations, identify complex emotions, and provide real-time insights. Advanced sentiment analysis algorithms and AI-driven platforms will likely be instrumental in improving the accuracy and scalability of opinion mining [35]. Additionally, there is a need for more sophisticated methods to predict and analyze public opinion over time. With the rise of social media, opinion mining has the potential to provide powerful insights to businesses and organizations that will help them better understand their customers and better meet their needs [36].
18 Exploring the potential of artificial
371
References [1]
[2]
[3] [4]
[5]
[6]
[7] [8]
[9] [10]
[11] [12]
[13]
[14] [15]
[16]
[17]
Yadav, S. P., Zaidi, S., Mishra, A., & Yadav, V. (2022). Survey on machine learning in speech emotion recognition and vision systems using a recurrent neural network (RNN). Archives of Computational Methods in Engineering, 29(3), 1753–1770. Khan, S., Saravanan, V. N. G. C., Lakshmi, T. J., Deb, N., & Othman, N. A. (2022). Privacy protection of healthcare data over social networks using machine learning algorithms. Computational Intelligence and Neuroscience, 2022, 1–12. Yadav, S. P. (2021). Emotion recognition model based on facial expressions. Multimedia Tools and Applications, 80(17), 26357–26379. Hewahi, N. M., & Baraka, A. M. (2012). Emotion recognition model based on facial expressions, ethnicity and gender using backpropagation neural network. International Journal of Technology Diffusion (IJTD), 3(1), 33–43. Gopi, B., Logeshwaran, J., Gowri, J., & Kiruthiga, T. (2022). The moment probability and impacts monitoring for electron cloud behavior of electronic computers by using quantum deep learning model. NeuroQuantology, 20(8), 6088–6100. Munagala, N. V. L., Saravanan, V., Almukhtar, F. H., Jhamat, N., Kafi, N., & Khan, S. (2022). Supervised approach to identify autism spectrum neurological disorder via label distribution learning. Computational Intelligence and Neuroscience, 2022, 1–9. Yadav, S. P., Mahato, D. P., & Linh, N. T. D. (Eds.). (2020). Distributed artificial intelligence: A modern approach. Boca Raton: CRC Press. Raja, S., Logeshwaran, J., Venkatasubramanian, S., Jayalakshmi, M., Rajeswari, N., Olaiya, N. G., & Mammo, W. D. (2022). OCHSA: Designing Energy-Efficient Lifetime-Aware Leisure Degree Adaptive Routing Protocol with Optimal Cluster Head Selection for 5G Communication Network Disaster Management. Scientific Programming, 2022. Saravanan, V., & Mohan Raj, V. (2016). Maximizing QoS by cooperative vertical and horizontal handoff for tightly coupled WIMAX/WLAN overlay networks. Cluster Computing, 19(3), 1619–1633. Yuvaraj, N., Karthikeyan, T., & Praghash, K. (2021). An improved task allocation scheme in serverless computing using gray wolf optimization (GWO) based reinforcement learning (RIL) approach. Wireless Personal Communications, 117, 2403–2421. Sumathi, A., & Saravanan, V. (2015). Bandwidth based vertical handoff for tightly coupled wimax/ wlan overlay networks. Journal of Scientific and Industrial Research, 74(10), 560–566. Logeshwaran, J., Malik, J. A., Adhikari, N., Joshi, S. S., & Bishnoi, P. (2022). IoT-TPMS: An innovation development of triangular patient monitoring system using medical internet of things. International Journal of Health Sciences, 6(S5), 9070–9084. Sivaram, M., Porkodi, V., Mohammed, A. S., Manikandan, V., & Yuvaraj, N. (2019). Retransmission DBTMA protocol with fast retransmission strategy to improve the performance of manets. IEEE Access, 7, 85098–85109. Yadav, S. P. (2020). Vision-based detection, tracking, and classification of vehicles. IEIE Transactions on Smart Processing and Computing, 9(6), 427–434. Kousik, N., Natarajan, Y., Raja, R. A., Kallam, S., Patan, R., & Gandomi, A. H. (2021). Improved salient object detection using hybrid Convolution Recurrent Neural Network. Expert Systems with Applications, 166, 114064. Yuvaraj, N., & Vivekanandan, P. (2013, February). An efficient SVM based tumor classification with symmetry non-negative matrix factorization using gene expression data. In 2013 International Conference on Information Communication and Embedded Systems (ICICES) (pp. 761–768). IEEE. Ramesh, G., Logeshwaran, J., & Rajkumar, K. (2022). The smart construction for image preprocessing of mobile robotic systems using neuro fuzzy logical system approach. NeuroQuantology, 20(10), 6354–6367.
372
Senthil Kumar C et al.
[18] Kaur, A., & Gupta, V. (2013). A survey on sentiment analysis and opinion mining techniques. Journal of Emerging Technologies in Web Intelligence, 5(4), 367–371. [19] Osimo, D., & Mureddu, F. (2012). Research challenge on opinion mining and sentiment analysis. Universite de Paris-Sud, Laboratoire LIMSI-CNRS, Bâtiment, 508, 1–13. [20] Vinodhini, G., & Chandrasekaran, R. M. (2012). Sentiment analysis and opinion mining: A survey. International Journal, 2(6), 282–292. [21] Cambria, E., Schuller, B., Xia, Y., & Havasi, C. (2013). New avenues in opinion mining and sentiment analysis. IEEE Intelligent Systems, 28(2), 15–21. [22] Chen, H., & Zimbra, D. (2010). AI and opinion mining. IEEE Intelligent Systems, 25(3), 74–80. [23] Ravi, K., & Ravi, V. (2015). A survey on opinion mining and sentiment analysis: Tasks, approaches and applications. Knowledge-based Systems, 89, 14–46. [24] Alexandridis, G., Varlamis, I., Korovesis, K., Caridakis, G., & Tsantilas, P. (2021). A survey on sentiment analysis and opinion mining in greek social media. Information, 12(8), 331. [25] Othman, M., Hassan, H., Moawad, R., & El-Korany, A. (2014). Opinion mining and sentimental analysis approaches: A survey. Life Science Journal, 11(4), 321–326. [26] Aquino, P. A., López, V. F., Moreno, M. N., Muñoz, M. D., & Rodríguez, S. (2020). Opinion mining system for Twitter sentiment analysis. In Hybrid Artificial Intelligent Systems: 15th International Conference, HAIS 2020, Gijón, Spain, November 11–13, 2020, Proceedings (Vol. 15, pp. 465–476). Springer International Publishing. [27] Liu, B., & Zhang, L. (2012). A survey of opinion mining and sentiment analysis. In: Charu C. Aggarwal, ChengXiang Zhai (Eds.). Mining Text Data (pp. 415–463). Boston, MA: Springer. [28] Keikhosrokiani, P., & Pourya Asl, M. (Eds.). (2022). Handbook of research on opinion mining and text analytics on literary works and social media. United States: IGI Global. [29] Lula, P., & Wójcik, K. (2011). Sentiment analysis of consumer opinions written in Polish. Economics and Management, 16(1), 1286–1291. [30] Navale, G. S., Dudhwala, N., Jadhav, K., Gabda, P., & Vihangam, B. K. (2016). Prediction of stock market using data mining and artificial intelligence. International Journal of Computer Applications, 134(12), 9–11. [31] Pang, B., & Lee, L. (2008). Opinion mining and sentiment analysis. Foundations and Trends® in Information Retrieval, 2(1–2), 1–135. [32] Haddi, E., Liu, X., & Shi, Y. (2013). The role of text pre-processing in sentiment analysis. Procedia Computer Science, 17, 26–32. [33] Kirilenko, A. P., Stepchenkova, S. O., Kim, H., & Li, X. (2018). Automated sentiment analysis in tourism: Comparison of approaches. Journal of Travel Research, 57(8), 1012–1025. [34] Zad, S., Heidari, M., Jones, J. H., & Uzuner, O. (2021, May). A survey on concept-level sentiment analysis techniques of textual data. In 2021 IEEE World AI IoT Congress (AIIoT) (pp. 0285–0291). New York City: IEEE. [35] Banker, S., & Patel, R. (2016). A brief review of sentiment analysis methods. International Journal of Information Sciences and Techniques (IJIST), 6(1/2), 89–95. [36] Grljević, O., Bošnjak, Z., & Kovačević, A. (2022). Opinion mining in higher education: A corpus-based approach. Enterprise Information Systems, 16(5), 1773542.
Pramendra Kumar, Gunawan Widjaja, Nikhil S. Patankar, B. Suresh Kumar
19 A novel blockchain-based artificial intelligence application for healthcare automation Abstract: The integration of blockchain and artificial intelligence (AI) technology can facilitate a comprehensive healthcare automation platform that can leverage the power of both AI and blockchain. This platform not only creates trust, enabling transparency of patient data sharing and record keeping, but also allows AI and machine learning algorithms to improve the efficiency and accuracy of healthcare technologies. It can provide healthcare data privacy, ease the access and sharing of medical records and other confidential health information, and enable fast and secure medical imaging transfers. Keywords: Blockchain, AI, technology, healthcare, automation
19.1 Introduction Artificial intelligence (AI)-driven analytics can be deployed to improve healthcare outcomes, leading to more efficient data processing and diagnosis. In addition, the combination of blockchain and AI can enable a platform to track all health-related data, including medical records, medical devices, laboratory data, medications, and other treatments with full traceability [1]. Furthermore, blockchain technology can create an immutable records system which fully stores medical and health data. This secure data storage system can ensure the privacy, accuracy, and security of medical records as well as protecting data against fraud, theft, and misuse.
19.1.1 Background Blockchain-based AI is an innovative application that has been developed to help automate certain aspects of medical and healthcare services. It combines the power of distributed ledger technology (DLT) with advanced AI capabilities to effectively automate end-to-end processes such as patient care, accurate diagnoses, data analysis and more [2]. By using AI-based systems such as machine learning and natural language processing (NLP), blockchain-based AI can analyze medical data in real time and deliver timely and accurate results [3]. This automated system can access the medical records https://doi.org/10.1515/9783111323749-019
374
Pramendra Kumar et al.
of patients and give them timely diagnoses while allowing medical staff to focus on providing quality healthcare instead of dealing with paperwork. Additionally, blockchain technology further enhances the security and privacy of patient data as encrypted data is stored on distributed ledgers, making it safe from hackers. The application of blockchain-based AI for healthcare automation is still in its early development stages. However, its potential is already being realized in many hospitals and clinics as it is helping to save time, money, and resources. In the near future, these systems will play a major role in revolutionizing healthcare and making it more efficient, secure, and affordable [4].
19.1.2 Objectives The objective of a blockchain-based AI application for healthcare automation is to automate and streamline the administrative processes in the healthcare sector. It can simplify and reduce paperwork, reduce costs, reduce human errors, and boost transparency and security. This technology also offers secure access to high-quality data to use for medical research, analysis, and decision-making. It can help monitor healthcare delivery, such as tracking resources, following up on patients’ treatments, and managing patient health records [5]. Through machine learning algorithms, the technology can recognize patterns, detect fraud and abuse, track medical appointments, and provide predictive analytics. Additionally, it can be used to develop innovative patient-driven care management strategies and healthcare interventions.
19.2 Healthcare automation Healthcare automation is the process of incorporating advanced technology, robotics, and AI into healthcare to improve service delivery and patient care. It includes both the automation of hospital processes and the integration of medical technologies. Automation in healthcare helps reduce the burden on healthcare professionals, makes healthcare more accessible and efficient, and enables better patient care through increased accuracy, improved patient safety, and enhanced data analysis. Automation in healthcare can be applied in a variety of forms, including remote patient monitoring, automated medication dispensing, automated scheduling, automated data entry, and telemedicine [6]. Automation in healthcare can improve the patient experience by freeing up healthcare professionals to focus more on providing customized patient care, streamlining processes and regimes, and providing faster care. It can also help reduce errors and improve efficiency by allowing healthcare professionals to devote more time to data entry and interpretation of patient health information. Addition-
19 A novel blockchain-based artificial intelligence application
375
ally, automation can provide better insights into population health by allowing for more granular analysis and aggregation of patient data.
19.2.1 Importance of healthcare automation Healthcare automation is an important tool for improving the speed, accuracy, and efficiency of medical processes, particularly those related to patient care. It can improve patient outcomes by reducing errors, eliminating unnecessary manual processes, increasing data accuracy, and providing easier access to patient records and medical information [7]. Automation can also help healthcare professionals stay organized and better coordinate patient care, reducing wait times and improving overall patient satisfaction. By increasing the overall efficiency of medical records and processes, healthcare automation can save time and money, reduce human error, and make the overall patient experience more efficient.
19.2.2 Challenges in healthcare automation There are several challenges to healthcare automation, including the cost and complexity of implementation, the need to maintain privacy and security, and the ability to ensure that the technology is reliable. Automation requires a significant investment, as it involves complex hardware and software. Additionally, healthcare systems must be extremely secure, as patient data and medical records are highly sensitive, and any compromise of patient safety and confidentiality could lead to legal action [8]. Finally, healthcare automation must be reliable and efficient, as medical personnel rely on the technology to provide medical care.
19.2.3 Role of AI in healthcare automation AI is a rapidly growing field that can be used to design, develop, and deploy solutions to a variety of real-world problems. AI-based solutions offer a host of benefits, such as automation, improved efficiency, enhanced accuracy, and reduced costs. AI has also found applications in the healthcare industry in recent years [9]. AI-based tools, such as machine learning algorithms, can be used to automate various processes, such as the assessment and diagnosis of medical conditions. AI-powered technologies can also crunch large amounts of medical data in order to suggest more accurate diagnoses and treatments [10]. Furthermore, AI can also be used to facilitate medical research, assist with data management, and provide better patient care. Additionally, AI can be used to analyze the effectiveness of healthcare services and make recommendations for improvement.
376
Pramendra Kumar et al.
In short, AI-based tools can be used to automate mundane tasks, improve the accuracy of medical diagnoses, and identify inefficiencies in the delivery of healthcare services. This technology has the potential to improve the quality of healthcare, reduce costs, and improve patient outcomes.
19.3 Blockchain technology in healthcare Blockchain technology has the potential to revolutionize healthcare by providing improved access to patient data, secured storage of patient records, better patient privacy and security, and more efficient operations. The technology is still in its early stages, but it offers the promise of a secure, distributed system for storing and sharing patient data across multiple organizations and devices, while maintaining the highest level of data protection and privacy [11]. By using a distributed ledger and cryptographic techniques, blockchain can provide faster, more secure access to patient records, eliminating the need for centralized databases. Additionally, it can also help streamline the various processes involved in the healthcare industry, including clinical trials, billing, claims processing, and pharmaceutical supply chains. With blockchain, almost all aspects of the healthcare industry can be digitized, resulting in better patient care and improved operational efficiency.
19.3.1 Introduction to blockchain technology Blockchain technology is a type of DLT that enables the secure storage, transfer, and management of data and digital assets, without the need for third-party intermediaries. It is a decentralized, cryptographic system of potentially permanent records and records that is available for anyone to view. The data is stored on a distributed network of computers, and is secured by networks of computers running the consensus algorithms, which verify the authenticity and accuracy of all the transactions within the network [12]. This DLT has the potential to revolutionize the way financial transactions are processed, and has become the foundation of a new economy. Blockchain also provides a platform for smart contracts, enabling the automation of multiple processes and operations within a company’s infrastructure.
19.3.2 Applications of blockchain in healthcare Blockchain technology is revolutionizing healthcare in remarkable ways. It offers numerous advantages over traditional systems, such as secure data storage, enhanced transparency, and improved accuracy [13].
19 A novel blockchain-based artificial intelligence application
377
Blockchain can be used in healthcare to securely store patient data, create an immutable audit trail of all patient interactions, and streamline the exchange of medical records between disparate systems. The technology can also be used to securely audit claims and reimbursement procedures, manage financial trading and market settlements, and manage clinical trials. Additionally, it can be used for population health analytics, fraud detection, and biomedical research [14]. By cryptographically securing patient data on a decentralized ledger, blockchains enable data to be securely shared between institutions and across different jurisdictions, reducing administrative costs and improving access to care. This also enables providers to have a better understanding of patient healthcare history, leading to more targeted treatments and improved patient outcomes [15]. In addition, blockchain can be used to create digital currency networks that facilitate the transfer of value within healthcare systems. This could significantly improve the efficiency of healthcare payments and create new opportunities for providers to offer innovative health services. Ultimately, blockchain technology will help healthcare systems reduce costs, improve quality, and create an environment of trust.
19.3.3 Benefits and challenges of blockchain in healthcare Benefits of blockchain in healthcare: – Improved security and privacy: Blockchain technology offers secure, reliable, and tamper-proof processing of patient medical information. It also provides robust authentication and encryption protocols for patient data. Encryption and authentication reduce the risks of stolen or corrupted data. – Improved data integrity and accuracy: Blockchain technology ensures that the patient data is accurate and up to date. The data cannot be altered or changed without permission, and the entire process is highly transparent. This reduces the chances of errors and ensures that the data remains reliable. – Reduction of manual processes: Blockchain technology eliminates the need for a central authority to manage health records. It also improves the speed of data transaction, and of course, since manual labor is not required. – Cost savings: By eliminating tedious manual processes, reducing redundancy, and eliminating third-party intermediaries, blockchain technology can help reduce operational costs for healthcare organizations. Challenges of blockchain in healthcare: – Scalability: As is common with new technologies, scalability is an issue. As the technology becomes more popular, the ability of the system to handle a larger volume of data is a challenge. – Regulations: Blockchain technology is a revolutionary one, and the implementation of it in healthcare is under the scrutiny of regulatory authorities. Stringent
378
–
–
Pramendra Kumar et al.
rules and regulations that must be followed are slowing down the adoption of blockchain technology in healthcare. Lack of knowledge and expertise: Even though blockchain technology is becoming increasingly popular, lack of knowledge and expertise is a challenge. There is a growing demand for experienced personnel to help implement and manage healthcare applications using blockchain technology. Costs Involved: Even though blockchain technology can help reduce the cost of some processes, the technology is pricey. Implementing and managing blockchain technology requires the purchase of specialized hardware and expert services, which can be costly.
19.4 Artificial intelligence techniques for healthcare automation AI has the potential to revolutionize the healthcare industry, allowing for more accurate diagnoses and improved treatment. AI techniques in healthcare automation include: – Machine learning: Machine learning algorithms are used to identify patterns in large amounts of data, allowing healthcare professionals to compare them to existing knowledge and make better decisions. – NLP: NLP algorithms and tools can be used to understand and interpret natural language, allowing healthcare providers to better interact with patients and process information. – Image recognition: AI-enabled image recognition tools can be used to identify medical images, such as medical radiology scans, allowing healthcare professionals to detect abnormalities faster and with less effort. – Expert systems: Expert systems are AI-driven solutions that use pre-programmed knowledge to provide healthcare professionals with advice in specific medical scenarios. They can be used for diagnosis, treatment planning, and more. These are just some of the AI techniques being used in healthcare automation today. With continued advances in technology, AI solutions are becoming increasingly sophisticated and advanced, allowing healthcare professionals to provide better care for their patients.
19.4.1 Machine learning in healthcare automation Machine learning in healthcare automation refers to a set of technologies and algorithms designed to automate certain healthcare processes and activities. Machine
19 A novel blockchain-based artificial intelligence application
379
learning can be used to automate administrative tasks and make informed decisions about patient care and treatment [16]. For example, machine learning algorithms can be used to identify patterns in doctor–patient interactions, to create personalized healthcare plans, and to detect anomalies in medical images. Machine learning also enables healthcare providers to use vast amounts of data from patient records, lab results, applications, and electronic health records (EHRs) to make decisions about diagnosis and treatment. By leveraging machine learning, healthcare providers can make more accurate decisions at a faster pace and reduce costs by streamlining processes. Additionally, machine learning can be used to create predictive models to anticipate the outcome of treatments and give insights on patient health [17].
19.4.2 NLP in healthcare automation NLP in healthcare automation is the use of machine-learning algorithms to automate the analysis and interpretation of natural language, including medical terms and patient records. NLP technology is being applied to healthcare automation to leverage the vast amount of health information that is currently only accessible to clinical staff. It is enabling clinicians to more quickly assess, diagnose, and recommend treatments for patients without having to manually review each record [18]. NLP is also being used to automate administrative and operational tasks such as EHR management, billing, and medical coding. Additionally, NLP is being used to generate reports and insights from health data to improve patient outcomes and drive efficiencies. For example, NLP algorithms can be used to analyze data from pharmacy, lab, and other clinical reporting systems to detect adverse events and prevent readmissions. The NLP is being used to develop chatbots that can assist patients with health inquiries and provide virtual assistants that can recommend patients for additional care [19].
19.4.3 Computer vision in healthcare automation Computer vision in healthcare automation is the use of AI algorithms and software to recognize and interpret medical images in order to automate processes that would otherwise be done manually. Specifically, this involves the use of Machine learning algorithms to analyze and optimize medical images such as X-rays, CT scans, and MRI scans in order to detect and classify various anomalies and diseases [20]. This technology can be used to automate medical diagnosis and save time, significantly reducing the amount of manual labor required for healthcare professionals. The use of computer vision is increasingly being implemented in hospitals, clinics, and hospitals, as well as in mobile and home healthcare applications. By automating routine tasks, healthcare professionals can focus their energy and attention on more complex tasks, as well as increase the accuracy of diagnoses [21].
380
Pramendra Kumar et al.
19.5 Integration of blockchain and artificial intelligence The integration of blockchain and AI is a powerful combination that can lead to innovative solutions in a multitude of industries, such as finance, healthcare, IoT, and more. Blockchain is a DLT that facilitates the secure exchange of digital assets, and AI is the development of intelligent systems that can think, understand, and take actions based on their own data or environment. Together, these two technologies provide an unprecedented level of accountability, security, transparency, and trust [22]. Through integrating blockchain and AI, users can leverage the features of both systems to store, share, and secure user data while also enabling intelligent autonomous decision-making tools. Examples of AI-based technologies include machine learning, NLP, image recognition, and more. These AI capabilities would be more secure thanks to decentralized blockchain networks, in which shared data gets stored and distributed on a global scale [23]. For instance, AI-based facial recognition systems could use blockchain-integrated databases to store user data for secure authentication across multiple networks. This AI-blockchain combination could also be valuable in the healthcare industry, allowing for secure sharing and storage of patient data and securely automated drug monitoring systems. Additionally, by combining the blockchain and AI, users can create more reliable predictive analytics systems that are resistant to manipulation through tampering with the data [24]. The integration of blockchain and AI has the potential to revolutionize various industries, making them more secure, efficient, and transparent. Blockchain provides a secure and reliable platform for making transactions, while AI can be used to automate process and decisions, resulting in more efficient workflows. Together, they can enable new use cases that aren’t currently possible using either technology alone [25].
19.5.1 Overview of the integration Integration is the process of combining separate, distinct elements into a cohesive whole. It can be used in multiple contexts, including technology, mathematics, and business. For example, technology-based integration can include the integration of different systems or components. This could involve connecting a database to an application, or making sure two systems share a common set of protocols [26]. In mathematics, integration is used to calculate the area under a curve by adding up the infinitesimal slices of area. And in business, integration often relates to companies joining forces, either through a merger or other assimilations. The core goal of any integration effort is to create a unified system from disparate parts.
19 A novel blockchain-based artificial intelligence application
381
19.5.2 Blockchain-based data sharing and security Blockchain-based data sharing and security is a secure way to send and receive data leveraging the power of blockchain technology. It creates an immutable, distributed ledger of transactions that all users can trust and verify. With this technology all participants of the network agree on a shared version of the data, eliminating the need for a central authority or a centralized storage [27]. Data is encrypted and stored in a distributed ledger, where it is shared and updated in real time between all participants in the network. This makes it much harder for data to be compromised or altered. Additionally, this type of data sharing ensures better data privacy by using public-key cryptography to encrypt data before it’s sent. With blockchain technology, data can be shared securely without relying on a single source of truth.
19.5.3 Smart contracts for automated healthcare processes A smart contract for healthcare is an automated and digital agreement between parties, designed to facilitate the execution of healthcare-related transactions. The purpose of these contracts is to enable improved efficiency, cost-effectiveness, accuracy, trust, and security in the healthcare industry [28]. Smart contracts can be used for a variety of healthcare activities, such as payment processing, insurance claims, and record keeping. By automating these processes, they reduce the need for face-to-face transactions and eliminate the potential for paper-based errors. By securely storing healthcare data, smart contracts help to prevent data loss. Furthermore, they allow for instant payment processing, secure archiving of patient records, and improved accuracy through automated healthcare analytics [29]. Smart contracts also provide enhanced security and privacy measures that are not possible using traditional healthcare models. Through encryption of sensitive data, improved authentication protocols, and improved access control, smart contracts minimize the risk of unauthorized access to confidential information. Furthermore, these contracts are designed to provide improved accountability and transparency in the healthcare industry, allowing all stakeholders to feel secure in the knowledge that their data is safe and secure [30].
19.5.4 Decentralized AI models and federated learning Decentralized AI models involve the usage of distributed ledgers and AI to enable lower latency transactions or data manipulation. These systems are considered to be more secure and efficient than traditional systems, as they are more dispersed [31].
382
Pramendra Kumar et al.
Federated learning is a machine learning approach that allows models to be trained on user data without the data ever leaving the user’s device. This allows companies to train a model without having to collect individual user data points. Additionally, due to the decentralized nature of the training process, federated learning models are able to achieve faster training times and higher levels of privacy and security [32].
19.6 Blockchain-based AI applications in healthcare Blockchain technology has been used to create healthcare applications that can help to provide more efficient and secure patient records, enable faster payment and reimbursement of healthcare services, improve clinical trial transparency, and share secure and reliable medical data. This technology has been used to develop AI-based applications, including algorithmic chatbots for monitoring health data, developing diagnostics systems, administering personalized health coaching, and providing other services [33]. The use of blockchain and AI technologies can help save time and money, while also improving the overall quality of care. Patient data can be securely stored on a blockchain – and selectively shared among healthcare providers – giving them a structured framework to grant access to and securely transmit patient information. This could also eliminate cumbersome paperwork and improve the speed of clinical decision-making, communication, and payment [34]. In addition, blockchain-enabled AI tools can provide timely and accurate healthcare recommendations based on patient data. This data could be used to facilitate research and discovery of more effective treatments, as well as to optimize treatment plans for individual patients. AI tools can also be used to alert healthcare providers to potential life-threatening conditions, as well as potential drug interactions. This technology is still in the early stages of development, but its potential for delivering improved healthcare services is undeniable [35].
19.6.1 Electronic health record management An EHR is a digital version of a patient’s medical history that is maintained by the provider over time. It contains information gathered from all patient encounters, including past medical history, diagnoses, medications, treatment plans, allergies, immunization status, lab results, and radiology images. EHRs offer a wide range of benefits to patients and providers, such as increased accessibility, improved care coordination, and easy data sharing [36]. However, this also means that EHRs need to be securely managed and maintained. EHRs are often maintained by healthcare IT pro-
19 A novel blockchain-based artificial intelligence application
383
fessionals – such as health information managers – who are responsible for the accurate collection, maintenance, storage, security, and sharing of patient data. These professionals must ensure that all data is accurately entered and updated, and must also adhere to all applicable HIPAA regulations. They also need to provide additional support to providers who are using EHRs, such as helping them learn how to enter data and interpret reports.
19.6.2 Medical supply chain tracking The medical supply chain tracking is a process of tracking the movement of medical products and services from the source of supply to the point of care. It involves the tracking of both physical goods and digital transactions. This process helps to ensure that the right product is delivered to the right patient at the right time, and that the inventory of each medical supplier is kept up to date. The medical supply chain tracking process starts at the supplier, tracking the delivery of products from the vendor to the doctor’s office, hospital, or other care provider. The tracking also includes the dispensing of the product or service, the receipt of payment, and the disposal or return of the product or service. Depending on the type of product or service being tracked, the tracking system may include additional information such as patient prepayments, orders, credit, or authorization codes. The tracking system can provide useful data for analyzing performance, identifying trends, and improving patient care. This data can be used to help identify and mitigate supply chain risks, such as stock levels and delivery times, as well as to ensure compliance with regulatory requirements and government standards. In addition, tracking systems can be used to automate functions such as inventory tracking and ordering processes, thereby reducing paperwork and improving efficiency.
References [1]
[2]
[3] [4]
Allen, B., Dreyer, K., Stibolt, R., Jr, Agarwal, S., Coombs, L., Treml, C., . . . & Wald, C. (2021). Evaluation and real-world performance monitoring of artificial intelligence models in clinical practice: Try it, buy it, check it. Journal of the American College of Radiology, 18(11), 1489–1496. Tabbussum, R., & Dar, A. Q. (2021). Performance evaluation of artificial intelligence paradigms – artificial neural networks, fuzzy logic, and adaptive neuro-fuzzy inference system for flood prediction. Environmental Science and Pollution Research, 28(20), 25265–25282. Hicks, S. A., Strümke, I., Thambawita, V., Hammou, M., Riegler, M. A., Halvorsen, P., & Parasa, S. (2022). On evaluation metrics for medical applications of artificial intelligence. Scientific Reports, 12(1), 5979. Li, L., Qin, L., Xu, Z., Yin, Y., Wang, X., Kong, B., . . . & Xia, J. (2020). Using artificial intelligence to detect COVID-19 and community-acquired pneumonia based on pulmonary CT: Evaluation of the diagnostic accuracy. Radiology, 296(2), E65–E71.
384
[5]
[6]
[7]
[8]
[9]
[10]
[11]
[12]
[13] [14]
[15] [16] [17]
[18]
[19] [20]
[21]
Pramendra Kumar et al.
Dumitrascu, O., Dumitrascu, M., & Dobrotǎ, D. (2020). Performance evaluation for a sustainable supply chain management system in the automotive industry using artificial intelligence. Processes, 8(11), 1384. Ghahramani, M., Qiao, Y., Zhou, M. C., O’Hagan, A., & Sweeney, J. (2020). AI-based modeling and data-driven evaluation for smart manufacturing processes. IEEE/CAA Journal of Automatica Sinica, 7(4), 1026–1037. Khanagar, S. B., Al-Ehaideb, A., Maganur, P. C., Vishwanathaiah, S., Patil, S., Baeshen, H. A., . . . & Bhandi, S. (2021). Developments, application, and performance of artificial intelligence in dentistry – A systematic review. Journal of Dental Sciences, 16(1), 508–522. Wu, E., Wu, K., Daneshjou, R., Ouyang, D., Ho, D. E., & Zou, J. (2021). How medical AI devices are evaluated: Limitations and recommendations from an analysis of FDA approvals. Nature Medicine, 27(4), 582–584. Amin, M. M., Cambria, E., & Schuller, B. W. (2023). Will affective computing emerge from foundation models and general AI? A first evaluation on chatgpt. In 2023 International Conference on computation and Language (pp. 761–768). IEEE, Singapore. Forghani, R., Savadjiev, P., Chatterjee, A., Muthukrishnan, N., Reinhold, C., & Forghani, B. (2019). Radiomics and artificial intelligence for biomarker and prediction model development in oncology. Computational and Structural Biotechnology Journal, 17, 995. Tobler, P., Cyriac, J., Kovacs, B. K., Hofmann, V., Sexauer, R., Paciolla, F., . . . & Hirschmann, A. (2021). AI-based detection and classification of distal radius fractures using low-effort data labeling: Evaluation of applicability and effect of training set size. European Radiology, 31, 6816–6824. AI-Atroshi, C., Rene Beulah, J., Singamaneni, K. K., Pretty Diana Cyril, C., Neelakandan, S., & Velmurugan, S. (2022). Automated speech based evaluation of mild cognitive impairment and Alzheimer’s disease detection using with deep belief network model. International Journal of Healthcare Management, 2022, 1–11. Gunning, D., & Aha, D. (2019). DARPA’s explainable artificial intelligence (XAI) program. AI Magazine, 40(2), 44–58. Zou, Y., Zheng, C., Alzahrani, A. M., Ahmad, W., Ahmad, A., Mohamed, A. M., . . . & Elattar, S. (2022). Evaluation of artificial intelligence methods to estimate the compressive strength of geopolymers. Gels, 8(5), 271. Yadav, S. P., & Yadav, S. (2020). Image fusion using hybrid methods in multimodality medical images. Medical and Biological Engineering and Computing, 58, 669–687. Irfan, D., Tang, X., Narayan, V., Mall, P. K., Srivastava, S., & Saravanan, V. (2022). Prediction of quality food sale in mart using the AI-based TOR method. Journal of Food Quality, 2022, 1–13. Yuvaraj, N., Srihari, K., Dhiman, G., Somasundaram, K., Sharma, A., Rajeskannan, S. M. G. S. M. A., . . . & Masud, M. (2021). Nature-inspired-based approach for automated cyberbullying classification on multimedia social networking. Mathematical Problems in Engineering, 2021, 1–12. Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). Energy‐efficient resource allocation model for device‐to‐device communication in 5G wireless personal area networks. International Journal of Communication Systems, 2023, 1–14. Yadav, S. P., Agrawal, K. K., Bhati, B. S., Al-Turjman, F., & Mostarda, L. (2022). Blockchain-based cryptocurrency regulation: An overview. Computational Economics, 59(4), 1659–1675. Rajalakshmi, M., Saravanan, V., Arunprasad, V., Romero, C. T., Khalaf, O. I., & Karthik, C. (2022). Machine learning for modeling and control of industrial clarifier process. Intelligent Automation and Soft Computing, 32(1), 339–359. Manikandan, V., & Yuvaraj, N. (2019, May). Advanced expert system using particle swarm optimization based adaptive network based fuzzy inference system to diagnose the physical constitution of human body. In Emerging Technologies in Computer Engineering: Microservices in
19 A novel blockchain-based artificial intelligence application
[22] [23]
[24]
[25]
[26]
[27] [28]
[29]
[30]
[31]
[32]
[33]
[34]
[35]
[36]
385
Big Data Analytics: Second International Conference, ICETCE 2019, Jaipur, India, February 1–2, 2019, Revised Selected Papers (Vol. 985, p. 349). New York: Springer. Sangeetha, R., Logeshwaran, J., Rocher, J., & Lloret, J. An improved agro deep learning model for detection of Panama wilts disease in banana leaves. AgriEngineering. 2023; 5(2):660–679 Yadav, S. P., Zaidi, S., Mishra, A., & Yadav, V. (2022). Survey on machine learning in speech emotion recognition and vision systems using a recurrent neural network (RNN). Archives of Computational Methods in Engineering, 29(3), 1753–1770. Saravanan, V., & Neeraja, A. (2013, January). Security issues in computer networks and stegnography. In 2013 7th International Conference on Intelligent Systems and Control (ISCO) (pp. 363–366). IEEE. Gobinathan, B., Mukunthan, M. A., Surendran, S., Somasundaram, K., Moeed, S. A., Niranjan, P., . . . & Sundramurthy, V. P. (2021). A novel method to solve real time security issues in software industry using advanced cryptographic techniques. Scientific Programming, 2021, 1–9. Logeshwaran, J., Kiruthiga, T., & Lloret, J., (2023). A novel architecture of intelligent decision model for efficient resource allocation in 5G broadband communication networks. ICTACT Journal on Soft Computing, 13 (3), 2986–2994 Yadav, S. P., Mahato, D. P., & Linh, N. T. D. (Eds.). (2020). Distributed artificial intelligence: A modern approach. Boca Raton: CRC Press. Singh, J., Deepika, J., Sathyendra Bhat, J., Kumararaja, V., Vikram, R., Jegathesh Amalraj, J., . . . & Sakthivel, S. (2022). Energy-efficient clustering and routing algorithm using hybrid fuzzy with grey wolf optimization in wireless sensor networks. Security and Communication Networks, 2022. Natarajan, Y., Kannan, S., & Mohanty, S. N. (2021). Survey of various statistical numerical and machine learning ontological models on infectious disease ontology. Data Analytics in Bioinformatics: A Machine Learning Perspective, 14(1), 431–442. Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). L-RUBI: An efficient load-based resource utilization algorithm for bi-partite scatternet in wireless personal area networks. International Journal of Communication Systems, 36(4), 1–17. Wan, Y., Tong, Y., Liu, Y., Huang, Y., Yao, G., Chen, D. Q., & Liu, B. (2022). Evaluation of the combination of artificial intelligence and radiologist assessments to interpret malignant architectural distortion on mammography. Frontiers in Oncology, 12, 1739. Hosamo, H. H., Nielsen, H. K., Kraniotis, D., Svennevig, P. R., & Svidt, K. (2023). Digital Twin framework for automated fault source detection and prediction for comfort performance evaluation of existing non-residential Norwegian buildings. Energy and Buildings, 281, 112732. Lee, D. H., Woo, S. E., Jung, M. W., & Heo, T. Y. (2022). Evaluation of odor prediction model performance and variable importance according to various missing imputation methods. Applied Sciences, 12(6), 2826. Markus, A. F., Kors, J. A., & Rijnbeek, P. R. (2021). The role of explainability in creating trustworthy artificial intelligence for health care: A comprehensive survey of the terminology, design choices, and evaluation strategies. Journal of Biomedical Informatics, 113, 103655. Orhan, K., Bayrakdar, I. S., Ezhov, M., Kravtsov, A., & Özyürek, T. A. H. A. (2020). Evaluation of artificial intelligence for detecting periapical pathosis on cone‐beam computed tomography scans. International Endodontic Journal, 53(5), 680–689. Schwartz, R., Dodge, J., Smith, N. A., & Etzioni, O. (2020). Green ai. Communications of the ACM, 63(12), 54–63.
Soumya Mishra, Priyanka Sharma, Bhoopathy Varadharajan, Bhola Khan
20 Enhancing industrial efficiency with AI-enabled blockchain-based solutions Abstract: The ever-evolving technologies are constantly making the industrial processes more efficient and providing new opportunities for cost saving and enhanced operations. Artificial intelligence (AI)-enabled blockchain-based solutions are making significant contributions to the enhancement of industrial efficiency. AI-enabled blockchain-based solutions enable the secure and frictionless exchange of digital assets and data among members of the chain while enabling trustless interaction between them. These solutions enable faster and error-free transactions with enhanced scalability, reliability, traceability, and privacy. Keywords: AI, blockchain, industrial, efficiency, privacy
20.1 Introduction Artificial intelligence (AI)-enabled blockchain-based solutions can enable automatic tracking of products through the supply chain, which helps reduce time and cost while ensuring the accuracy of data in real time. These solutions also make data sharing more secure, as they enable the secure storage of transaction information which can be utilized for data analytics in the future [1]. Additionally, AI-enabled blockchain-based solutions can be utilized to identify and analyze potential fraudulent activities within the chain. This helps in reducing the risks associated with fraudulent activities, and increases safety and security. AI-enabled blockchain-based solutions are valuable tools that are making industrial processes more efficient.
20.1.1 Background and objectives The objective of enhancing industrial efficiency with AI-enabled blockchain-based solutions is to create a secure and efficient digital industrial ecosystem in order to drive economic growth. This objective requires the implementation of a blockchain-based platform that will link industrial businesses with each other and enable more efficient communication and collaboration [2]. Meanwhile, AI-enabled technology can be integrated into blockchain technology to make it even more powerful and efficient. With AI, blockchain technology can create a system that is intelligent, secure, and costeffective. Ultimately, these combined technological advancements will help to create https://doi.org/10.1515/9783111323749-020
388
Soumya Mishra et al.
smarter and more efficient industrial operations and processes, leading to improved productivity and cost savings for businesses.
20.2 Industrial efficiency: challenges and opportunities Industrial efficiency is a measure of the amount of production a business achieves with a given level of resources. This measure can be applied to specific processes, pieces of machinery, production lines, or even entire factories or industries. Efficiency can be improved by streamlining processes, reducing overhead costs, improving employee productivity, and other measures [3]. It is an important measure for any business, as it is directly related to profits.
20.2.1 Overview of industrial efficiency The industrial efficiency overview refers to the systematic assessment of the industrial production process. It encompasses operational efficiency, capital efficiency, utilization efficiency, inventory efficiency, energy efficiency, process efficiency and product quality. A thorough overview of industrial efficiency is beneficial in reducing production costs, advancing productivity, and enhancing the competitiveness of the industry. The industrial efficiency measures also help companies to utilize resources more efficiently and to reduce the carbon footprint in production process. The overview includes an assessment of the entire production process from raw material inputs to end-products and services delivery [4, 5]. It assesses and monitors operational efficiency, capital efficiency, utilization efficiency, inventory efficiency, and energy efficiency. It further examines process performance, product quality, and customer satisfaction measures. The data collected helps to enhance industry operations and encourage corporate sociocultural and environmental benefits.
20.2.2 Challenges in industrial efficiency Industrial efficiency refers to the use of systems, machines, and technologies that reduce inflation and improve outputs. This helps companies cut costs, become more efficient, and maximize productivity [6]. However, like any other business process, industrial efficiency can be quite challenging. Some of the challenges include:
20 Enhancing industrial efficiency with AI-enabled blockchain-based solutions
–
–
–
–
–
389
Economic volatility: Industrial efficiency is particularly vulnerable to changes in global economic climates. As resources become costlier, companies have to figure out ways to become more efficient without increasing costs. Restrictive regulations: Strict regulations often create barriers to improved industrial efficiency. This could require a company to invest heavily in new technologies or equipment that is incompatible with existing systems. Technological complexity: Industrial efficiency is made difficult by the technical complexity of operating systems and equipment. Companies often need to invest time and money into training the workforce on the best practices for operating the systems. Skilled labor: In spite of technological advances, the need for skilled labor is still present in many industries. Without the right personnel in place, industrial efficiency initiatives can become very difficult to implement. Resource scarcity: Lack of resources further complicates the challenge of industrial efficiency. A lack of access to the right tools, equipment, and resources can prevent a company from reaching its potential.
20.2.3 Opportunities for improvement Opportunities for improvement can involve any aspect of a business or organization – from its products to its processes to its people. Improving on a few key elements can lead to competitive advantage, higher customer satisfaction, and improved overall performance [7]. Common opportunities for improvement include improving quality, product maintenance, product innovation, customer service, cost control, operational efficiency, and employee development.
20.3 Blockchain technology in industrial applications Blockchain technology is rapidly gaining traction in industrial applications. By providing an immutable, distributed ledger of data, blockchain technology can securely store data and transfer ownership of assets without third-party involvement. This technology is being used to facilitate efficient, secure transactions in a range of industries such as finance, healthcare, logistics, and automotive [8]. In the automotive industry, for example, blockchain can be used in a variety of ways to improve processes. It can be used to ensure accurate tracking of parts and components, and to securely store information on the history of a vehicle, such as its previous owners and maintenance records. Blockchain can also be used to verify customer
390
Soumya Mishra et al.
IDs and ensure better supply chain traceability through smart contracts, streamlining the process of sharing and managing data in the automotive sector [9]. In the logistics industry, blockchain technology can be used to create efficient, automated systems for verifying asset location, tracking, and transfer of ownership. This technology can be used to facilitate secure transactions, such as payments, with the assurance that digital assets can be transferred safely and securely. Additionally, blockchain can also enable faster and more secure delivery of goods and services, as well as improve the tracking and verification of shipments. Blockchain technology also has applications in healthcare, such as with its ability to securely store and manage patient data. Through blockchain, data can be shared with multiple parties safely and securely, while ensuring regulatory compliance obligations are met. This technology can also be used to create digital identities for patients, allowing them to manage their healthcare records securely and efficiently [10]. The blockchain technology is revolutionizing many different industries, from finance to automotive to logistics and healthcare. It offers a secure, reliable, and efficient platform for managing digital assets, creating digital identities, and tracking data. As the technology continues to evolve and gain traction, the potential for its use in industrial applications is growing.
20.3.1 Applications of blockchain in industrial settings Blockchain technology has a vast number of potential applications in industrial settings. Blockchain is a highly secure and decentralized distributed ledger system, which can facilitate digital interactions and automated trust. This makes it an excellent tool for simplifying and improving many aspects of the industrial supply chain, such as tracking, authentication, and payment processing. It can also be used to improve communications, optimize processes, and reduce costs [11]. One of the major industrial applications of blockchain technology is in the supply chain. By using blockchain, the entire supply chain can be digitized and recorded in a secure and tamper-proof ledger system. This allows for better transparency and traceability throughout the chain, reducing errors and fraud. In addition, smart contracts can be used to automate several parts of the supply chain, allowing for faster and more efficient transactions. The industrial sector can also make use of blockchain for better credit and identity management. Automated identity verifications and asset transfers can all be made more secure and efficient with the help of this technology. In addition, blockchain also makes it easier to store and verify production data such as quality control, inventory, testing, and more [12]. Similar applications in the industrial sector include the tracking of goods and materials with IoT sensors, automated contract processing, secure and transparent asset management, improved data security procedures, and more. With the help of block-
20 Enhancing industrial efficiency with AI-enabled blockchain-based solutions
391
chain, industrial organizations can work more efficiently and smarter than ever before.
20.3.2 Benefits and challenges of blockchain in industrial efficiency Benefits: – Increased transparency: Blockchain technology provides a level of transparency that is not seen in traditional recordkeeping systems. This could help reduce fraud and ensure that operations are more efficient and accurate. – Enhanced security: Blockchain technology is designed to be secure and cannot be modified or changed, as each transaction is linked to the previous one and cryptographically secured. This can help reduce the risk of data breaches and attacks from hackers. – Automation: Blockchain technology can be used to automate many industrial processes, reducing the need for manual labor and streamlining operations. – Data integrity: The immutability of blockchain technology helps ensure that data is accurate and consistent across multiple nodes. This can help reduce errors and improve productivity. Challenges: – Complexity: Blockchain technology is relatively new and still evolving, which can make it difficult for some organizations to implement. Moreover, there are still issues concerning scalability, interoperability, and data privacy that must be addressed. – Cost: Implementing blockchain technology can require significant investment in infrastructure and resources. This could be a barrier for some companies that are not willing or able to invest in this technology. – Regulation: As blockchain technology is still relatively new, there is a lack of regulatory framework in many jurisdictions. This can hamper the development and adoption of blockchain technology, as organizations may be unwilling to take risks without proper guidance.
20.4 Artificial intelligence for industrial efficiency AI for industrial efficiency refers to a variety of technologies that can help industrial companies and factories to be more efficient, productive, and profitable. AI can support a variety of industrial processes, including analytics, automation, digital twins, predictive maintenance, and more [13]. AI technologies can be used to make decisions
392
Soumya Mishra et al.
for production, design, maintenance, and other processes and to detect defects. AI can also be used to monitor the operational environment, provide recommendations for process improvement, and automate tasks to reduce human labor. AI applications can help industrial companies save time, money, and resources in production, while keeping quality at a high level. This can enable companies to be more agile and adaptive and respond more quickly to changing needs and opportunities.
20.4.1 Machine learning in industrial efficiency Machine learning in industrial efficiency includes the use of AI and machine learning algorithms to automate the processes, tasks, workflows, and operational decisions that make up traditional industries. By offering both automated decision-making capabilities and instances of valuable insights on how to improve operations, machine learning helps companies operate more efficiently and effectively [14]. Industrial applications, such as logistics and supply chain management, quality assurance, maintenance scheduling, production optimization, safety, and predictive analytics, can be improved through the application of machine learning. In addition to algorithmic decision-making, machine learning can also be used to pre-process, analyze, and interpret data from sensors, machines, and other devices throughout a facility. This data can then be used to create digital twins of real assets, create probabilistic models about uncertain future events and conditions, and provide predictive insights about potential maintenance needs. For example, machine learning can be used to forecast future asset performance, alert operators of potential maintenance needs before they occur, and provide advice on the best course of action for the situation [15]. It can also be used to improve efficiency across multiple machines by suggesting adjustments to control parameters that achieve maximum yields or minimum energy consumption, for example. Ultimately, machine learning can help organizations reduce costs, reduce downtime, increase yield, optimize production, and increase overall safety.
20.4.2 Robotics and automation in industrial settings Robotics and automation are becoming increasingly important in industrial settings. Robotics and automation allow industrial processes to be more efficient and effective, as they automate repetitive and labor-intensive tasks. In industrial settings, robotics and automation are used to increase production speeds, reduce labor costs, and maintain quality levels. They are also used to reduce or eliminate some of the more dangerous and unpleasant aspects of manufacturing [16].
20 Enhancing industrial efficiency with AI-enabled blockchain-based solutions
393
Robots are the most common form of industrial automation. They are computercontrolled, programmable machines that can be programmed to perform a wide range of tasks. These tasks can range from highly complicated tasks such as welding and painting to simpler tasks such as picking and packing. Robots are often used in manufacturing to perform repetitive assembly tasks that would otherwise require a large number of manpower. Automation systems are also used in industrial settings. These systems are networks of machines and software that can be programmed to perform multiple tasks. Automation systems can be used for data processing, material movement, quality control, and production scheduling. Automation systems can also be used for controlling complex systems such as production lines and assembly operations [17]. The robotics and automation are becoming increasingly important in industrial settings. They provide companies with the ability to increase production rates, reduce labor costs, and maintain quality standards. With the right implementation, robotics and automation can greatly improve the efficiency and effectiveness of a manufacturing operation.
20.4.3 Predictive maintenance and fault detection Predictive maintenance and fault detection is a type of preventive maintenance that uses in-depth pattern recognition and analytics to predict and detect when a piece of equipment or system may fail or require maintenance before it actually fails. Predictive maintenance and fault detection uses a variety of methods and technologies such as data collection, sensors, machine learning, analytics, and AI to accurately predict when an equipment or system will start to experience a fault and allow the user to take corrective action [18]. This helps to reduce downtime, improve safety, and decrease maintenance costs. This type of maintenance is used most commonly in industrial settings and in the oil and gas, energy, and other industries.
20.5 Integration of blockchain and AI for industrial efficiency Blockchain and AI can be integrated to industrial efficiency by leveraging the concept of “smart contracts” which can automate the various required tasks to be carried out in the production process. Blockchain can provide a secure, transparent, and flexible platform for the data to be stored in, and AI can use the data to make predictions about product demand, forecast inventory needs, and optimize production sequences to increase efficiency and reduce production costs. Blockchain could also be used to track the data provided by AI-enabled systems, so as to ensure that it is accurate and
394
Soumya Mishra et al.
up-to-date. This would help in reducing fraud and ensuring the efficient utilization of resources. Furthermore, blockchain can also be employed to monitor energy usage and enforce energy consumption rules which have been sanctioned by the government authorities [19]. In addition, it can also provide an immutable ledger to record the transactions that take place between different entities in the supply chain. This would allow for greater accountability, appropriate pricing mechanisms, and secure data sharing.
20.5.1 Overview of the integration Integration is the process of combining different components or subsystems into a larger system in order to achieve a desired objective or goal. This process involves the interchange of data, the flow of information, the crossing of functional boundaries, the implementation of protocols, and the exchange of services. The process of integration is closely related to interoperability and application integration [20]. By leveraging integration, businesses can gain a new competitive edge by optimizing processes, boosting productivity, and leveraging resources. This provides a number of costsaving benefits. Additionally, businesses can become more agile and remain ahead of the competition.
20.5.2 Blockchain-based supply chain management Blockchain-based supply chain management is the use of blockchain technologies to manage and monitor the activities within a supply chain. By leveraging blockchain technology, businesses have the potential to promote transparency and reduce their overhead costs. Additionally, blockchain-based supply chain management can better protect goods from counterfeiting, reduce delivery times, and improve operational efficiency across the supply chain network [21].
20.5.3 AI-enabled smart contracts and autonomous systems AI-enabled smart contracts and autonomous systems are blockchain-based systems, in which the entire transaction process is automated using smart contracts and AIbased technologies. These contracts represent a set of terms and conditions that are written into a code that can be deployed onto a blockchain platform. Smart contracts use complex algorithms and AI to manage and execute the transactions and verifications that need to be conducted in the process, eliminating the need for manual entry or verification. Autonomous systems leverage AI technologies such as natural language processing and machine learning that are used for data analysis, decision-
20 Enhancing industrial efficiency with AI-enabled blockchain-based solutions
395
making, and process automation [22]. Autonomous systems are self-governing and can operate without any human intervention. They are designed to ensure accuracy, reliability, and speed of the contractual process. AI-enabled smart contracts and autonomous systems are revolutionizing the way we manage and execute contracts and transactions. They provide a secure environment and help reduce time, costs, and operational risks.
20.5.4 Data integrity and security in industrial processes Data integrity and security in industrial processes is a set of processes and technologies that are needed to ensure the safe and secure operation of industrial process systems. By implementing data and system security measures, organizations can ensure that the data that is collected and stored is not compromised or inappropriately accessed. This is done by employing techniques such as encryption, firewalls, user authentication, and access control [23]. This data security layer helps protect against unauthorized access and malicious attacks. Keeping data integrity also means that data is free from unauthorized use, alteration, and destruction. Data integrity is maintained through controls such as user authentication, data encryption, data backup, and storage policies. Additionally, companies must have a continuous process in place to keep data secure and to monitor the security of their data systems.
20.6 AI-enabled blockchain solutions in industrial efficiency AI-enabled blockchain solutions in industrial efficiency are solutions designed to leverage the synergies between. I and blockchain technology to create a powerful toolset to improve industrial production efficiency. These solutions use machine learning, predictive analytics, and other data-driven approaches to optimize operations and reduce downtime [24]. For example, AI-enabled blockchain solutions can detect inefficiencies in operational flows, target specific locations, and enable automated maintenance systems, all of which can optimize production. Moreover, they can provide secure and reliable data sharing between different stakeholders and create immutable records of production and operations data, helping to streamline processes and reduce errors.
20.6.1 Supply chain optimization Supply chain optimization is the process of making a supply chain more efficient by continuously analyzing and improving the activities and processes used to deliver
396
Soumya Mishra et al.
products or services. It requires an examination of all elements involved in the supply chain, from raw material acquisition, to manufacturing, distribution, and logistics, in order to identify areas for improvement. The goal of optimization is to reduce costs, improve efficiency, and maximize customer satisfaction [25]. This is usually done through increased automation, better management of inventory and resources, improved collaboration between supply chain partners, and better use of analytics to identify trends and make timely decisions.
20.6.2 Energy management and sustainability Energy management and sustainability refer to the practice and process of using energy in a way that can improve overall energy efficiency. This includes the use of energy from renewable sources, the use of energy efficient technologies, and the creation of sustainable energy plans. Sustainability is an approach to energy management that seeks to ensure that energy resources are used more responsibly over time, while minimizing environmental impacts. The goal of energy management and sustainability is to reduce energy consumption and improve the efficiency of energy use, which in turn reduces resource consumption and emissions. Energy efficiency saves money, reduces the need for electricity or fuel supply, and reduces greenhouse gas emissions [26]. Sustainability initiatives such as energy audits, benchmarking, and implementing smart energy strategies can improve energy utilization, enhance system functionality, and reduce operating costs.
20.6.3 Quality control and defect detection Quality control (QC) is the process of ensuring that a product or service meets a certain set of quality standards. QC includes inspecting, testing, and validating a product or delivery service to meet customer requirements and expectations. QC is used to identify defects in products or services, and to guarantee that they conform to established standards. Defect detection is the process of identifying defects or errors in a product or service and taking corrective action. Defects can be physical defects or deficiencies in the functionality of the product or service. Defect detection can be done through manual and automated methods. Manual inspection is typically used for low production rates, while automated testing is used when production rates are high.
20 Enhancing industrial efficiency with AI-enabled blockchain-based solutions
397
20.7 Challenges and future directions The main challenge that blockchain technology faces in industrial applications is the lack of understanding and trust between various entities and industry organizations. As blockchain technology is relatively new, many industrial companies have yet to come to terms with the full potential of the technology and tend to be overly cautious in adopting it. Additionally, the industry’s legacy systems are typically not built in a way that can leverage blockchain technology [27]. Furthermore, there are several governance and scalability issues that need to be addressed. In order to effectively integrate blockchain technology into industrial applications, there is a need for a comprehensive framework outlining the security and scalability requirements for each implementation. Such a framework should cover the full spectrum of issues from legal and financial implications to data privacy and compliance. Additionally, the framework must be able to handle multiple technologies and manage different parts of the blockchain stack. In the future, industrial applications will benefit greatly from the adoption of distributed ledger technology (DLT), which is already being leveraged by financial services and other industries. As technology progresses, blockchain technology will become more standardized and easier to manage, paving the way for wider adoption across multiple sectors [28]. Additionally, more businesses will seek to leverage the technology to create smart contracts and automate processes on the blockchain. Finally, we can expect the emergence of more decentralized networks built on blockchain technology, enabling a truly global scale of data sharing.
20.7.1 Limitations of AI-enabled blockchain solutions in industrial efficiency The industrial use of AI-enabled blockchain solutions has tremendous potential, but it does face some limitations [29]. Although the accuracy of AI-enabled blockchain solutions can be quite reliable, the general uncertainty and complexity of human decisions and preferences can cause decisions to be imprecise, leading to lower levels of efficiency. Additionally, due to the inherent complexity of blockchain technologies and limited access to short- and long-term data, blockchain-enabled systems may be unable to make the most optimal decisions to maximize efficiency in certain situations. Furthermore, lack of data privacy and security can be a major concern, as malicious actors may attempt to gain access to and modify sensitive data, leading to system instability and increased risk of unauthorized data access/usage. Finally, current blockchain technologies lack scalability and lack prospective practical riskreward balance for a wide range of applications in industrial settings [30].
398
Soumya Mishra et al.
20.7.2 Ethical considerations in AI-enabled industrial efficiency Ethical considerations in AI-enabled industrial efficiency have become increasingly important as more companies begin to employ AI-based technologies. Companies must understand the potential risks and ethical implications of using AI-based technologies in their production processes to ensure compliance with both legal and ethical standards. Some of the most common ethical considerations when dealing with industrial efficiency include the potential for overreach or the misuse of data, as well as the lack of transparency in implementation and decision-making [31]. Companies should ensure that any data collected is used responsibly and securely, within the bounds of the law. Privacy considerations need to be taken into account, both for users of the technology and data subjects. Additionally, companies should ensure that algorithms are tested and monitored to identify and address bias, as well as consider the potential implications of any decisions made by algorithms. Companies should also consider the impact of AI-enabled technology on the workforce and employ policies that prioritize the safety and wellbeing of employees. Finally, companies should consider potential issues of unintended consequences, such as reduced innovation and employment opportunities, as well as environmental and economic implications associated with the use of AI.
20.7.3 Future trends and research directions The future of AI in industry is bright. AI technologies are being adopted in a wide variety of industrial sectors, from robotics to healthcare. As AI advances and its applications further evolve, new trends and research directions will emerge [32]. One current trend is the use of AI for predictive maintenance (PdM). AI-assisted PdM can detect potential problems in equipment before they actually happen, offering insights that can help improve operational efficiency and save costs. Another trend is the development of AI-driven automation. Automation allows machines to be programmed with routines that reduce time-consuming manual labor while allowing cost savings and increased efficiency. AI-driven automation can be used in a variety of applications, including warehousing, logistics, and manufacturing. In addition to the current trends in the application of AI for industrial efficiency, there are several research directions for the future development of AI in industry. One area of research is the application of AI for smarter decision-making. AI-driven decision-making systems can evaluate data and make decisions based on the data in near real time, enabling companies to adapt their strategies quickly and save time and costs significantly [33]. Another upcoming trend is the use of AI for distributed systems across multiple locations. In a distributed system, AI can be used to control and monitor equipment from remote locations, improving effectiveness and reducing downtime.
20 Enhancing industrial efficiency with AI-enabled blockchain-based solutions
399
In conclusion, AI is revolutionizing industry with its potential to boost operational efficiency and productivity. Current trends and research directions are giving us an idea of where the future of AI in industry is headed. As AI continues to progress, new trends and research areas will emerge, offering exciting new possibilities for the future of industrial efficiency.
References [1]
Haenlein, M., & Kaplan, A. (2019). A brief history of artificial intelligence: On the past, present, and future of artificial intelligence. California Management Review, 61(4), 5–14. [2] Zhang, C., & Lu, Y. (2021). Study on artificial intelligence: The state of the art and future prospects. Journal of Industrial Information Integration, 23, 100224. [3] Schwalbe, N., & Wahl, B. (2020). Artificial intelligence and the future of global health. Lancet I, 395(10236), 1579–1586. [4] Schiff, D. (2021). Out of the laboratory and into the classroom: The future of artificial intelligence in education. AI and Society, 36(1), 331–348. [5] Belfield, H. (2020, February). Activism by the AI community: Analysing recent achievements and future prospects. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society (pp.15–21). [6] Biswas, S., Carson, B., Chung, V., Singh, S., & Thomas, R. (2020). AI-bank of the future: Can banks meet the AI challenge. New York: McKinsey & Company. [7] Cave, S. (2020, February). The problem with intelligence: Its value-laden history and the future of AI. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society (pp.29–35). [8] Sur, J., Bose, S., Khan, F., Dewangan, D., Sawriya, E., & Roul, A. (2020). Knowledge, attitudes, and perceptions regarding the future of artificial intelligence in oral radiology in india: A survey. Imaging Science in Dentistry, 50(3), 193. [9] Aggarwal, K., Mijwil, M. M., Al-Mistarehi, A. H., Alomari, S., Gök, M., Alaabdin, A. M. Z., & Abdulrhman, S. H. (2022). Has the future started? The current growth of artificial intelligence, machine learning, and deep learning. Iraqi Journal for Computer Science and Mathematics, 3(1), 115–123. [10] Sangeetha, R., Logeshwaran, J., Rocher, J., & Lloret, J. (2023). An improved agro deep learning model for detection of Panama wilts disease in banana leaves. AgriEngineering, 5(2), 660–679. [11] Singh, P. K., Chauhan, S. S., Sharma, A., Prakash, S., & Singh, Y. (2023). Prediction of higher heating values based on imminent analysis by using regression analysis and artificial neural network for bioenergy resources. Proceedings of the Institution of Mechanical Engineers, Part E: Journal of Process Mechanical Engineering, 09544089231175046. [12] Chauhan, S. S., Yadav, S. P., Awashthi, S., & Naruka, M. S. (2023). Sentimental analysis using cloud dictionary and machine learning approach. In: Kaushal Kishor, Neetesh Saxena, Dilkeshwar Pandey (Eds.), Cloud-based Intelligent Informative Engineering for Society 5.0 (pp. 157–169). Boca Raton: Chapman and Hall/CRC. [13] Awasthi, S., Naruka, M. S., Yadav, S. P., & de Albuquerque, V. H. C. (Eds.). (2023). AI and IoT-based intelligent health care & sanitation. United Arab Emirates: Bentham Science Publishers. [14] Yadav, S. P., Zaidi, S., Nascimento, C. D. S., De albuquerque, V. H. C., & Chauhan, S. S. (2023, January). Analysis and design of automatically generating for GPS based moving object tracking system. In 2023 International Conference on Artificial Intelligence and Smart Communication (AISC) (pp. 1–5). IEEE. [15] Saravanan, V., Sumathi, A., Shanthana, S., & Rizvana, M. (2013). Dual mode mpeg steganography scheme for mobile and fixed devices. International Journal of Engineering Research and Development, 6, 23–27.
400
Soumya Mishra et al.
[16] Cihon, P., Maas, M. M., & Kemp, L. (2020). Fragmentation and the future: Investigating architectures for international AI governance. Global Policy, 11(5), 545–556. [17] Stahl, B. C. (2021). Artificial intelligence for a better future: An ecosystem perspective on the ethics of AI and emerging digital technologies (p. 124). New York: Springer Nature. [18] Allam, Z., Dey, G., & Jones, D. S. (2020). Artificial intelligence (AI) provided early detection of the coronavirus (COVID-19) in China and will influence future Urban health policy internationally. Ai, 1 (2), 156–165. [19] Saravanan, V., & Chandrasekar, C. (2012). Qos-continuous live media streaming in mobile environment using vbr and edge network. International Journal of Computer Applications, 53(6), 1–12. [20] Lakshminarayanan, R., & Yuvaraj, N. (2020). Design and analysis of an improved deep learning algorithm on classification of intervertebral discs. International Journal of Advanced Science and Technology, 29, 4019–4026. [21] Gupta, S., Saravanan, V., Choudhury, A., Alqahtani, A., Abonazel, M. R., & Babu, K. S. (2022). Supervised computer-aided diagnosis (CAD) methods for classifying alzheimer’s disease-based neurodegenerative disorders. Computational and Mathematical Methods in Medicine, 2022, 1–9. [22] Indhumathi, R., Amuthabala, K., Kiruthiga, G., Yuvaraj, N., & Pandey, A. (2023). Design of task scheduling and fault tolerance mechanism based on GWO Algorithm for attaining better QoS in cloud system. Wireless Personal Communications, 128(4), 2811–2829. [23] Zhang, K., & Aslan, A. B. (2021). AI technologies for education: Recent research & future directions. Computers and Education: Artificial Intelligence, 2, 100025. [24] Vaishya, R., Javaid, M., Khan, I. H., & Haleem, A. (2020). Artificial intelligence (AI) applications for COVID19 pandemic. Diabetes and Metabolic Syndrome: Clinical Research and Reviews, 14(4), 337–339. [25] Aruna R, D., Surendran S, D., Yuvaraj N, D., & Debtera, B. (2022). An enhancement on convolutional artificial intelligent based diagnosis for skin disease using nanotechnology sensors. Computational Intelligence and Neuroscience, 2022, 1–11. [26] Berente, N., Gu, B., Recker, J., & Santhanam, R. (2021). Managing artificial intelligence. MIS Quarterly, 45(3), 1433–1450. [27] Yuvaraj, N., Praghash, K., & Karthikeyan, T. (2022). Privacy preservation of the user data and properly balancing between privacy and utility. International Journal of Business Intelligence and Data Mining, 20(4), 394–411. [28] Ramesh, G., Logeshwaran, J., Kiruthiga, T., & Lloret, J. (2023). Prediction of energy production level in large PV plants through AUTO-Encoder Based Neural-Network (AUTO-NN) with restricted Boltzmann feature extraction. Future Internet, 15(2), 46. [29] Saravanan, V., Sreelatha, P., Atyam, N. R., Madiajagan, M., Saravanan, D., & Sultana, H. P. (2023). Design of deep learning model for radio resource allocation in 5G for massive iot device. Sustainable Energy Technologies and Assessments, 56, 103054. [30] Borges, A. F., Laurindo, F. J., Spínola, M. M., Gonçalves, R. F., & Mattos, C. A. (2021). The strategic use of artificial intelligence in the digital era: Systematic literature review and future research directions. International Journal of Information Management, 57, 102225. [31] Lu, Y. (2019). Artificial intelligence: A survey on evolution, models, applications and future trends. Journal of Management Analytics, 6(1), 1–29. [32] Logeshwaran, J., Shanmugasundaram, N., & Lloret, J. (2023). L-RUBI: An efficient load-based resource utilization algorithm for bi-partite scatternet in wireless personal area networks. International Journal of Communication Systems, 36(4), 1–17. [33] Fouad, H., Hassanein, A. S., Soliman, A. M., & Al-Feel, H. (2020). Analyzing patient health information based on IoT sensor with AI for improving patient assistance in the future direction. Measurement, 159, 107757.
Index abnormalities 215 account 277, 279–280, 289, 292, 294 accountability 189–190, 196–198, 203–206, 208 accountable 190, 192, 194, 198 accuracy 1–2, 6, 10–11, 14, 93, 104–105, 107, 125–141, 145–149, 153–155, 157, 159–160, 162, 168–170, 173–183, 186, 253, 269–274, 276, 278, 281 accurate 23–24, 26, 33, 35–36, 38, 41, 49, 55–56, 58, 62, 67–69, 71, 73–75, 77–80, 216–217, 220, 226–227 acquisition 224 ad hoc network 333 adaptable 90 addressing 202, 206, 208 adjustable 174 administrators 339 adversarial training 121 advertising 246 affordable 374 agencies 137 agents 118–121, 271, 274–275 aggregating 153, 168, 170, 177 agile 282 agreement 128–129 AI 21–23, 27–28, 30, 41, 43–49, 345–346, 350, 352–357, 373–376, 378–382, 387, 391–395, 397–399 AI-driven 251–254, 256–259, 261–262, 264, 266 algorithm 126–127, 131, 134–135 algorithms 1, 3–4, 5, 7, 9–13, 16, 21–26, 29–32, 34, 36–39, 42, 46, 53, 61–65, 68, 70–76, 79, 81–83, 167–168, 171–173, 175, 177–178, 180–185, 189–195, 200, 208, 253–255, 257–261, 263–264, 266, 288, 291–296, 302–303, 305, 332, 334–340 allocations 218 alternative 109 Amazon 311 ambiguities 347 ambiguity 364 analysis 345–357, 361–363, 365, 367–370 analytics 310, 315, 317–319, 323, 326 analyze 345–346, 348, 350–352, 354, 357 annealing 155, 292 anomalies 21, 23, 36, 40, 42, 47 anomaly 24, 173, 175, 233, 235 https://doi.org/10.1515/9783111323749-021
anonymization 196, 317 anticipate 254 antiquated 39 appearances 40 applications 2, 5–6, 7, 9–12, 15, 17, 154–155, 161, 233, 235, 287–291, 293, 297, 302–304, 331, 333, 335–339 approach 167, 170, 173–174, 180–181, 183, 185 approaches 109, 113, 115–118 Apriori 87, 95, 106 Apriori algorithm 87, 95, 106 architectures 5, 299 artificial 1–2, 3, 4, 9, 13–14, 213, 215–217, 219, 224–226, 228, 233, 235–236, 238, 241, 243–246, 251, 253–258, 261–262, 269, 287, 291, 361–362, 365, 367–368 artificial intelligence 1, 3–4, 5, 6, 7, 8, 9, 10, 11, 13–18, 291–292, 294, 297, 345, 348, 350, 352, 354–355 assessments 129 association 24, 345 assumptions 152, 154, 161, 200, 205 asymmetry 99 audiences 315, 321 audit 204 augmentation 190, 199 augmented 328 authentication 310, 316–318, 325 authorized 317 automate 233–241, 246, 287–288, 291, 293, 295–297, 302–305 automated 2–3, 6, 9, 13, 361–363, 367, 370 automation 22, 41, 43, 45, 192, 201, 233–234, 238–239, 242, 246, 269–274, 276–278, 280–283, 373–376, 378–379 automotive 389–390 autonomous 23, 37–38, 40, 42, 44, 214, 219 autonomous systems 7 autonomy 302 average 130, 133–134, 168–169, 171 bagging 153, 167, 169–171, 176–181, 184 banks 18 Bayesian 235 behaviors 297, 303 beneficial 201, 203, 207 benefit 241, 288
402
Index
benefits 251–252 bias 128, 133, 135–141, 168, 171, 173, 180, 184, 189–193, 200–202, 205, 208, 324, 326 bike 236 bioinformatics 91, 98, 101, 104 Blending 169, 176 blockchain 304, 373–374, 376–378, 380–382, 387, 389–391, 393–395, 397 blood 334 Bloom 106 Boltzmann 114 boosting 167, 169, 171–173, 176–181, 184, 186 Bootstrap 152–153, 158, 170 bootstrapping 145, 151–155, 158–161 borders 265 burden 271 business 6, 13, 179, 201–202, 233–234, 236, 251–252, 256–257, 265, 269–271, 273, 277, 281–283, 287–289, 296–298, 301–302, 305, 309, 316, 319, 325 businesses 54, 69, 309–311, 314, 316–318, 320–321, 323–325, 327–328, 361–362, 368, 370 calibration 151 Calinski-Harabasz 92 camera 182 candidate 95 capabilities 269, 271, 273, 282 capitalization 8 capturing 99 carbon 244, 388 categorizations 366 categorize 275 challenge 160, 370 challenges 193, 199, 207–208 characteristics 349–350 chatbots 224, 274, 309, 313, 322, 327, 365 churn 315 classes 353 classification 53–54, 56, 58–65, 67–75, 78–80, 82–83 classifiers 169–170 cleaning 105 cloud 293, 300, 304 clustering 23–24, 26, 29, 34, 87–94, 99–107, 148, 193 clusters 87, 89–94, 99–101, 103–104, 106 coefficients 151 cohesion 92
collaboration 387, 396 collisions 214 combinations 135 combining 74, 167–171, 175–177, 182–183, 185 commands 274 communication 109, 118, 120, 273 companies 289, 295, 298 compatible 299–300 competition 252, 259 complex 125, 140, 155, 162 complexity 158, 161, 264 compliance 288, 302, 304 compressing 334 computers 214 computing 184 conclusions 158 confidence 87, 94, 96 confusion matrix 126–127, 131 consequences 189, 191, 195, 197–198, 204 consumer 54–56, 59, 79 consumption 228–229, 332–333, 337, 340 contextual 364 contribution 182 control 331, 333–334, 336–337, 340 conversation 370 conversations 346, 357 co-occurrence 87, 94–96, 102–103 coordination 56, 109, 117–120, 122 corrections 179 correlations 151 corrupted data 377 costs 214–216, 219–221, 227–229 creativity 39 credit 173, 183, 242, 383 cross-selling 312 cross-validation 133–135, 141, 145–150, 152–157, 159–161 current 313, 320, 399 curve 126, 130 customer 21–22, 23, 25, 32, 34–35, 38, 41, 45–48, 174, 183, 213–214, 217–220, 222–224, 233–241, 244–247 customers 309–312, 314–316, 318, 320–327 cyber 288, 294, 301 data 53–54, 55, 56, 57, 58, 60–83, 145–162, 189–208, 251–262, 264–265, 345–351, 353–354, 356–357 data exploration 88
Index
403
data quality 139–140 databases 238, 245 dataset 132–134, 136, 140–141 datasets 21, 24, 26, 28, 54–55, 58, 62–65, 67, 70, 72–73, 81, 312 DBSCAN 87, 89–90, 99–100 decision 125, 128, 130, 136–137, 139–140, 251–262, 264–266 decision trees 53, 59–61, 76, 168, 170–174, 176–177, 179–180 decision-making 59–60, 81 decisions 213, 216–217, 219–220, 224, 227, 230 deep learning 1, 5, 21, 25, 30, 181, 309, 311–313 demographics 311, 315 dendrogram 90 dependability 149 dependencies 95–96 dependency 347 deployment 146, 162–163, 189, 191, 197–200, 202, 204–205, 207 design 189–191, 198, 204 developers 192–193, 204, 207, 289 devices 28, 31, 43 diagnose 332, 339 diagnoses 373–376, 378–379, 382 diagnosis 339 digital 361 dimensionality 336 dimensionality reduction 89, 105 disagreement 128–129 discrepancies 161 discrimination 189, 198, 202, 204 diseases 183 dissatisfaction 364 dissimilarity 92, 101 distorted outcomes 148 diversity 26, 37 documents 349, 351 domains 89–90, 95, 98, 100–102, 106 drawbacks 155, 162 drone racing 119 drones 40, 44 dynamic environments 121–122
educators 225–226 effectiveness 125–127, 131–132, 138, 240, 244 efficiency 214–215, 217, 221, 226–227, 229, 233–234, 236–239, 241, 244–245, 251, 253, 255, 257–258, 331, 336, 387–389, 392–399 efficient 213–215, 217, 219–220, 223–225, 228 efficiently 290, 294–295, 299, 304 electricity 396 elements 194, 200, 203 embeddings 347, 352 emotions 346, 350, 352, 355–357 employ 309, 311 empower 233 encrypted 195 encryption 195–196, 198, 200, 208 energy 42, 44, 48, 102, 119, 219, 227–229, 244–245, 331–332, 334–335, 337, 340, 388, 392, 394, 396 enforcement 137 engagement 311, 314, 361–362 enthusiast 30 entropy 60–61 environment 377, 380 epsilon-greedy 114 equations 30 equipment 389, 393, 398 equipped 215, 219, 221–222 errors 190, 374–375, 377, 381 establishing 194, 203 estimations 170 estimator 151–152 ethical principles 14, 16 ethics 189, 207, 209 evaluate 191 evaluation 125–126, 132, 134, 136–137, 140–141 exclusionary 191 experience 21, 27–28, 39, 41, 48 experiences 233–234, 240–241, 244, 246 experiments 133, 177 expert systems 1, 8 explainability 192–194, 254, 265 exploration 333 extract 345, 348, 350, 354
e-commerce 2–3, 32, 95, 97, 309–311, 314, 316–317, 324–328 economic 200, 207 economy 13 education 288, 303
facial recognition 3–4, 7, 15, 38, 40 factors 318, 320, 324 fairness 189–193, 196–198, 203, 206 fallibility 199 false negatives 126
404
Index
false positives 126 feature 89, 101, 104, 106 features 336 federated 196 feedback 3, 6, 236, 238, 241, 245, 346–348, 350 feels 355 file 106 filtering 311 finance 21, 25, 33, 45–46, 49, 189, 213, 217–218 financial 60, 78, 214, 217–218, 241–242, 245 fine-tune 310 flagging 317 flaws 153 flexibility 331 food 226–227 footage 235 footprint 47, 388 forecasting 54, 57–59, 69, 76 forecasts 169, 171, 173–174, 178 forests 173–175, 179 formulation 280 Forum 207 framework 27, 30–31 frameworks 11, 27–28 fraud 335 fraud detection 9, 15 fraud prevention 32 fraudulent 310, 316–317, 387 fulfillment 238–239, 310, 319 future 291, 296, 303, 305, 387, 392, 397–399 fuzzy 335, 340 game playing 109 Gaussian distributions 90, 100 gender 302 generalizability 133, 135, 141 generalizable 168, 185 generalization 109, 121 generalize 167, 169, 175, 185 genetic 155, 292 geometric 103 goals 300–301, 305 Google 27–28 governance 184, 205–207 governments 11, 13, 195, 200, 205, 207 GPU 27–29 gradient 62, 79 granular 317
granularities 90 guidance 391 haptic 222 hardware 375, 378 harmonic 127, 131 harvesting 44, 49 HCI 3 healthcare 1–2, 3, 6–7, 11, 15–18, 189, 198, 202, 213–216, 222, 233–234, 243–244, 373–382 heavy 277 heuristics 3 hidden patterns 88 hierarchies 100 high-accuracy 73 hitting 229 homogeneous 155 homomorphic 196 hospitality 361 Hubs 207 human 155 human brain 2 Human Resources 261 human thinker 350 hybrid approaches 101 hyperparameter 150, 161, 175, 178 hyperparameters 133–135, 146, 150, 154, 156 hyperplane 63–64 identification 293 identify 345–348, 350, 352–357 image 21, 24–25, 31, 36–37, 39 image recognition 2, 4, 6–7, 16 immutability 391 impact 150, 160, 162 impacting 202 implementing 310, 316, 318, 321–322 implications 189, 191, 197, 201, 203–206, 208, 251–252, 264–266 inclusion 155 inconsistency 290 individuals 190–192, 194–195, 198, 201–205, 207–208 industrial 331, 333, 337, 339, 387–393, 395, 397–399 industries 2, 6, 9–10, 12, 17, 22, 33, 41–42, 44–45, 49, 361, 365 industry 233–234, 236, 239, 334, 337 influence 25, 42
Index
influences 110 information 310, 312–314, 316, 318, 322, 324, 326, 328 initialization 102, 104 initiatives 389, 396 innovation 288 input 5–7, 9, 11 input data 5 insight 126, 133–134, 137 insights 270–271, 273–274, 276–277, 280–282, 375, 379 instability 170 institutions 205 integration 254 integrity 139–140 intelligence 1, 3–4, 185, 189, 192–193, 195–196, 203–205, 208, 213–217, 219, 224–226, 228, 233, 235–236, 238, 241, 243–246, 251, 253–258, 261–262, 269–275, 277–278, 280, 287, 291, 361–362, 365 intelligent 155 intelligent machines 1 interactions 240–241, 244 interfaces 222 interpret 59 interpretability 107, 125, 138–140, 364 Interpretable 193 intervention 288 interventions. 374 intrusions 334, 337 inventory 22, 45, 233, 236, 238–239 investments 217 irrelevant 176, 181 iteration 95 Jaccard 93, 103 job 131 judgment 94 jurisdictions 377 Kappa statistics 128 Keras 28–29 keywords 367 knowledge 1–2, 4, 8, 10, 114, 118, 121, 258, 291, 366, 369 labor 7, 14, 16, 18, 40–41, 43, 48–49, 233–234, 252, 261, 265, 389, 391–393, 398 laboratory 373
405
language 289, 291, 296, 303 laundering 217 leakage 140–141 learning 167–185, 190–192, 194, 196, 199, 208, 331, 335–336, 364–369 learning dynamics 109 ledgers 374 leverage 191 leveraged 328 leveraging 27, 39, 214, 219, 223, 269, 271, 278, 280, 282 lexical 347 libraries 27–28, 30–31 library 27–31 limitations 252 linear regression 53, 76–77, 79 linguistic 351–352, 355–356 logic 1, 3, 7, 12 logistic regression 56–58, 79, 82 logistics 389–390, 392, 396, 398 loop 147, 149–150 low cost 333 loyalty 295, 298, 309, 312, 314–315, 321–322, 327, 362 machine 167–173, 175–185 machine learning 3–5 machine learning algorithms 53 machines 214–215, 226 malicious 194–195, 199–200 manage risk 32 management 270–272, 274, 276, 278–279, 281, 283 managers 18, 271, 276 manipulation 251 manufacturing 214, 221–222, 233, 237–239, 270, 278, 282 map 115, 118 market 252–255, 259–260, 263, 291, 295–296, 304 market basket analysis 91, 97 marketers 277–278, 315–316 marketing 214, 223–224 Markov 109–111 masking 196 material 225 mathematical 30, 36 medical 215, 332–333, 339–340 medical diagnoses 129 medical illnesses 21 medicine 213, 216
406
Index
merging 169–170 meta-learning 121, 167–168, 177 meteorology 333 methodology 153, 158, 160 metrics 87, 92–94, 107, 125–127, 131, 134, 137, 141 military applications 15 mining 87, 94–98, 101, 105–107, 345–346, 348, 354–355 minority 191, 204, 353, 357 misalignment 121 misuse 373 model 147 models 190–193, 195, 199–200, 205, 234–235, 242, 244 monitor 271, 279–280, 331, 333, 337, 339 monitoring 292, 294, 299–301, 303–304, 331–334, 337–340 multinomial 58, 68 multirobot 119–120 mundane 270–271, 277, 281 naive Bayes 65–69 natural language 1–2, 3, 5–7, 9, 12, 15–18 nearest neighbors 53, 69–70, 72 negotiation 257 Netflix 311 network security 332 networks 1–2, 3, 5, 7, 9, 153, 155–156, 331, 338–340, 362, 367–368 neural networks 5, 9, 25–26, 28–29, 31–33, 109, 112, 115, 167–168, 176–177 neuroevolution 26 neuron 25 NLP 2, 6, 18 noise 90, 99, 105 nonstationarity 109 numerous situations 129 obligations 390 observational 42 observations 149, 151–152, 161 occurrence 126 online algorithms 105 ontology 8 operations 30, 32, 44, 47–49, 252, 256–257 opinions 345–346, 350, 354, 361–362, 364, 368, 370 opportunities 91, 191, 201, 251, 255–257, 260–261, 270, 273, 277, 280
optimization 22, 27, 30, 109, 113, 287–290, 292–295, 297, 299, 301, 303, 305 options 252–253 organization 287, 291, 294, 297, 301–302 organizations 137, 191, 194–195, 197, 199–201, 205, 207 outcome 5, 9 outcomes 287, 295, 300–303, 305 overfitting 55–56, 58, 62, 67, 71–73, 77, 81, 146–147, 149–151, 153, 156–157, 168, 170–171, 175, 177, 181, 184 overlapping sections 147 Oversampling 148 parameters 25, 29, 236–237, 331–333 patient care 373–376, 379, 383 patterns 2, 4–5, 9, 15–16, 251–252, 254–256, 259–260, 271–274, 276, 278, 289, 292, 295, 297, 303–305 payments 377 penalties 9 percentage 182 perception 21, 41 performance 3, 17, 24, 27–28, 30, 36, 56–57, 62, 73, 75, 79–80, 82, 125–128, 130–136, 139–141, 168–169, 171, 175–182, 184–185, 388–389, 392 personal 288, 293, 302 personalized 233–234, 238, 241, 243–244, 246 picture 73–75 planning 290–291, 299–300 plans 262 playback 222 polarity 356 policy 25 polynomial 53, 63, 76–77, 79, 82 population 147–148, 151–152, 154, 158, 161, 375, 377 positives 56 potential 153, 158, 161–162, 251–255, 257, 259–261, 263–266, 311, 315, 317–318, 324–325, 327, 363, 370 power 331–332, 334 powerful 167, 169, 171–172, 177, 181 precision 22, 26, 40–41 prediction 22, 26, 34, 53, 55, 58, 61, 65, 67, 69, 71–76, 78–79 predictions 53–56, 58, 61–62, 66–67, 71–73, 75–78, 81–82, 168–176, 178–184
Index
predictive 170–172, 174–175, 177–182, 184, 186, 237, 240, 245 predictive model 128 preferences 309, 311–312, 314–316, 320, 322, 326–327 pressure 270 presumption 129 prices 183 pricing 309, 315, 318–321, 324, 326 principal component analysis 87, 105 principles 189, 196, 203–208 prior 191, 199, 204 privacy 189, 194–195, 204, 272, 283, 387, 391, 397 privacy and security 107 probability 255, 260 process 287–288, 290–292, 294–297, 300, 302–303 processes 233–234, 236–241, 243 processing 21, 23, 25, 28–36 production 22, 41, 44–45, 47–49, 136, 141 productivity 21, 41, 44, 269–270, 272–273, 281, 388, 391, 394, 399 products 22–23, 33–34, 39, 42, 44, 46–48 professionals 374–375, 378–379, 383 profiles 91, 102, 104, 241 profits 318, 320–321 programming 23, 29, 43, 167 Project 299 projected 145 projects 28–29, 31, 55, 68 promote 201, 203, 206 protocols 287, 290, 293–294, 299–301, 335, 340 pruning 95–96, 103, 176 psychological 57 punishments 25 Python 28–30 Q-learning 113, 115–116 quality 216, 221–222, 226–228, 374, 376–377, 382 qubits 196 queries 313–314, 323 questions 269, 274 random 135 random forests 62 real time 280 real-time 270, 275, 279–281 reasoning 2–4, 10, 12 recognition 21, 24–26, 28, 31, 36, 38–39, 45 recommendation 173–174
407
recommendations 201 record-keeping 197 records 373–377, 379, 381–382 recruiters 261 recruitment 261 reference 151 regression 53–54, 56–62, 69–79, 81–83, 151, 156, 160, 168, 173–176, 180–181, 335 regulations 122, 265, 283 regulatory 390–391 reimbursement 382 reinforcement 25, 30, 109–110, 113, 115–117, 121, 234–236, 365–366 relationship 332 relationships 54, 60, 66, 69, 76, 79, 87–91, 94–98, 106, 173, 185 reliability 125, 128–129, 140, 193, 196, 207 reliable 149, 152–153, 158, 160, 162 representation 115 representatives 289 requirements 288–290, 299–302, 304 resampling 145, 152, 158, 162 research 5–7, 12, 17 researchers 193, 199, 203–204, 207 resource 287, 290–291, 297 resources 287–288, 290–293, 297–300, 303–305, 388–389, 391–392, 394, 396 responsible 251, 254 restrictions 45 results 146–147, 149–150, 156–160, 162 reward 109–110, 115–117, 121 ride-sharing apps 18 rights 222, 227, 229 risk 14–15, 18, 168, 177, 251, 255, 259–260, 265 risks 251, 255–256, 259–260, 265 robotics 7, 9, 21, 23, 37, 40–42, 221, 228 robots 21–22, 23, 40–44, 214–215, 221–222, 233, 338 rules 87, 94–99 safety 215–216, 218–220, 222, 228, 272, 283 sales 291, 296, 298, 309–312, 314, 327 sampling 114, 190 sanitization 196 satisfaction 202, 288–289, 298, 300 savings 388, 398 scalability 109, 120, 269, 273, 331, 334 scanning 293 scheduling 278
408
Index
scoping 299 seamless 106 security 107, 139–140, 189, 191–192, 195–196, 198–201, 203–204, 208, 272, 282–283, 309–311, 316–317, 325, 328 segmentation 174 segmenting 23, 38 self-configuration 332 semantic 313 semantics 352 sensor 237, 331–335, 338–339 sentiment 68–69, 71, 75, 222–223, 316, 322–323, 345–357, 361–365, 367–370 sentiment analysis 345–349, 351–357 series 150–151, 158 service 288–289, 291, 295, 297–298, 305 services 233–234, 238, 240–241 shipments 319, 390 shipping 47 shopping 310–311, 314, 328 signals 315, 321 significance 109, 112 simulated 292 social networks 91, 101, 104 Sockets 294 software 136, 375, 379 soil 337 solutions 153, 269, 272, 274, 282, 362, 365 sophistication 251, 270 spam filtering 68 speech 236 speed 22, 28, 30, 35–36, 43–44, 269–271, 274, 278 splitting 148–150, 156, 162 stacking 73–74, 169, 176–178, 181 stakeholders 192, 194, 206, 209 standards 190, 194, 198, 203, 205–209, 289, 300–301 statistical 145, 152, 154, 156, 158, 160 stock 183, 241, 350 stoppages 292 strategies 109, 113, 115–117, 119–121 strategy 57, 61–62, 67 streamline 269–270, 276, 279 streamlining 374, 379 structure 153, 158 subsets 145–148, 150, 154, 157, 161 superior 168–169, 173, 178, 182 supervised 3, 54–56, 59–65, 71, 76–78 supervised learning 54–56, 59–60, 62–63, 65, 76, 78
supervision 366 supply chain 120, 318 support vector machine 167 sustainability 396 swarm 336 synthesizes 276 system 125–127, 130, 137–138 systems 361–366, 370 tactile 222 target 314–315, 320, 327 targeting 277, 312, 315, 321 tasks 21–23, 25, 27–29, 33–34, 37, 40–41, 43–48, 167–168, 175, 177, 183, 185 taxation 276 taxonomies 8 taxonomy 2, 8 technologies 213, 218–219, 221–222, 224–228, 230, 270–274, 276, 278, 280 technology 2, 4, 6–7, 8, 9, 10, 13–18, 251, 255, 258, 260–265, 271–280, 282–283, 291, 295, 297, 303–305, 309, 313, 319, 322, 324, 326, 373–382, 387, 389–391, 394–395, 397–398 tedious 258 telemedicine 374 temperatures 337 TensorFlow 27–30 Testing 55–56, 77 text 361–365, 367–369 text recognition 73, 76 theft 317–318 themes 346, 348 Thompson 114 threshold 130 thresholds 130 throughput 270 tokenization 32 Toolkit 27 traceability 373 tracking 331–333, 337–338 train 301–302 training 24, 26–28, 31, 33, 37–38, 54–55, 62, 67, 69, 71–74, 78–82, 125, 131–135, 140, 145–151, 153–157, 161–162, 288 transactional data 91, 97 transition 110–112, 115 translation 26, 31, 33, 35 transparency 11, 14, 16, 189, 192–194, 196, 198, 203–206, 208, 299–302
Index
transparent 191–192, 194, 197–198, 205–206 transportation 44, 119, 213–214, 219–220, 228–229 treatment 97 treatments 2, 18, 215–216, 243–244, 373–375, 377, 379, 382 trends 289, 291, 295–296, 303–304 truck 238 true negatives 126 trust 189, 191–196, 198, 204, 206–207, 373, 377, 380–381 tumors 216 tuning 134–135, 141, 175 unauthorized 194–195, 198 understanding 345–348, 352–353, 355–357 unintended 189, 198, 204 unlock 269 unlocked 206 unmanned aerial vehicles 339 unprocessed 349 unsupervised 3, 235, 365–366 unsupervised learning 23–25 users 238, 240, 247 validation 145–151, 153–157, 159–161 variability 151–152, 156, 158
variables 173–174, 182 variety 170, 173, 178, 180–182 vector 53–54, 62, 64, 74, 77–79, 82 vehicles 213–214, 219–221 vendor 287–290, 293–294, 296, 298, 300, 302, 304 vendors 287, 289–290, 294–296, 298, 300, 304 verification 390, 394 vibration 237 video 112, 117, 119 videos 303 virtual agents 9 virtual reality 7, 16 visualization 88–89, 94, 100, 194 visualizations 328 voice recognition 310 vote 169, 171–172, 176 vulnerability 293 waste 40–41, 47–48 weigh 253 window 150–151 wireless 331–334, 338–339 wireless sensor networks 332 yield 238
409