Interactive Collaborative Robotics: 8th International Conference, ICR 2023, Baku, Azerbaijan, October 25–29, 2023, Proceedings (Lecture Notes in Artificial Intelligence) 3031431103, 9783031431104

This book constitutes the refereed proceedings of the 8th International Conference on Interactive Collaborative Robotics

138 69 51MB

English Pages 396 [395] Year 2023

Report DMCA / Copyright

DOWNLOAD PDF FILE

Table of contents :
Preface
Organization
Contents
Attention Guided In-hand Mechanical Tools Recognition in Human-Robot Collaborative Process
1 Introduction
2 Related Work
2.1 Hand Object Interaction
2.2 Occlusion-Aware Object Detection
3 Proposed Approach
3.1 Criss-Cross Attention Module
3.2 SimAM Attention Module
4 Experiments
4.1 Dataset Preparation
4.2 Implementation Details
4.3 Analysis and Comparisons
5 Conclusion
References
Design and Implementation of a Multimodal Combination Framework for Robotic Grasping
1 Introduction
2 Grasp Understanding and Detection
2.1 Language Text Feedback
2.2 6-DoF Grasp Generation
3 Robotic Motor Skills Generalization
3.1 Human Motion Demonstration
3.2 DMP and Obstacle Avoidance
4 Experiments
4.1 Evaluation of Robotic Grasp Understanding
4.2 Evaluation of Grasp Detection
4.3 Evaluation of Robotic Trajectory Learning
5 Conclusion
References
Experimental Validation of an Interface for a Human-Robot Interaction Within a Collaborative Task
1 Introduction
2 Related Work
3 Materials and Methods
3.1 UR Robots – Virtual Control Application
3.2 Workcell Description
3.3 Collaborative Task
4 Experiment Description
4.1 Training Stage
4.2 Main Stage
4.3 Evaluation
5 Results
6 Discussion
7 Conclusion
References
Fail It till You Make It: Error Expectation in Complex-Plan Execution for Service Robots
1 Introduction
2 Related Work
3 Error Expectation in General Purpose Service Robot (GPSR)
4 Experiments and Results
4.1 Object Recognition
4.2 Object Detection
5 Conclusions
References
Moving Person Detection Based on Modified YOLOv5
1 Introduction
2 Related Works
2.1 Dynamic SLAM Based on Deep Learning
2.2 Person Detection in Indoor Environments
3 New Model Architecture and Methodology
3.1 Overview of YOLOv5 Network
3.2 Convolutional Block Attention Module (CBAM)
3.3 Bidirectional Feature Pyramid Network (BiFPN)
3.4 New Model Architecture
4 Experimental Results
4.1 Datasets Preparation and Training Results
4.2 Performance Evaluation on TUM RGB-D Dataset
4.3 Performance Evaluation on Real-Time Situation
5 Conclusion and Future Work
References
Autonomous Robot Navigation System as Part of a Human-Machine Team Based on Self-organization of Distributed Neurocognitive Architectures
1 Introduction
2 Autonomous Robot Navigation System Based on Multi-agent Neurocognitive Architecture
3 Experiments
4 Conclusion
References
3D-CNNs-Based Touchless Human-Machine Interface
1 Introduction
2 Related Works
3 Proposed Method
3.1 Extraction of the Hand Region
3.2 Moore–Neighbor Approach
3.3 Detecting Hand Gestures Target-Person Locking
4 Simulations, Experimentations and Discussion of the Results
5 Conclusion
References
Development of a Mechanism for Recognizing the Emotional State Based on the Unconscious Movements of the Subject
1 Introduction
2 The Need for a New Modality
3 Creation of a Mechanism for Recognizing the Emotional State of the Subject Based on His Unconscious Actions
3.1 Designing a Mechanism for Recognizing the Emotional State of the Subject
3.2 Collect Spatial Characteristics of the Mouse Cursor Position and Create Own Dataset
3.3 Research and Data Processing
4 Conclusion
References
Development of a Device for Post-traumatic Ankle Rehabilitation
1 Introduction
2 Tasks of Ankle Rehabilitation
3 The Circuit Diagram of the Active-Passive Mechanotherapy Device
4 Diagrams for Measuring Leg and Platform Interaction Forces
5 Functional Diagram of the Control System
6 Description of the Unit’s Operating Modes
7 Conclusion
References
Evaluation of EEG Data for Zonal Affiliation of Brain Waves by Leads in a Robot Control Task
1 Introduction
2 Collecting and Preparing Statistical Data
3 Obtaining Representative Features
4 Evaluation of the Representativeness of the EEG Data
5 Conclusion
References
Comparison of ROS Local Planners for a Holonomic Robot in Gazebo Simulator
1 Introduction
2 System Setup
2.1 Mecanum Wheel Robot
2.2 Virtual Environments
3 ROS Local Planners
4 Performance Comparison
5 Conclusion
References
Movement Along the Trajectory of a Home Quadruped Robot
1 Introduction
2 Model of Home Robot
3 Movement Along the Trajectory
3.1 Piecewise Linear Approximation
3.2 Gait with Variable Turning Radius (VTR Gate)
4 Experiments
5 Conclusion
References
Study of Path Planning Methods in Two-Dimensional Mapped Environments
1 Introduction
2 Overview of Motion Planning Methods
3 Mathematical Description of the Robot Movement and the Environment
4 Modified Rapidly-Growing Random Trees Algorithm
5 Comparison of the Original and the Modified RRT Algorithm
6 Conclusion
References
DHC-R: Evaluating “Distributed Heuristic Communication” and Improving Robustness for Learnable Decentralized PO-MAPF
1 Introduction
2 Background
3 Related Work
4 Method
4.1 Architecture
4.2 Observations
4.3 Training
5 Empirical Evaluation
6 Conclusion
References
Ground Mobile Robot Localization Algorithm Based on Semantic Information from the Urban Environment
1 Introduction
2 Description of the Training Data Collection Methodology for Semantic Segmentation
3 Segmentation Model
4 The Algorithm Developed
5 Results Analysis
6 Conclusion
References
Remote Control Robotic System for the Perimeter Security
1 Introduction
2 Principle of Operation of the Security System
3 Choosing the Encryption Algorithm of the Security System
4 Structure of the Security System
5 Description of the Security Robotic System
6 Security Robotic System Operation
6.1 Preparation for Operation
6.2 Operation Algorithm
7 Results of Laboratory and Field Tests of Security System
8 Conclusion
References
Development of a Robot for Agricultural Field Scouting
1 Introduction
2 Movement Organization
2.1 Precise Positioning with RTK
2.2 Building a Map in an Unknown Area
2.3 Obstacle Avoidance
3 Pipeline of Robot Works
3.1 Robot Concept
3.2 Choice of Components
3.3 Pipeline of Our Model
4 Conclusion
References
Optimization of the Placement of Measurement Points and Control of the Power of Moving Sources in Rod Heating
1 Introduction
2 Formulation of the Problem
3 Approach and Formulas for the Numerical Solution of the Problem
4 Numerical Experiments
5 Conclusion
References
Design of Hybrid Control System for Nonaffine Plants
1 Introduction
2 Formulation of the Problem
3 Solution of the Problem
4 Example of Control Design and Simulation Results
5 Conclusion
References
Sliding-Mode Control of Phase Shift for Two-Rotor Vibration Setup
1 Introduction
2 Servo System Model of Two-Rotor Vibration Setup
3 Robust Phase-Shift Control Algorithm
4 Exploring Sliding Mode Potential
4.1 Analytical Study
4.2 Numerical Study
5 Conclusion
References
GBMILs: Gradient Boosting Models for Multiple Instance Learning
1 Introduction
2 Preliminaries
2.1 Multiple Instance Learning
2.2 A Brief Introduction to GBM
3 Gradient Boosting for MIL
3.1 Aggregating Functions
3.2 Hard Max Aggregation
3.3 Simple Attention Aggregation
3.4 Ensemble of GBMs and the Attention Neural Networks
4 Numerical Experiments
5 Conclusion
References
Approach to Numerical Solution of Nonlinear Optimal Feedback Control Problems
1 Introduction
2 Problem Statement and Solution Scheme
3 Results of Numerical Experiments
4 Conclusion
References
Model-Based Policy Optimization with Neural Differential Equations for Robotic Arm Control
1 Introduction
2 Related Work
3 Background
4 Method
5 Experiments
6 Conclusion
References
Monitoring the State of Robotic Systems Based on Time Series Analysis
1 Introduction
2 Problem Statement
3 Suggested Method
4 Approbation of the Method
4.1 Practical Implementation
4.2 Results
5 Conclusion
References
Resource-Saving Multiobjective Task Distribution in the Fog- and Edge-Robotics
1 Introduction
2 Resource-Saving Tasks Distribution: Basic Approach
3 Multiobjective Tasks Distribution: Problem Statement
4 Some Assumptions and General Technique for Resource-Saving Multiobjective Task Distribution
5 Experimental Comparison of the Greedy and Multiobjective Solutions
6 Experimental Multiobjective Tasks Distribution
7 Conclusion
References
Reliability of Robot’s Controller Software
1 Introduction
2 Simulation of Robot’s Digital Control System
3 Causes of Software Failures
4 Model of Failure Flux Forming
5 Example
6 Conclusion
References
Ontological Approach to the Organization of Computing in Distributed Monitoring Systems with Mobile Components Based on a Distributed Ledger
1 Introduction
2 Ontological Approach to the Organization of Computations in Distributed Systems
3 Development of a Domain Ontological Model
3.1 Methodology for the Development of a Domain Ontological Model
3.2 Analysis of the Subject Area of Operation of a Distributed Monitoring System with Mobile Components Based on a Distributed Ledger
3.3 Domain Ontological Model of Distributed Monitoring System Functioning with Mobile Components Based on DL Technologies
4 Experimental Studies
5 Results
6 Conclusion
References
Improved Model of Greedy Tasks Assignment in Distributed Robotic Systems
1 Introduction
2 Robotics and Distributed Computing Domains Intersection
3 Tasks Distribution Efficiency Estimation
4 Experimental Results
4.1 Experiment 1. Small Data Volume to Transmit, Route №1
4.2 Experiment 2. Large Data Volume to Transmit, Route №1
4.3 Experiment 3. Small Data Volume to Transmit, Route №2
4.4 Experiment 4. Large Data Volume to Transmit, Route №2
5 Conclusion
References
Construction of a Three-Dimensional UAV Movement Planner When the Latter Moves in Conditions of Difficult Terrain
1 Introduction
2 Problem Statement
3 Generalized Global Scheduler Algorithm
4 Algorithm for Thinning a Piecewise Polyline and Reducing its Total Length
5 Modified Dubins Method
6 Trajectory Correction Algorithm Taking into Account Kinematic Constraints
7 Simulation Results
8 Conclusion
References
Identification of the Quadcopter Rotational Dynamics for the Tilt Angle
1 Introduction
2 Finite Frequency Identification Case
3 Nonlinear Model Identification
4 Conclusion
References
Development of a Firmware for Multirotor UAV Flight Controller Implemented on MCU MDR 32
1 Introduction
2 Problem Statement
3 Model of the Flight Controller
4 Design of the Flight Controller Hardware
5 Software Implementation of the Flight Controller
6 Experimental Testing
7 Conclusion
References
Autonomous Landing Algorithm for UAV on a Mobile Robotic Platform with a Fractal Marker
1 Introduction
2 Analysis of Existing Research
3 Materials and Methods
3.1 Used Models
3.2 The Landing Control Algorithm
4 Experiments
5 Analysis of Obtained Results
6 Conclusion
References
Curl-Free Vector Field for Collision Avoidance in a Swarm of Autonomous Drones
1 Introduction
2 Curl-Free Vector Field-Based Approach to Drone Swarm Control
2.1 Drone Swarm System Model
2.2 Control Algorithm for Collision Avoidance
3 Numerical Simulation Results and Discussion
4 Conclusion
References
Author Index
Recommend Papers

Interactive Collaborative Robotics: 8th International Conference, ICR 2023, Baku, Azerbaijan, October 25–29, 2023, Proceedings (Lecture Notes in Artificial Intelligence)
 3031431103, 9783031431104

  • 0 0 0
  • Like this paper and download? You can publish your own PDF file online for free in a few minutes! Sign Up
File loading please wait...
Citation preview

LNAI 14214

Andrey Ronzhin Aminagha Sadigov Roman Meshcheryakov (Eds.)

Interactive Collaborative Robotics 8th International Conference, ICR 2023 Baku, Azerbaijan, October 25–29, 2023 Proceedings

123

Lecture Notes in Computer Science

Lecture Notes in Artificial Intelligence Founding Editor Jörg Siekmann

Series Editors Randy Goebel, University of Alberta, Edmonton, Canada Wolfgang Wahlster, DFKI, Berlin, Germany Zhi-Hua Zhou, Nanjing University, Nanjing, China

14214

The series Lecture Notes in Artificial Intelligence (LNAI) was established in 1988 as a topical subseries of LNCS devoted to artificial intelligence. The series publishes state-of-the-art research results at a high level. As with the LNCS mother series, the mission of the series is to serve the international R & D community by providing an invaluable service, mainly focused on the publication of conference and workshop proceedings and postproceedings.

Andrey Ronzhin · Aminagha Sadigov · Roman Meshcheryakov Editors

Interactive Collaborative Robotics 8th International Conference, ICR 2023 Baku, Azerbaijan, October 25–29, 2023 Proceedings

Editors Andrey Ronzhin St. Petersburg Federal Research Center of the Russian Academy of Sciences St. Petersburg, Russia

Aminagha Sadigov Institute of Control Systems of Ministry of Science and Education of the Republic of Azerbaijan Baku, Azerbaijan

Roman Meshcheryakov V.A. Trapeznikov Institute of Control Sciences of the Russian Academy of Sciences Moscow, Russia

ISSN 0302-9743 ISSN 1611-3349 (electronic) Lecture Notes in Artificial Intelligence ISBN 978-3-031-43110-4 ISBN 978-3-031-43111-1 (eBook) https://doi.org/10.1007/978-3-031-43111-1 LNCS Sublibrary: SL7 – Artificial Intelligence © The Editor(s) (if applicable) and The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 This work is subject to copyright. All rights are reserved by the Publisher, whether the whole or part of the material is concerned, specifically the rights of translation, reprinting, reuse of illustrations, recitation, broadcasting, reproduction on microfilms or in any other physical way, and transmission or information storage and retrieval, electronic adaptation, computer software, or by similar or dissimilar methodology now known or hereafter developed. The use of general descriptive names, registered names, trademarks, service marks, etc. in this publication does not imply, even in the absence of a specific statement, that such names are exempt from the relevant protective laws and regulations and therefore free for general use. The publisher, the authors, and the editors are safe to assume that the advice and information in this book are believed to be true and accurate at the date of publication. Neither the publisher nor the authors or the editors give a warranty, expressed or implied, with respect to the material contained herein or for any errors or omissions that may have been made. The publisher remains neutral with regard to jurisdictional claims in published maps and institutional affiliations. This Springer imprint is published by the registered company Springer Nature Switzerland AG The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland Paper in this product is recyclable.

Preface

The International Conference on Interactive Collaborative Robotics (ICR) was launched in 2016 and the 8th International Conference on Interactive Collaborative Robotics was held in Baku, Azerbaijan from October 25–29, 2023. The conference brought together experts and scholars from different fields to discuss the use and challenges of human-robot collaboration in different spheres such as industry, society, healthcare, and education. The main focus of the conference was on the foundations and means of collaborative behavior of one or more robots physically interacting with humans in operational environments configured with embedded sensor networks and cloud services under uncertainty and environmental variability. The International Conference on Interactive Collaborative Robotics (ICR) is deservedly recognized and supported by experts in the field internationally, so the decision to have ICR 2023 in Baku was right to the point; this year’s conference was cohosted by the Institute of Control Systems of the Ministry of Science and Education of the Republic of Azerbaijan (Baku, Azerbaijan), the St. Petersburg Federal Research Center of the Russian Academy of Sciences (St. Petersburg, Russia), and Wenzhou University (Wenzhou, China); moreover, many outstanding professionals from all over the world accepted to act as committee chairs and members. During the conference, scientists, industry experts, and scholars from research and business at home and abroad were given a venue to share their ideas and discuss cuttingedge technologies, industrial products, and industry trends, making it a highly influential event in the robotics industry. More details are available at the conference website: https:// icr.cyber.az/. This volume contains a collection of 33 papers presented at ICR 2023, each thoroughly reviewed by members of the Program Committee consisting of more than 30 top specialists in the conference subject. The papers were selected from 56 submissions in a single-blind peer review process, with each submission receiving at least 3 reviews. Theoretical and more general contributions were presented in oral sessions. Problem-oriented sessions as well as discussions then brought together specialists in limited problem areas with the aim of exchanging knowledge and skills resulting from research projects of all kinds. Special thanks and appreciation go to the members of the Program Committee and Organizing Committee for their diligence and enthusiasm during the conference organization. We extend our gratitude to all participants for their valuable contribution to the success of ICR 2023 and look forward to meeting all of you at the next International

vi

Preface

Conference on Interactive Collaborative Robotics, in 2024. More details are available at: http://icr.nw.ru/. October 2023

Andrey Ronzhin Aminagha Sadigov Roman Meshcheryakov

Organization

General Chairs Ali Abbasov Andrey Ronzhin Min Zhao

Institute of Control Systems of the Azerbaijan National Academy of Sciences, Azerbaijan St. Petersburg Federal Research Center of the Russian Academy of Sciences, Russia Wenzhou University, China

Program Committee Chairs Roman Meshcheryakov Aminagha Sadigov

V.A. Trapeznikov Institute of Control Science of the Russian Academy of Sciences, Russia Institute of Control Systems of the Azerbaijan National Academy of Sciences, Azerbaijan

Program Committee Members Kamil Aida-zade Fikret Aliev Ramiz Aliguliyev Elchin Aliyev Rasim Alizade Yasar Ayaz Branislav Borovac Ivan Ermolov Vagif Gasymov Viktor Glazunov Mehmet Guzey Vagif Ibrahimov

Institute of Control Systems of the Azerbaijan National Academy of Sciences, Azerbaijan Baku State University, Azerbaijan Institute of Information Technology, Azerbaijan Institute of Control Systems of the Azerbaijan National Academy of Sciences, Azerbaijan Azerbaijan Technical University, Azerbaijan National University of Sciences and Technology, Pakistan University of Novi Sad, Serbia Ishlinsky Institute for Problems in Mechanics of the Russian Academy of Sciences, Russia Azerbaijan Technical University, Azerbaijan Mechanical Engineering Research Institute of the Russian Academy of Sciences, Russia Sivas University of Science and Technology, Turkey Baku State University, Azerbaijan

viii

Organization

Ismayil Ismayilov Dimitrios Kalles Alexey Kashevnik Anis Koubaa Evgeni Magid Ilshat Mamaev Kamil Mansimov Geylani Panahov Adalat Pashayev Fahrad Pashayev Viacheslav Pshikhopov Mirko Rakovic Ramin Rzayev Hooman Samani Yulia Sandamirskaya Jesus Savage Evgeny Shandarov Lev Stankevich Sergey Yatsun Alena Zakharova

Azerbaijan National Aviation Academy, Azerbaijan Hellenic Open University, Greece St. Petersburg Federal Research Center of the Russian Academy of Sciences, Russia Prince Sultan University, Saudi Arabia Kazan Federal University, Russia Karlsruhe Institute of Technology, Germany Institute of Control Systems of the Azerbaijan National Academy of Sciences, Azerbaijan Azerbaijan Institute of Mathematics and Mechanics, Azerbaijan Institute of Control Systems of the Azerbaijan National Academy of Sciences, Azerbaijan Institute of Control Systems of the Azerbaijan National Academy of Sciences, Azerbaijan Southern Federal University, Russia University of Novi Sad, Serbia Institute of Control Systems of the Azerbaijan National Academy of Sciences, Azerbaijan University of Hertfordshire, UK Intel, Switzerland National Autonomous University of Mexico, Mexico Tomsk State University of Control Systems and Radioelectronics, Russia Peter the Great St. Petersburg Polytechnic University, Russia Southwest State University, Russia V.A. Trapeznikov Institute of Control Science of the Russian Academy of Sciences, Russia

Organization Committee Chairs Tofig Babayev Polina Chernousova Zengling Ma

Institute of Control Systems of the Azerbaijan National Academy of Sciences, Azerbaijan St. Petersburg Federal Research Center of the Russian Academy of Sciences, Russia Wenzhou University, China

Organization

Organization Committee Members Aygun Aliyeva Tahir Alizada Marina Astapova Ekaterina Cherskikh Natalia Dormidontova Etiram Karimov Elchin Khalilov Dmitriy Levonevskiy Alyona Lopotova Alina Mikhailus Anna Motienko Irina Podnozova

Institute of Control Systems of the Azerbaijan National Academy of Sciences, Azerbaijan Institute of Control Systems of the Azerbaijan National Academy of Sciences, Azerbaijan St. Petersburg Federal Research Center of the Russian Academy of Sciences, Russia St. Petersburg Federal Research Center of the Russian Academy of Sciences, Russia St. Petersburg Federal Research Center of the Russian Academy of Sciences, Russia Institute of Control Systems of the Azerbaijan National Academy of Sciences, Azerbaijan Wenzhou University, China St. Petersburg Federal Research Center of the Russian Academy of Sciences, Russia St. Petersburg Federal Research Center of the Russian Academy of Sciences, Russia St. Petersburg Federal Research Center of the Russian Academy of Sciences, Russia St. Petersburg Federal Research Center of the Russian Academy of Sciences, Russia St. Petersburg Federal Research Center of the Russian Academy of Sciences, Russia

ix

Contents

Attention Guided In-hand Mechanical Tools Recognition in Human-Robot Collaborative Process . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Guo Wu, Xin Shen, and Vladimir Serebrenny

1

Design and Implementation of a Multimodal Combination Framework for Robotic Grasping . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Congyu Huang, Ziyang Wang, Haoran Zhu, Jie Li, and Xiaofeng Liu

13

Experimental Validation of an Interface for a Human-Robot Interaction Within a Collaborative Task . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Maksim Mustafin, Elvira Chebotareva, Hongbing Li, and Evgeni Magid

23

Fail It till You Make It: Error Expectation in Complex-Plan Execution for Service Robots . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Luis Contreras, Jesus Savage, Stephany Ortuno-Chanelo, Marco Negrete, Arata Sakamaki, and Hiroyuki Okada Moving Person Detection Based on Modified YOLOv5 . . . . . . . . . . . . . . . . . . . . . Xin Shen, Guo Wu, and Vadim Lukyanov Autonomous Robot Navigation System as Part of a Human-Machine Team Based on Self-organization of Distributed Neurocognitive Architectures . . . . . . . Inna Pshenokova, Kantemir Bzhikhatlov, Olga Nagoeva, Idar Mambetov, and Alim Unagasov 3D-CNNs-Based Touchless Human-Machine Interface . . . . . . . . . . . . . . . . . . . . . . Ali Asgarov and Ali Parsayan Development of a Mechanism for Recognizing the Emotional State Based on the Unconscious Movements of the Subject . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Yaroslava Gorbunova and Gleb Kiselev Development of a Device for Post-traumatic Ankle Rehabilitation . . . . . . . . . . . . Andrey Knyazev, Sergey Jatsun, Andrey Fedorov, and Jamil Safarov

36

47

59

70

81

93

Evaluation of EEG Data for Zonal Affiliation of Brain Waves by Leads in a Robot Control Task . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103 Daniyar Wolf, Yaroslav Turovsky, Anastasia Iskhakova, and Roman Meshcheryakov

xii

Contents

Comparison of ROS Local Planners for a Holonomic Robot in Gazebo Simulator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 116 Artem Apurin, Bulat Abbyasov, Edgar A. Martínez-García, and Evgeni Magid Movement Along the Trajectory of a Home Quadruped Robot . . . . . . . . . . . . . . . 127 Dmitry Dobrynin Study of Path Planning Methods in Two-Dimensional Mapped Environments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 137 Nizar Hamdan, Viacheslav Pshikhopov, Mikhail Medvedev, Dimitry Brosalin, Maria Vasileva, and Boris Gurenko DHC-R: Evaluating “Distributed Heuristic Communication” and Improving Robustness for Learnable Decentralized PO-MAPF . . . . . . . . . . . 151 Vladislav Savinov and Konstantin Yakovlev Ground Mobile Robot Localization Algorithm Based on Semantic Information from the Urban Environment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 164 Artur Podtikhov and Anton Saveliev Remote Control Robotic System for the Perimeter Security . . . . . . . . . . . . . . . . . . 175 Azad Bayramov and Samir Suleymanov Development of a Robot for Agricultural Field Scouting . . . . . . . . . . . . . . . . . . . . 185 Olga Mitrofanova, Ivan Blekanov, Danila Sevostyanov, Jia Zhang, and Evgenii Mitrofanov Optimization of the Placement of Measurement Points and Control of the Power of Moving Sources in Rod Heating . . . . . . . . . . . . . . . . . . . . . . . . . . . 197 Kamil Aida-Zade and Vugar Hashimov Design of Hybrid Control System for Nonaffine Plants . . . . . . . . . . . . . . . . . . . . . . 208 Anatoly Gaiduk, Viacheslav Pshikhopov, Mikhail Medvedev, Vladislav Gissov, Ali Kabalan, and Evgeny Kosenko Sliding-Mode Control of Phase Shift for Two-Rotor Vibration Setup . . . . . . . . . . 221 Nikolay Kuznetsov, Boris Andrievsky, Iuliia Zaitceva, and Elizaveta Akimova GBMILs: Gradient Boosting Models for Multiple Instance Learning . . . . . . . . . . 233 Andrei Konstantinov, Lev Utkin, Vladimir Muliukha, and Vladimir Zaborovsky

Contents

xiii

Approach to Numerical Solution of Nonlinear Optimal Feedback Control Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 246 Samir Guliyev Model-Based Policy Optimization with Neural Differential Equations for Robotic Arm Control . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 258 Andrey Gorodetskiy, Konstantin Mironov, and Aleksandr Panov Monitoring the State of Robotic Systems Based on Time Series Analysis . . . . . . 267 Viktor Semenov Resource-Saving Multiobjective Task Distribution in the Fogand Edge-Robotics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 279 Anna Klimenko and Arseniy Barinov Reliability of Robot’s Controller Software . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 289 Eugene Larkin, Tatiana Akimenko, Alexey Bogomolov, and Vadim Sharov Ontological Approach to the Organization of Computing in Distributed Monitoring Systems with Mobile Components Based on a Distributed Ledger . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 300 Eduard Melnik and Irina Safronenkova Improved Model of Greedy Tasks Assignment in Distributed Robotic Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 311 Anna Klimenko Construction of a Three-Dimensional UAV Movement Planner When the Latter Moves in Conditions of Difficult Terrain . . . . . . . . . . . . . . . . . . . . . . . . . 322 Vladimir Kostyukov, Igor Evdokimov, and Vladislav Gissov Identification of the Quadcopter Rotational Dynamics for the Tilt Angle . . . . . . . 334 Vadim Alexandrov, Ilya Rezkov, Dmitrii Shatov, and Yury Morozov Development of a Firmware for Multirotor UAV Flight Controller Implemented on MCU MDR 32 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 345 Daniyar Wolf, Vadim Alexandrov, Dmitrii Shatov, Ilya Rezkov, Peter Trefilov, and Roman Meshcheryakov Autonomous Landing Algorithm for UAV on a Mobile Robotic Platform with a Fractal Marker . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 357 Dmitry Anikin, Artem Ryabinov, Anton Saveliev, and Alexander Semenov

xiv

Contents

Curl-Free Vector Field for Collision Avoidance in a Swarm of Autonomous Drones . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 369 Tagir Muslimov Author Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 381

Attention Guided In-hand Mechanical Tools Recognition in Human-Robot Collaborative Process Guo Wu(B)

, Xin Shen , and Vladimir Serebrenny

Bauman Moscow State Technical University, Moscow 105005, Russia [email protected]

Abstract. The task of recognition of human behavior in a collaborative robotic system is crucial for the organization of seamless and productive collaboration. We design a vision system for the industrial scenario for riveting a metal plate and concentrate on the task of recognizing in-hand mechanical tools. However, there is a severe occlusion problem during hand-object interaction process. Incorporating attention modules into the backbone part are often utilized to handle occlusion and enhance the ability of extract features with contextual information. In view of that, three modified occlusion-aware models based on YOLOv5 for in-hand mechanical tools recognition are proposed: by adding SimAM into each of bottleneck network in the backbone part, inserting a Criss-Cross attention layer between the last C3 block and the SPPF block of the back-bone network, and replacing the last C3 block of the backbone network with Criss-Cross attention layer. We create a dataset specifically for our task of in-hand mechanical tools recognition and validate four modified models after training separately, which proves the effectiveness of SimAM module and ineffectiveness of Criss-Cross attention module. The realtime detection is still imperfect under the occlusion of various directions of the hands. Keywords: Human Robot Collaboration · Hand Object Interaction · Occlusion-Aware Object Detection · Attention Mechanism

1 Introduction In the realm of human-robot collaboration (HRC), the ability of robots to understand and interact with human is crucial for seamless and efficient collaboration [1]. In order to implement HRC in a collaborative robot cell, the vision system not only needs to achieve safety assurance, but also requires the speculation of human intention and the tracking of human behavior. In some papers [1–3], people are in an open working space, and the human body is always moving, and the visual system recognizes and classifies the behavior of the human body. However, in some limited working spaces, especially in some industrial scenes, there is little physical movement of human body, and the work is mainly carried out with hands. Worker’s actions are dictated by tools held in hand, so this paper mainly focuses on recognizing them. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 1–12, 2023. https://doi.org/10.1007/978-3-031-43111-1_1

2

G. Wu et al.

We build a system for the industrial scenario for riveting a metal plate: between the manipulator and the worker there is a metal plate with holes drilled in it. The worker shows the mechanical tool. The vision system recognizes the tool and continues the dialogue if it’s the right one. Then the worker should touch the hole to be riveted with his finger. When the worker approves the outcome of the hole’s coordinate calculation, they signal their readiness to commence the task. The manipulator approaches the point. The vision system keeps track of the tools held in the hand while the manipulator is working. The manipulator stops working and drives off when the worker gives end gesture. Hand behavior includes making gestures, but there have been many excellent results on gesture recognition, including static gestures [4] and dynamic gestures [5], this paper will not go into details on how to perform gesture recognition. One fundamental aspect of this interaction is the recognition and understanding of mechanical tools that are commonly manipulated by humans during various industrial and manufacturing processes. Over the past few years, there has been a significant focus on the study of handobject interaction, which primarily revolves around comprehending and modeling the spatial relationships and interdependencies within the hand and tools during interaction. [6]. Whether it is cooking in an indoor kitchen [7], an outdoor daily activity [8], or an industrial scene [9], recognizing in-hand objects always poses several challenges. The first challenge is that hands and objects occlude each other. There has been a lot of work on hand pose estimation under various occlusions [10–12], but few papers have devoted to how to recognize objects in the hands under occlusions. The second challenge is that tools are always in a dynamic environment when manipulated by humans due to the complexity of tool shape and size, and the of variability orientation. While most mechanical tools are held in a grasped pose, the poses of hands and objects are constantly changing as industrial tasks progress. Commonly used handheld mechanical tools in industrial tasks may include wrenches, hammers, screwdrivers, pliers and drills. Traditional approaches to tool recognition have primarily based on object detectors, often dividing into two categories: two-stage detectors such as state-of-the-art Faster RCNN [13] and one-stage detectors including YOLO series [14–16]. However, due to severe hand occlusion and dynamic changes, performance of existing object detectors is still far from satisfactory. The attention mechanism, which was originally applied in the field of natural language processing, is now often used as an improvement method for the object detectors [17, 18]. The neural network model can enhance its ability to evaluate the significance of various sections of the input sequence by incorporating attention mechanism, enabling it to prioritize relevant information and make more precise and contextually aware predictions [17]. The fact that the hand’s position holds significant contextual information is motivating, making it worthwhile to further explore the enhanced method utilizing the attention mechanism. We focus on developing a robust, reliable and effective vision system that can accurately recognize in-hand mechanical tools in human-robot collaborative riveting process. The main contribution of this paper consists of three parts. Firstly, we propose three occlusion-aware models for in-hand mechanical tools recognition, which incorporates SimAM and Criss-Cross attention modules into the backbone part to enhance the ability of feature extraction and improve the accuracy of detection. Secondly, we create a

Attention Guided In-hand Mechanical Tools Recognition

3

dataset specifically for recognizing mechanical tools held in hand, which can be utilized for additional visual tasks. Thirdly, one of our improved models achieves superior performance when compared to the original object detector YOLOv5. The rest of this paper is organized as follows. In Sect. 2, we provide a brief review of the existing literature on hand-object interaction and occlusion-aware object detection. Section 3 details our proposed network architecture. In Sect. 4, we present dataset preparation, implementation details, experimental results and evaluation metrics. In conclusion, Sect. 5 concludes the paper by providing a summary of the study’s contributions and its overall importance.

2 Related Work 2.1 Hand Object Interaction Studying hand object interaction presents a difficulty as it requires simultaneous consideration of both aspects during the interaction, while the presence of self-occlusion between the hand and the object further complicates it. However, the focus of these studies varies from each other. For each aspect, they present several visual tasks such as detection, segmentation, action recognition, 3D pose estimation, and 3D reconstruction among these studies. This paper mainly concentrates on detection of object that is grasped by hands, which is almost equivalent to recognizing the behavior of hands or action of human. Nevertheless, the majority of the datasets employed in the study of hand-object interactions are inadequate for our particular task. For instance, EPIC-KITCHENS [7], the most commonly utilized dataset, primarily consists of ingredients and tableware objects. This makes our work become more difficult in applying transfer learning, as the source domain differs significantly from the target domain, which requires mechanical tools. Similarly, CoRe50 [8] contains only daily used handheld objects. MECCANO [9] focuses more on industrial parts rather than mechanical tools although these objects are used in industrial-like setting. Objects in WorkingHands [19] are exactly mechanical tools, but their annotations are utilized for the purpose of semantic segmentation analysis. Therefore, as mentioned above, we deliberate on creating a dedicated dataset. 2.2 Occlusion-Aware Object Detection Inter-class occlusion refers to the situation where objects are frequently hidden or obstructed by stationary elements or another type of objects [20]. Typically, the detector performs multi-scale processing and incorporates more anchor boxes based on the task’s complexity. However, despite these enhancements, they are insufficient in addressing the issue of occlusion. In the data preparation stage, augmentation techniques can be used to randomly occlude the annotated objects [21], also occluded instances can also be used for fine-tuning after training [22]. There are also some studies that use the GAN network to generate more training data [23] or recover the occluded regions after segmenting the object [24]. The occlusion problem is even more intractable in the segmentation tasks, among which appears a new direction amodal instance segmentation [25] proposed in

4

G. Wu et al.

recent years, aiming to identify invisible or occluded regions. Nonetheless, annotating dataset for amodal instances is a very time-consuming task. By integrating information from different modalities such as RGB images and depth maps, in the proposed multi-stream deep convolutional neural networks to recognize egocentric hand action. Utilizing contextual information proves highly beneficial, where in computer vision it refers to considering the surrounding area as well as the known shape or size of the occluded object. In [26] they firstly detect the localization of the hand and the center point of the in-hand object, and then classified this object after cropping the region of it. However, this two-stage sequential reasoning is slow and cannot achieve realtime performance. Attention mechanism or Transformer module is often adopted to incorporate spatial context, specifically in [25] features extracted from the region of hand intersecting with object are taken as context (key and value) to enhance the object features (query). In the case of video input data, temporal cues can also alleviate the problem of undetectable frames caused by occlusion. In [6–10] they proposed spatial position encoder and interaction unit, which presents multi-head transformers with hands as query and objects as key and value. Also, tracking methods help to maintain identity of object and hand, for example in [11] they used SORT to enhance the missing detections of hands. Due to these successful outcomes attained in the visual task of hand-object interaction, we aspire to explore more attention modules that is better suited for our needs.

3 Proposed Approach The proposed approach relies on implementing the YOLO model to detect objects. In terms of speed, the one-stage algorithms in the YOLO series outperform the two-stage algorithms found in the RCNN series. In addition, the most notable feature of YOLOv5 is that the model is very lightweight, specifically YOLOv5s model size is only 14 MB. The architecture of the YOLOv5 model is shown in Fig. 1. The Neck section incorporates the FPN and PANet modules, enhancing the variety and robustness of characteristics. The head part uses multi-scale detection headers, which is inherited from YOLOv3. The backbone part, which extracts general feature representations, adds CSPDarknet53 structure as an improvement. The attention mechanism allows the model to selectively capture relevant regions or features in the input image. Therefore, we assert that it would be more suitable to incorporate attention module within the backbone part. Typically, there are three different types in the attention mechanism: self-attention, spatial-wise attention, and channel-wise attention. Taking Vision Transformer (ViT) [27] for example, self-attention mechanism allows the model to build a larger global receptive field and capture global contextual information. On the contrary, we want to pay more attention to local information of hand-tool region. Utilizing channel-wise attention mechanism alone is also not sufficient to address occlusion issues, as it usually deals with correlation between different channels in the feature map. As an illustration, the SE module [17] enhances relevant features and diminishes irrelevant ones for the given task by modifying the attention weights of different channels in the feature map.

Attention Guided In-hand Mechanical Tools Recognition

5

Fig. 1. The architecture of the YOLOv5 model.

3.1 Criss-Cross Attention Module The model with spatial-wise attention mechanism modifies its attention weights based on the spatial location of the pixels. Through learning, the model can assign higher attention weights to areas surrounding occluded regions, enabling the model to prioritize its focus on the unoccluded regions in the image. Therefore, we resorted to widely well-known spatial-wise attention modules, specifically noting our interest in Criss-Cross attention module (CCA) [28]. The core mathematical principle of the CCA module in computer vision is to generate adaptive attention maps by performing crisscross attention operations, which capture long-range dependencies and model spatial relationships between pixels in an image, enabling the network to focus on important regions and improve segmentation accuracy. The crisscross operations mean that computes the similarity between each pixel and the pixels in its row and column (forming a cross pattern), and through two iterations of the same operation, indirectly calculates the similarity between each pixel and every other pixel. This approach reduces the spatial complexity from H × W to (H + W − 1), where H and W represent the height and width of the image, respectively. As showed in the following Fig. 2.

Fig. 2. Crisscross operation for pixels.

The CCA module introduces a lightweight and parameter-efficient design. It can be seamlessly integrated into existing convolutional neural network architectures without significantly increasing the model’s computational complexity or memory requirements. So, this gave us big motivation to have a try this module in our modified model. As

6

G. Wu et al.

shown in Fig. 3, given a local feature map H ∈ RC×W ×H , the module first applies two convolutional layers with 1 × 1 filters on H to generate two feature maps Q and K,  respectively, where {Q, K} ∈ RC ×W ×H .C  is the number of channels, which is less than C for dimension reduction. After obtaining Q and K, we further generate an attention is map A ∈ R(H +W −1)×(W ×H ) via Affinity operation. Then the contextual  +W information −1 collected by an Aggregation operation defined as follows. Hu = H A  + i,u i,u i=0 Hu , where Hu is a feature vector in H ∈ RC×W ×H at position, clearly the output gets the same size as the input, this is why it can be seamlessly integrated into existing CNN architectures.

Fig. 3. The structure of CCA module.

Therefore, we made the following two attempts, replacing the last C3 block of the backbone network with CCA and inserting a layer of CCA between the last C3 block and the SPPF block of the backbone network. For the convenience of recording, we will name the two improved methods as CCA1 and CCA2. 3.2 SimAM Attention Module There exist some attention modules such as CBAM [18] that sequentially combines two separate dimensions of channel and spatial. However, the two-step manner in CBAM takes too much calculation time. In [29] they proposed a novel straightforward method for generating 3d weights named SimAM, which stands for simple and parameter-free attention module.

Fig. 4. Estimation of full 3D weights of SimAM.

Figure 4 presents the process of estimation of full 3D weights for attention. Inspired by the concept of spatial suppression in neuroscience, the simplest implementation of

Attention Guided In-hand Mechanical Tools Recognition

7

finding these neurons that should be given higher priority is to measure the linear separability between a target neuron and other neurons. They define an energy function for each neuron: 2 2  1 M −1  y0 − xi , (1) et (wt , bt , y, xt ) = yt − ˆt + i=1 M 

where M is the total number of neurons on that channel. Adopting binary labels and adding regularizer, a closed-form solution was found, thus, the minimum energy is:   4 σˆ 2 + λ et∗ =  , (2) 2 t − μˆ + 2σˆ 2 + 2λ  2 1 M 1 M ˆ2 = M ˆ . These neurons groups all energy where μˆ = M i=1 xi and σ i=1 xi − μ functions across channels and spatial dimensions. The importance of these neurons is proportional to the inverse of the minimum energy, so it functions as 3D weight to enhance feature map of each pixel. Since SimAM does not increase number of parameters in the network, this module can be flexibly added in many places. Experiments show that the representation ability of the backbone network can be improved by SimAM. Consequently, we designed to add this attention module into each of bottleneck network in the backbone part, which is part of the C3 block. Modified bottleneck network is shown as Fig. 5.

Fig. 5. Modified bottleneck network with SimAM module.

4 Experiments 4.1 Dataset Preparation In our vision system, we need only five classes: screwdriver, hammer, wrench, drill and pliers. Employing a Python script of crawler, we collected 10000 images from Google and Bing to create our own dataset at the beginning, including five categories of in-hand mechanical tools mentioned above. The keywords we searched included the tool name

8

G. Wu et al.

and the tool name plus the words “hold in hand”, so that the dataset contains both tools with a clean background and images of them being held in the hand. At the same time, each tool in the dataset has various scales, from covering the entire image to almost invisible. Some pictures were not clear thus we removed them from our dataset. After our screening, only 9302 images are left. 70% of images are divided into the training set, 20% into the validation set and 10% into the test set. YOLOv5 model need to be trained on labeled data, for this, the Roboflow platform was used, which allows not only labeling images, but also exporting labeled data sets in YOLOv5 PyTorch format. YOLOv5 performs online augmentation during training, so before exporting dataset two preprocessing steps are simply applied: auto-orientation and resizing. 4.2 Implementation Details We chose the pre-trained weights from the small version of YOLOv5s model. To evaluate the impact of incorporating two attention modules into the YOLOv5 backbone, we trained YOLOv5 with SimAM and YOLOv5 with Criss-Cross attention module on our own tool recognition dataset. Additionally, we conducted an ablation experiment by training the original YOLOv5 network without the attention module on the same dataset. Before training, we specified the model hyperparameters: batch size 32, image size 640, initial learning rate 0.01 using SGD with 0.001 weight decay and a momentum of 0.9, and number of epochs 300, for the rest we used the default values, such as IOU threshold 0.45 and confidence threshold 0.25. Our experiments were conducted on an Ubuntu 20.04 with NVIDIA GeForce RTX 4060 laptop GPU with 8 GB memory. The camera to validate our model in real-time is Intel RealSense Depth Camera D435. 4.3 Analysis and Comparisons During the training process, we found that the original YOLOv5 converged around 200 epochs, while the loss value of the improved network with the attention mechanism dropped much slower. It can be seen that adding the attention mechanism will slow down the training speed. To obtain quantitative results, the models were tested on the validation set after training. To compare the results of three different deep learning models for mechanical tool detection, a set of standard metrics used to evaluate machine learning models was applied. Table 1 shows the standard indicators: precision, recall, [email protected] (average AP accuracy at IoU of at least 0.5) and [email protected]:.95 (average AP accuracy at threshold values IoU in the range from 0.5 to 0.95 in steps of 0.5). The number of frames that the model can process in one second (FPS) is also calculated separately. YOLOv5s + SimAM model has 1.9%, 0.2%, and 0.3% higher precision, [email protected], and [email protected]:.95 than original YOLOv5 respectively, but 1.1% lower recall than YOLOv5s, which is not that vital in our situation. At the same time, YOLOv5s + CCA1 model only has 0.9% higher precision, but 0.9%, 0,4%, 0.2% lower recall, [email protected], and [email protected]:.95. YOLOv5s + CCA2 model has 2.1%, 0.9%, 1.4% and 3.9% lower precision, recall, [email protected], and [email protected]:.95 than YOLOv5. All of three improved models fall short of the original YOLOv5 in terms of the FPS indicator. According to the table, it is evident that the two modified models incorporating

Attention Guided In-hand Mechanical Tools Recognition

9

the CCA attention module do not meet expectations. Meanwhile, the SimAM attention module demonstrates the ability to enhance the performance of the original YOLOv5 model. Table 1. Quantitative comparisons of four models. Model

Precision

Recall

[email protected]

[email protected]:.95

FPS

YOLOv5s

0.873

0.844

0.900

0.685

182

+SimAM

0.892

0.833

0.902

0.698

154

+CCA1

0.882

0.835

0.896

0.683

169

+CCA2

0.852

0.835

0.886

0.646

169

We perform real-time qualitative analysis with camera. In each experiment, the operator has to stretch out his arm and execute motions to continuously move or rotate one of the five mechanical tools within the camera’s view. Motivated by [11], we employ a readily available 2D bounding box tracker that utilizes Kalman filtering to extract tool object trajectories from the noisy detections obtained in each frame. Figure 6 shows screenshot of test results in the lab. The results of the three lines are respectively using the original YOLOv5, adding SimAM, and adding the CCA1 module. The confidence scores shown on images of the second row is generally a little higher than that of the first row. Even so, none of the four models can achieve real-time detection of mechanical tool. When the area occluded by the hand exceeds a threshold, tools, especially pliers, cannot always be detected. The possible reason is that the direction and position of hand occlusion in the data set are not diverse enough. If the data set is in the form of video, there will be better results.

10

G. Wu et al.

Fig. 6. Screenshot of test results in the lab.

5 Conclusion In this paper, we presented three modified occlusion-aware models based on YOLOv5 for in-hand mechanical tools recognition. The first modification is adding SimAM into each of bottleneck net-work in the backbone part. The second modification is inserting a Criss-Cross attention layer between the last C3 block and the SPPF block of the backbone. The third modification is replacing the last C3 block of the backbone with Criss-Cross attention layer. We created a dataset specifically for our task of in-hand mechanical tools recognition. Validation of four models proves the effectiveness of SimAM module and ineffectiveness of Criss-Cross attention module. Real-time detection is still imperfect if the occlusion caused by hands are too severe, and just a small part of tools could be visible. The outcomes of this research have significant implications for various domains, including manufacturing, assembly line operations, and industrial automation. Further research will be aimed at hand-object sematic segmentation considering the pose of hands.

References 1. Serebrenny, V., Lapin, D., Mokaeva, A.: The concept of an aircraft hull structures assembly process robotization. In: AIP Conference Proceedings, vol. 2171, no. 1 (2019) 2. Ranz, F., Hummel, V., Sihn, W.: Capability-based task allocation in human-robot collaboration. Proc. Manuf. 9, 182–189 (2017)

Attention Guided In-hand Mechanical Tools Recognition

11

3. Tsarouchi, P., Michalos, G., Makris, S., Athanasatos, T., Dimoulas, K., Chryssolouris, G.: On a human-robot workplace design and task allocation system. Int. J. Comput. Integr. Manuf. 30(12), 1272–1279 (2017) 4. Mazhar, O., Navarro, B., Ramdani, S., Passama, R., Cherubini, A.: A real-time humanrobot interaction framework with robust background invariant hand gesture detection. Robot. Comput.-Integr. Manuf. 60, 34–48 (2019) 5. Qi, W., Ovur, S.E., Li, Z., Marzullo, A., Song, R.: Multi-sensor guided hand gesture recognition for a teleoperated robot using a recurrent neural network. IEEE Robot. Autom. Lett. 6(3), 6039–6045 (2021) 6. Fan, H., Zhuo, T., Yu, X., Yang, Y., Kankanhalli, M.: Understanding atomic hand-object interaction with human intention. IEEE Trans. Circ. Syst. Video Technol. 32(1), 275–285 (2021) 7. Damen, D., et al.: Scaling egocentric vision: the epic-kitchens dataset. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 720–736 (2018) 8. Lomonaco, V., Maltoni, D.: CORe50: a new dataset and benchmark for continuous object recognition. In: Conference on Robot Learning, pp. 17–26 (2017) 9. Ragusa, F., Furnari, A., Livatino, S., Farinella, G.M.: The MECCANO dataset: understanding human-object interactions from egocentric videos in an industrial-like domain. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 1569–1578 (2021) 10. Nguyen, K., Todorovic, S.: A weakly supervised amodal segmenter with boundary uncertainty estimation. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 7396–7405 (2021) 11. Hasson, Y., Varol, G., Schmid, C., Laptev, I.: Towards unconstrained joint hand-object reconstruction from RGB videos. In: 2021 International Conference on 3D Vision (3DV), pp. 659–668 (2021) 12. Bambach, S., Lee, S., Crandall, D.J., Yu, C.: Lending a hand: detecting hands and recognizing activities in complex egocentric interactions. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1949–1957 (2015) 13. Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI) 39(6), 1137–1149 (2017) 14. Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 779–788 (2016) 15. Bochkovskiy, A., Wang, C., Liao, H.Y.M.: YOLOv4: optimal speed and accuracy of object detection. arXiv preprint arXiv:2004.10934 (2020) 16. Wang, C.Y., Bochkovskiy, A., Liao, H.Y.M.: YOLOv7: trainable bag-of-freebies sets new state-of-the-art for real-time object detection. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7464–7475 (2023) 17. Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7132–7141 (2018) 18. Woo, S., Park, J., Lee, J.-Y., Kweon, I.S.: CBAM: convolutional block attention module. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11211, pp. 3–19. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01234-2_1 19. WorkingHands Dataset. https://www3.cs.stonybrook.edu/~minhhoai/downloads.html. Accessed 01 June 2023 20. Saleh, K., Szénási, S., Vámossy, Z.: Occlusion handling in generic object detection: a review. In: 2021 IEEE 19th World Symposium on Applied Machine Intelligence and Informatics (SAMI), pp. 477–484 (2021)

12

G. Wu et al.

21. Zhong, Z., Zheng, L., Kang, G., Li, S., Yang, Y.: Random erasing data augmentation. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, no. 07, pp. 13001– 13008 (2020) 22. Zhou, C., Yuan, J.: Occlusion pattern discovery for object detection and occlusion reasoning. IEEE Trans. Circ. Syst. Video Technol. 30(7), 2067–2080 (2019) 23. Wang, X., Shrivastava, A., Gupta, A.: A-Fast-RCNN: hard positive generation via adversary for object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2606–2615 (2017) 24. Ehsani, K., Mottaghi, R., Farhadi, A.: SeGAN: segmenting and generating the invisible. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6144–6153 (2018) 25. Tang, Y., Wang, Z., Lu, J., Feng, J., Zhou, J.: Multi-stream deep neural networks for RGB-D ego-centric action recognition. IEEE Trans. Circ. Syst. Video Technol. 29(10), 3001–3015 (2018) 26. Lee, K., Kacorri, H.: Hands holding clues for object recognition in teachable machines. In: Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, pp. 1–12 (2019) 27. Dosovitskiy, A., et al.: An image is worth 16 × 16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020) 28. Huang, Z., Wang, X., Huang, L., Huang, C., Wei, Y., Liu, W.: CCNet: Criss-Cross attention for semantic segmentation. In: Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pp. 603–612 (2019) 29. Yang, L., Zhang, R.Y., Li, L., Xie, X.: SimAM: a simple, parameter-free attention module for convolutional neural networks. In: International Conference on Machine Learning (ICML), pp. 11863–11874 (2021)

Design and Implementation of a Multimodal Combination Framework for Robotic Grasping Congyu Huang , Ziyang Wang , Haoran Zhu , Jie Li , and Xiaofeng Liu(B) College of IoT Engineering, Hohai University, Changzhou 213100, China [email protected]

Abstract. Robotic grasping plays a crucial role in manipulation tasks. However, due to the complexity of human-robot interaction, service robots still face significant challenges in handling task-oriented operations in real-world environments. To address this issue and better meet practical interaction needs, we propose a multimodal combination framework for robotic grasping. It leverages language texts to facilitate communication and detects and grasps target objects based on point clouds and feedback. The framework comprises several multimodal components, including ChatGPT, stereo cameras, and wearable devices, to complete instruction processing, grasp detection, and motion execution. To enable effective interaction, ChatGPT facilitates basic communication and responds to instructions between humans and robots. Additionally, the robot can detect the 6-DoF grasp of objects based on point clouds obtained by stereo cameras. These grasps are combined with the feedback provided by ChatGPT to further meet the requirement from human. Finally, we utilize wearable devices to teach robots generalized motor skills. This enables the robot to learn corresponding movements and perform them effectively in various scenarios, further improving its manipulation abilities. The experimental results from simulated conversations and real-scene tasks highlight that our proposed framework provides logical communication, stable grasping, and effective motion. Keywords: Robotic Grasping · Point Cloud · ChatGPT · Wearable Device · Multimodal combination · Human-Robot Interaction

1 Introduction As the aging population grows and the labor shortage intensifies, service robots are increasingly demanded to enter households to provide further assistance to humans [1]. Robotic grasping is a kind of basic skills that is almost the first step of all manipulation tasks. The service robots should be able to grasp the corresponding target object based on human demands. In previous studies on robotic grasping, Pas et al. [2] proposed that robotic grasping can be divided into two subprocesses: grasp detection and motion execution. Robots with multimodal information processing capabilities can achieve more effective interaction with human. Thus, service robots that can understand language text and combine it with visual information can better execute commands from humans and perform corresponding operations. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 13–22, 2023. https://doi.org/10.1007/978-3-031-43111-1_2

14

C. Huang et al.

Recent studies in natural language processing have demonstrated that large language models (LLMs) possess powerful reasoning abilities. For instance, Kojima et al. [3] proposed Zero-shot-CoT, which enables LLMs to elicit chains of thought across various reasoning tasks. Similarly, Madaan et al. [4] utilized LLMs to generate structured commonsense and convert them to Python code via COCOGEN, presenting a promising direction for structural commonsense reasoning. These studies have opened up new avenues for research on robotic systems [5]. By combining LLMs with robots, they can acquire fundamental commonsense knowledge and comprehend language instructions. Moreover, ChatGPT has displayed impressive performance in communication, reasoning, and computation. If these capabilities can be implemented in service robots, it will significantly enhance their understanding and expression abilities. Previous studies have extensively explored robotic grasping, yet robots still lack basic knowledge and cannot directly obtain grasp postures from visual information. To address this gap, Liang et al. [6] proposed PointNetGPD, which employs PointNet [7] to capture point features and analyze grasp qualities from point clouds. Similarly, Ni et al. [8] utilized PointNet++ [9] to design an end-to-end spatial grasp generation method for sparse point clouds, which predicts poses, categories, and scores of grasps directly. Despite these advancements, neither approach considered task-oriented options and they cannot provide a basis for grasping special target objects. For effective humanrobot interaction, robots must be able to take objects according to the specific needs of humans. Despite advances in robotics, there are still challenges in motion control due to limitations in algorithms and other factors. Learning from Demonstration (LfD) [10], also known as Imitation Learning (IL), provides a promising direction for robots. With the help of LfD, robots can learn from human motion demonstrations and quickly master various skills. To grasp objects, robot arms must follow specific trajectories while navigating environmental conditions to reach the intended location. Grasp Understangding

Motor Skill Learning & Executation

6-Dof Grasp Detection Camera

I'm a bit hungry. Please give me the apple on the table.

Inerial Sensor Node Decision Making Motion Trajectory

Human

Okay, please wait a moment. I'll get the apple for you right away.

Point Set Processing

Robotic Motion Planning

Wearable Device

Thank you!

ChatGPT Human

Fig. 1. Overall design and multimodal structure of the framework.

Based on the above-mentioned content, this paper proposes a multimodal combination framework for robotic grasping, as shown in Fig. 1. The framework leverages ChatGPT to bridge the gap between humans and robots by utilizing the text processing capabilities of large language models for effective understanding of command requests and feedback responses. Moreover, to obtain the grasp pose of objects, point clouds

Design and Implementation of a Multimodal Combination Framework

15

are utilized to provide sufficient visual information for optimal grasp detection based on robot’s hardware configuration. Finally, human body movements are captured with wearable devices and algorithms are used to teach robots smoother movement skills thereby enhancing their efficacy and efficiency of operations.

2 Grasp Understanding and Detection The process of grasp generation integrates information from language and vision, providing robots with effective references for practical operations. For instance, once receiving instructions from humans, robots should effectively understand the language and provide feedback. Moreover, the images captured by the camera can provide robots with the distribution of objects in real-world scenes. Subsequently, grasping detection based on visual information can be performed to obtain the grasping pose of each object. 2.1 Language Text Feedback Undoubtedly, natural language processing has provided significant support for robot research. As a large language model, ChatGPT has demonstrated powerful performance in common sense understanding, reasoning, and computation. In this article, we consider incorporating ChatGPT into the robot system in the context of grasping. As shown in Fig. 2, we have built a chat node for ChatGPT using Robot Operating System (ROS) to facilitate language communication between humans and robots. The chat node will serve as a relay station to send the instruction text from humans to ChatGPT and receive the output from ChatGPT. Afterwards, the robot should receive the feedback sent by the chat node and integrate it with the grasping detection results as the basis for motion execution. We believe that by integrating ChatGPT into the robot system, more flexible and smoother human-robot interaction can be achieved. Human Robot

Instruction Texts Grasp Detection Target Response & Execution

ChatGPT Chat Node

Feedback

Fig. 2. Structure of grasp understanding module combined with ChatGPT.

2.2 6-DoF Grasp Generation Figure 3 illustrates the structure of the grasp detection module. The original point cloud captured by the camera often includes excessive noise and irrelevant information. Therefore, preprocessing such as filtering, down sampling, and segmentation is necessary to

16

C. Huang et al.

extract object points from the raw point cloud. Subsequently, robots should conduct grasp sampling on the segmented object points to obtain a series of candidates. During sampling, it is crucial to select the correct robot configuration, including the appropriate gripper type, and follow specific grasping strategies. Each candidate represents a 6-DoF grasp and should be transformed for use in the following stages. To meet the requirements of the evaluation network, the point cloud corresponding to each grasp candidate will be used as training input. By allowing the network to learn the features of the point cloud, it can correlate different grasp qualities and output their respective scores. Thereby, the robot can select the highest-scoring grasp candidate for each object as the optimal grasp. Robot Camera

Languages Grasp Sampling Chat Node

Point Cloud Processing

Quality Evaluation

Optimal Grasp

Motion Module

Fig. 3. Structure of the grasp detection module.

3 Robotic Motor Skills Generalization Despite advancements in robot technology, there are still limitations in their motion skills, particularly in achieving compliant control. To enhance a robot’s motion capabilities, we have incorporated the idea of Learning from Demonstration, utilizing wearable devices to capture human motion trajectories. By learning from these motion features, robots can better understand human movement and improve their ability to interact with humans efficiently and seamlessly. 3.1 Human Motion Demonstration Without a doubt, humans are the best teachers for robots. Throughout the long process of evolution, humans have mastered a wealth of motion skills, and are able to proficiently utilize their limbs to perform various actions. By means of human-robot teaching, robots can quickly acquire a large number of skills.

Design and Implementation of a Multimodal Combination Framework

17

The motion capture system framework is shown in Fig. 4, which primarily comprises wireless sensor nodes that are worn on different human body joints. Each node is equipped with a low-cost MEMS inertial sensor that can record the posture of the joint during movement. Due to the possibility of errors in the raw data from different nodes, it is necessary to carry out error calibration and implement suitable posture algorithms to enhance the quality of the data. Finally, we utilized a multi-sensor data fusion algorithm to merge the data acquired from various joints. Inerial Sensor Node

Robot

Motion Trajectory

Motion Learning

Inertial Data Acquisition Motion Modeling Error Calibration Real-time Mapping Attitude Algorithm

3D Posture Tracking

Fig. 4. The framework of motion capture system.

We developed a human motion model based on multiple joint nodes and performed real-time mapping processing. In light of the robotic arm’s grasping requirements, we primarily captured and processed the hand and arm’s motion trajectories, which could be utilized for robot motion learning, as shown in Fig. 5.

a) Human Body Movement

b) Motion Capture

Fig. 5. Results of human motion capture.

3.2 DMP and Obstacle Avoidance DMP is a trajectory modulation method that relies on dynamic systems and decomposes actions into a sequence of primitives. It adjusts the motion primitives’ weights to accomplish the trajectory fitting. DMP inherits many benefits from linear dynamic systems,

18

C. Huang et al.

such as conditional convergence, robustness to disturbances, and time independence. Additionally, nonlinear external forces can be incorporated to represent smooth motion. To improve the robot’s motion capabilities, we utilized the DMP algorithm to enable it to learn and generalize human motion trajectories. Our framework also includes an obstacle avoidance module. Obstacle avoidance is a crucial and difficult task during robotic arm operations. In complex and unpredictable environments with various obstacles, robots must be capable of autonomously planning their path and avoiding obstacles to complete tasks efficiently while maintaining safety and stability. Therefore, developing efficient obstacle avoidance algorithms and systems is a significant and challenging area of research, design, and development for robotic arms. Ginesi et al. [11] proposed an autonomous, efficient, and safe method for avoiding obstacles that utilizes potential functions to calculate the trajectory of obstacle avoidance. They introduced a velocity-dependent potential function that allows for smoother obstacle avoidance behavior compared to static (i.e., position-dependent) potentials, and the algorithm has demonstrated excellent performance in obstacle avoidance tasks.

4 Experiments To evaluate the performance of our framework, experiments for evaluating the grasp understanding, grasp detection, and robotic motor skills have been conducted in this section. During the test, a UR3 robot equipped with a two-finger gripper and a D455 RealSense camera is used to validate the framework. Table 1. Conversation results of ChatGPT under grasping condition. Human

ChatGPT

I am a bit hungry, please give me the banana on the table

Certainly, I can give you a banana. Please hold on for a moment as I retrieve one for you

Please give me a bottle of water

Okay, I will pour you a bottle of water. Please be careful not to spill it

Hello, I’m about to go out. Can you give me a Sure, I can do that. Here’s the mask, I’ll hand it mask? to you Thank you very much for your help!

You’re welcome, I’m always here to assist you. If you have any other questions or need help, please feel free to let me know

I am going to wash my face, can you give me I’m sorry, there’s no facial cleanser on the the facial cleanser on the table? table. Do you need me to help prepare a bottle of facial cleanser for you? Okay, thank you

I really appreciate your understanding. What kind of facial cleanser would you like? Do you prefer one for oily, normal or dry skin?

Design and Implementation of a Multimodal Combination Framework

19

4.1 Evaluation of Robotic Grasp Understanding To evaluate the robot’s understanding of human language commands in grasping scenarios, we conducted various dialogue tests on ChatGPT. To meet the experimental requirements, we made some settings for ChatGPT and required it to interact with users. It should be noted that we did not set the content of the conversation, and only required ChatGPT to simulate a conversation with a human as a service robot, while providing partial visual information for reference. Part of the conversation results can be seen in Table 1. 4.2 Evaluation of Grasp Detection To assess the detection performance of our proposed framework, we selected various common objects found in everyday life, such as apples, bowls, and cookies, which vary in shape, type, and color. The objects were randomly placed on a table with different orientations and positions. As depicted in Fig. 6, the optimal grasp output by the grasp detection for each object is represented by the blue gripper. As we can see, our framework can provide stable grasping for different objects regardless of changes in object state and position.

a) Surgical Mask

c) Shampoo

b) Medicine Box

d) Trash Bag

Fig. 6. Detection results in single-object scenes. (Color figure online)

Considering practical grasping requirements, detecting grasping in multi-object scenes has become increasingly common. Additionally, as the number and variety of

20

C. Huang et al.

objects increases, the difficulty of grasp detection also increases. In this case, we designed various scenarios and continuously increased the types and quantities of objects to test the performance of our framework. The detection results of our framework in clutter scenes can be found in Fig. 7. The results show that our framework can meet the robot’s grasping requirements in various scenarios, greatly improving the effectiveness of human-robot interaction.

a) Toothpaste, apple and mango

c) Beer, kiwi, facial cleanser, etc.

b) Cookies, cola, bowls, etc.

d) Potato chips, pear, cookies, etc.

Fig. 7. Detection results in clutter scenes.

4.3 Evaluation of Robotic Trajectory Learning To assess the motion module’s performance, we utilized a motion capture device to record human motion trajectories and enabled the robot to learn from them. In order to generate a similar trajectory, we input the initial trajectory into DMP for learning and adjusted the starting and target points, as depicted in Fig. 8. By modifying the external force term, we could regulate the trajectory’s shape and guide the robotic arm to write numbers of varying shapes based on the modified starting and ending points. As shown in Fig. 8, we conducted experiments in both simulated and real environments and achieved satisfactory outcomes. In order to verify the effectiveness of our obstacle avoidance algorithm, we designed corresponding motion scenarios for the robot and set up obstacles along its path of movement, as shown in Fig. 9. The red trajectory represents the original path that would have resulted in a collision with the obstacle. In contrast, the black trajectory indicates the

Design and Implementation of a Multimodal Combination Framework

a) Human body trajectory

b) DMP calculation results

21

c) Robot motion trajectory

Fig. 8. Results of robot trajectory learning.

path after obstacle avoidance, which ensures safer execution of the robot arm’s grasping and delivery tasks.

Fig. 9. Results of Robot Obstacle Avoidance Motion. (Color figure online)

5 Conclusion This paper presents a multimodal combination framework for robotic grasping. Specifically, we employed ChatGPT as the bridge between humans and robots, inspired by the powerful performance of LLMs. With the powerful reasoning ability of ChatGPT, the robot is capable of effectively understanding human language commands and providing positive feedback. Additionally, we designed a visual detection module for the robot to process point cloud information, enabling it to reliably determine grasping poses for target objects. The results showed that our method performed well both in single-object scenarios and in more complex cluttered environments. Finally, we utilized the Learn from Demonstration approach, where human motions were captured using wearable devices and effectively used to teach the robot new skills. Future research could focus on further perfecting the multimodal combination framework, as this study only provided a basic application of ChatGPT without integrating a

22

C. Huang et al.

real embedded artificial intelligence system. Such a system would offer a more comprehensive and autonomous large language model for robots. In addition, we are considering using wearable devices to teach robots a wider range of more complex motion skills. Acknowledgments. This work was supported in part by the National key R&D program of China 2018AAA0100803, in part by the National Natural Science Foundation 62276090 and 62203150, in part by the Key Research and Development Program of Jiangsu under grants BK20192004B, in part by the China Postdoctoral Science Foundation under Grant 2021M701051, in part by the Jiangsu Province Excellent Post-doctoral Program 2022ZB192, in part by the Changzhou Basic Research Program (Application Program) CJ20220051, in part by the Guangdong Forestry Science and Technology Innovation Project under grant 2020KJCX005.

References 1. Suwa, S., et al.: Home-care professionals’ ethical perceptions of the development and use of home-care robots for older adults in Japan. Int. J. Hum.-Comput. Interact. 36, 1295–1303 (2020) 2. ten Pas, A., Gualtieri, M., Saenko, K., Platt, R.: Grasp pose detection in point clouds. Int. J. Robot. Res. 36(13–14, SI), 1455–1473 (2017) 3. Kojima, T., Gu, S.S., Reid, M., Matsuo, Y., Iwasawa, Y.: Large language models are zero-shot reasoners, pp. 2205–11916 (2022) 4. Madaan, A., Zhou, S., Alon, U., Yang, Y., Neubig, G.: Language models of code are few-shot commonsense learners. arXiv e-prints, 2210–07128 (2022) 5. Liang, J., et al.: Code as policies: language model programs for embodied control. arXiv e-prints, 2209–07753 (2022) 6. Liang, H., et al.: PointNetGPD: detecting grasp configurations from point sets. In: 2019 IEEE International Conference on Robotics and Automation (ICRA), pp. 3629–3635 (2019) 7. Charles, R.Q., Su, H., Kaichun, M., Guibas, L.J.: PointNet: deep learning on point sets for 3d classification and segmentation. In: 30TH IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2017), pp. 77–85 (2017) 8. Ni, P., Zhang, W., Zhu, X., Cao, Q.: PointNet++ grasping: learning an end-to-end spatial grasp generation algorithm from sparse point clouds. In: 2020 IEEE International Conference on Robotics and Automation (ICRA), pp. 3619–3625 (2020) 9. Qi, C.R., Yi, L., Su, H., Guibas, L.J.: PointNet plus plus: deep hierarchical feature learning on point sets in a metric space. In: Advances in Neural Information Processing Systems 30 (NIPS 2017), vol. 30 (2017) 10. Ambhore, S.: A comprehensive study on robot learning from demonstration. In: 2020 2nd International Conference on Innovative Mechanisms for Industry Applications (ICIMIA), pp. 291–299 (2020) 11. Ginesi, M., Meli, D., Calanca, A., Dall’Alba, D., Sansonetto, N., Fiorini, P.: Dynamic movement primitives: volumetric obstacle avoidance. In: 2019 19th International Conference on Advanced Robotics (ICAR), pp. 234–239 (2019)

Experimental Validation of an Interface for a Human-Robot Interaction Within a Collaborative Task Maksim Mustafin1(B) , Elvira Chebotareva1 and Evgeni Magid1,3

, Hongbing Li2

,

1 Intelligent Robotics Department, Institute of Information Technology and Intelligent Systems,

Kazan Federal University, 35, Kremlyovskaya st., Kazan 420111, Russia [email protected] 2 Department of Instrument Science and Engineering, Shanghai Jiao Tong University, Minhang, Shanghai 200240, China 3 Tikhonov Moscow Institute of Electronics and Mathematics, HSE-University, 34, Tallinn st., Moscow 123458, Russia

Abstract. This paper presents a prototype of a non-contact UR robot based Virtual Control (UR-VC) system for collaborative robots of the UR family, which is based on computer vision techniques and a virtual interaction interface. A control method involved specific hand movements within a field of view of a web camera, which was connected to a laptop with the running UR-VC system. We present the UR-VC system and the results of an experimental validation. To inquire if the UR-VC system is comfortable and user-friendly for an interaction with collaborative robots and to study opportunities for a further development and expansion directions of the system, we designed a test case that simulates a joint product assembly in a collaborative workspace. The constructed collaborative workspace included the UR3e robot, the laptop with the running UR-VC system and assembly parts for a collaborative task. 24 participants were involved in the experiments. First, the participants learned how to control the robot using the UR-VC system. After the training, all participants successfully controlled the robot using the proposed interface for performing the collaborative task. Participants’ experience of operating the robot was analyzed via surveys, their unconstrained comments and video recordings of the experiments. Keywords: Human-Robot Interaction · Human-Robot Collaboration · Collaborative Assembly · Virtual Control

1 Introduction Currently, collaborative robotics has a great potential for application in industry and manufacturing [1, 2]. Automation with robots can significantly improve quality, safety, and efficiency of production processes [3]. However, full automation of processes can be difficult or impossible for various reasons [4]. Some stages of production may not be © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 23–35, 2023. https://doi.org/10.1007/978-3-031-43111-1_3

24

M. Mustafin et al.

automated and require a human intervention [5]. Additionally, full automation may be infeasible due to a high cost and complexity of an implementation. These are particularly relevant for small and medium-sized industries. In the latter case, a production process can be arranged in such a way that production steps can be shared between a human and a robot working collaboratively in a shared workspace [6]. An example would be a process that involves a collaborative assembly or processing of a product, where the product or its parts are passed alternatively between a human and a robot. Human-robot collaboration (HRC) implies an existence of one or more communication methods between a human and a robot. A robot control system can process verbal and non-verbal operator signals [7] and may rely on speech, gesture, and gaze recognition, tactile control, or multimodal interfaces [8]. In some cases, the most convenient method for controlling a robot involves a non-verbal communication based on operator’s gestures and movements. This approach enables both simple and complex interactions between an operator and a robot and helps to integrate robots into existing workflows. In this paper, we present a new virtual control system based on computer vision and augmented reality (AR) techniques for non-contact control of a collaborative manipulator during joint assembly tasks. An experimental validation of the system demonstrated a successful HRC during a joint assembly task.

2 Related Work Overviews of modern collaborative robots (cobots) used in industry and service fields demonstrated a broad variety of approaches and particular applications [9, 10]. Design issues of cobot control systems’ reviews focus on existing sensor-based control methodologies [11] and consider general issues in the management of cobots [12]. A special place among cobot control systems is occupied by AR-based methods, which are a promising direction in industrial robotics. Costa et.al. [13] stated that replacing a purely manual control with a collaborative scenario using AR reduces a production cycle time and improves an operator’s ergonomics and identified four types of user interfaces: head-mounted displays (HMD), projector-based interfaces, hand-held displays (HHDs), and Fixed Screens. They noted that HMDs and projector-based interfaces are used much more frequently compared to HHDs and Fixed Screens in research and emphasized that a usage of HMDs for AR in collaborative robotics may be hindered by hardware aspects, such as a narrow field of view, occlusions, and weight, which may have a negative impact on an operator’s sense of safety. We believe this implies an emergence of risks associated with a negative impact on a user experience (UX) while a positive UX in human-robot interaction (HRI) is essential for an efficient organization of HRC processes [14]. When developing a contactless control interface for a cobot during a joint assembly, maintaining a balance between safety, efficiency, and ergonomics in a design of collaborative assembly processes is important [15], as well as design recommendations based on international standards, research, and real-world use cases [16]. Typically, performing joint assembly and processing tasks requires a human to perform some work manually, with their hands. As a result, for a collaborative assembly and processing, contactless methods of a robot control that do not necessitate a constant presence of

Experimental Validation of an Interface for a Human-Robot Interaction

25

operator’s hands on a control panel are preferred. One of these methods is controlling a robot using gestures. A significant part of modern gesture-based control of cobots relies on methods for classifying and recognizing gestures using machine learning techniques, including such particular examples as learning semantics in experiments with a gesture-based control system in a collaborative assembly task [17], a new taxonomy for gestures classification [18], an online static and dynamic gesture recognition framework for HRI [19], a robothuman interface [20] based on MediaPipe solution [21], and others. In our previous work [22], we conducted a series of pilot experiments on gesturebased control of UR5e robot in a collaborative assembly task. The experiments revealed the general user satisfaction with the contactless control method using gestures, however, they identified a number of disadvantages of this approach. The first issue was a necessity to select a universal gesture system for the robot control. Even though all users successfully employed the gestures we had proposed for the control, some users noted that particular gestures were not quite familiar to them. In light of this, we encountered a challenge of fine-tuning a command set according to users’ preferences. Simultaneously, expanding the gesture vocabulary requires additional research, which may not necessarily guarantee a development of a universal set of gestures that accommodates all users’ preferences. The first disadvantage was a necessity to employ all fingers, which can affect a biomechanical load on a user’s hand, their comfort level, and focus. As a number of commands increases, a user must not only operate different hand joints but also memorize all the commands. Considering the abovementioned literature analysis and our own experimental experience, in order to develop a new interface for a virtual robot control system we abandoned the gesture-based approach in favor of a mixed method that involves AR elements and a single gesture of closing a thumb and a forefinger.

3 Materials and Methods This section overviews our virtual control interface concept, a robot control system architecture, and a workcell configuration. Additionally, we describe a collaborative task that was used for the system testing. Using our previous research as a starting point [22], we aimed to develop a new computer vision-based method for interaction and control of a cobot. The new approach was designed to be adaptable to a wide range of users and scalable to future needs, including new functionalities and features integration. It was important to develop an application, which does not generate haptic feedback but provides a feedback to a user via audio (application sounds) and visual signals (interface appearance changes). The use of a contactless control of a cobot through a virtual interface during collaborative assembly tasks was supported by a number of arguments. Firstly, the contactless control reduces a biomechanical load on an operator; for example, when an operator controls a cobot using a teach pendant, the operator needs to hold it in a hand, which causes an arm muscles fatigue. Secondly, if the operator’s hands are dirty, the contactless control prevents a further contamination of work area surfaces (the teach pendant, objects within the cobot workspace); thus, the contactless control allows operator’s working environment to stay clean and tidy for a long time. Thirdly, to control the cobot with the teach

26

M. Mustafin et al.

pendant the operator needs to devote some time learning and practicing pendant’s capabilities. Therefore, a user-friendly and intuitive application that uses computer vision and simple interaction commands (which may also include all functions of the teach pendant) will optimize time and efforts of the operator. 3.1 UR Robots – Virtual Control Application UR robots – Virtual Control (UR-VC) application was programmed in Python3 and uses Pygame and Playsound libraries at the frontend. The Pygame was used to draw and animate interface elements. The Playsound was used to play predefined sounds when an operator selects a button or presses a button. The backend of UR-VC application employed CVZone, MediaPipe, OpenCV, and NumPy libraries for hand detection and data processing. The UR-VC application User Interface (Figs. 1 and 2) contains the following elements: 1. 2. 3. 4. 5. 6. 7. 8. 9. 10.

11.

A current robot program state (takes values “Playing”, “Paused” or “Stopped”). A last command of a user (which button was clicked). UR robot responses to user’s commands (UR log). “E-STOP” (Emergency Stop) button – a user can stop the robot immediately, which ends an execution of a current robot program. “Power On” button turns on the robot. “Play” button launches a robot program. “Pause” button pauses a robot program. A main cursor is located at a fingertip of a user’s index finger in the interface of UR-VC. The main cursor allows a user to select any button. A clicking cursor is located at a fingertip of a user’s thumb in the interface. The user can click any button using this cursor and together with the main cursor. A progress bar for clicking is designed to indicate a remaining time, which a user should keep his/her pointing fingers together in order to produce a button click. The bar was designed to exclude accidental and unintentional clicks. “Next detail” button is responsible for sending the robot a command to proceed in order to assembly the next product (a fidget spinner). The button appears after clicking “Power On” button, waiting the robot to turn on and loading its program.

Experimental Validation of an Interface for a Human-Robot Interaction

27

Fig. 1. UR-VC application User Interface.

Fig. 2. Using UR-VC application showcase.

3.2 Workcell Description A designed workcell included a robot table and a small additional table where necessary for experiments items were set up. The robot table allows placing objects on its top and fixing them with screws. The workcell included the following items (Fig. 3): 1. The UR3e manipulator; this collaborative robot has a small size, which improved participants’ safety and left more free space within the workcell. 2. The starting location of the robot (marked with a A4 paper sheet with “Start” label); this was a location where the robot moved after launching a loaded program.

28

M. Mustafin et al.

3. A waiting location of the robot for a command from the operator (marked with a A4 paper sheet with “Waiting for the “Next detail” command” label); this was a location where the robot moved after assembling one product and waited a user to click “Next detail” button in UR-VC application. 4. A felt-tip pen in the robot’s gripper was used for drawing squares by the robot during the training stage for the participants. 5. Assembly parts, fidget spinner frames, were used for assembling the product (the fidget spinner) during the main stage. 6. A plastic mold for the fidget spinner frame for the collaborative assembly was a location where the robot placed a fidget spinner frame to assembly one product. 7. Assembly parts, 3D printed plastic bearings in a pallet, were used for assembling the product during the main stage. 8. A sheet of paper was attached to a robot’s desktop for drawing squares on it during the training stage for the participants. 9. Felt-tip pens for participants for the training stage; the participants used these felttip pens for drawing on the paper inside the squares (that were drawn by the robot) during training stage. 10. A laptop with the UR-VC application running on it. 11. An experiment instruction listed steps for the participants to complete all stages of experiment. In addition, the workcell contained a siting chair, which participants could use at their will.

Fig. 3. The experimental workcell.

3.3 Collaborative Task Collaborative assembly task is to assemble the product, the fidget spinner, using the UR3e robot. An operator takes the fidget spinner frame and puts it in the plastic mold

Experimental Validation of an Interface for a Human-Robot Interaction

29

for the robot. Next, operator waits for the robot to insert four plastic bearings into the fidget spinner frame (Fig. 4). In Fig. 5, the assembly parts and the finished product are presented. The proposed in this paper collaborative assembly task is much simpler than the original task from our previous work [22]. For example, screws tightening by the UR5e robot in the original task made the experiment process rather long and difficult. Therefore, this time we intentionally simplified the original task to allow a participant concentrating on a developed UR-VC system evaluation rather than on the task complexity.

Fig. 4. The process of assembling the fidget spinner with the UR3e robot.

Fig. 5. Assembly parts and the finished product.

4 Experiment Description This section describes the experimental setup aimed at testing the developed interface. The experiments had two stages: a training stage to teach the participants operating the system and the main experimental stage that was performed in order to evaluate the proposed system. 4.1 Training Stage In total, the experiments involved 24 participants (Fig. 6): four laboratory members and 20 not professional robot operators. The participants were divided into two groups, which

30

M. Mustafin et al.

differed by dates of experiments. In the first stage (a learning stage) of the experiment, the participants learned to control the robot via the UR-VC application. The learning stage consisted of two parts: theoretical and practical.

Fig. 6. All participants of the experiments.

In the theoretical part, the participants watched a prerecorded video, which taught basics of working with the UR-VC application. In the practical part the participants had to apply the obtained theoretical knowledge (of the theoretical part) in order to develop basic practical skills of working with the robot via the UR-VC application and get acquainted with a concept of a collaborative assembly by completing two tasks. These two tasks required a sequential execution of steps (clicking the UR-VC buttons in a specific order and explaining outcomes of the clicks), which were described in the instruction. The first task allowed the participants to interact with the UR-VC application for the first time and to understand a correspondence of the robot’s movement and the user commands. The second task was a simple collaborative task, in which the robot drew squares one by one with the felt-tip pen at a command of a participant, and the operator drew a number inside the square. A fragment of the participant training experiment is presented in Fig. 7.

Experimental Validation of an Interface for a Human-Robot Interaction

31

Fig. 7. A participant of the experiment during the training stage.

4.2 Main Stage A main stage or a collaborative assembly task stage was built around a comparison of a manual and a collaborative assembly of products by the participants. At a beginning, the participants need to assemble five products manually. Then they needed to assemble five fidget spinners with the UR3e robot using the UR-VC application. A participant took the fidget spinner frame and put it in the plastic mold for the robot. After that, the participant clicked “Next detail” button in the UR-VC application and waited for the robot to insert four plastic bearings into the fidget spinner frame. When the robot was done, the participant took out the finished product and put it in a special box. All these steps were listed in the instruction. Additionally, during the collaborative assembly, the participants had to completely stop the robot program execution and move the last four bearings in a certain way so that the robot could immediately begin assembling the last fidget spinner after the program was launched. A fragment of the main part of the experiment is presented in Fig. 8.

Fig. 8. A participant of the experiment during the testing stage.

32

M. Mustafin et al.

4.3 Evaluation After the experiment, the participants were asked to take a survey that consisted of seven questions (Table 1). Additionally, the experiments were recorded and we could postprocess the videos in order to evaluate all informal comments of the participants. Table 1. The survey questions. Number Question Q1

Was it easier for you to assembly the products with the robot assistance than without the robot?

Q2

How accurately did the robot execute your commands?

Q3

How quickly did the robot respond to your commands?

Q4

How comfortable were you while working with the robot?

Q5

How good did the robot perform its task?

Q6

What disadvantages in the robot and the UR-VC app operation could you note (if any)?

Q7

What changes, improvements, innovations would you like to propose for the UR-VC application?

We used a 5-point Likert-type scale for questions Q1-Q5 and a free form for questions Q6 and Q7. During the participants’ behavior observation, we evaluated the following factors: time spent for training (average ~16 min), time spent for performing the main collaborative task (average ~13 min), comments made during interactions with the robot and the users’ reactions to unexpected situations.

5 Results All participants successfully coped with the task of controlling the robot using the URVC application. Figure 9 presents the results of the survey with questions Q1–Q5. The x-axis (horizontal axis) represents the questions’ number, while the y-axis (vertical axis) shows the number of participants who selected a specific answer option. In general, the participants noted a positive experience of using the application and did not notice any significant disadvantages. However, some participants pointed out that the robot could work faster. One participant complained that it was not very convenient to turn around and click the buttons in the UR-VC after each operation with the robot. Another participant reported that sometimes it was inconvenient to control the UR-VC with a single hand. The participants made valuable suggestions on the UR-VC application potential improvements, including adding a support for controlling the robot with both hands, adding a controlling hand selection, adding voice commands for emergency stop, enabling tracking of operator’s performance of their part of the work in a collaborative assembly, and selecting a robot’s operating speed mode.

Experimental Validation of an Interface for a Human-Robot Interaction

20

1

15

2

10

3

5

4

0

33

5 Q1

Q2

Q3

Q4

Q5

Fig. 9. The survey results. Avg.: Q1 – 4.5, Q2 – 4.6, Q3 – 3.8, Q4 – 4.2, Q5 – 4.7.

6 Discussion To evaluate the experimental results, we conducted an analysis of the surveys and compared them with our previous research outcome where a pure gesture-based interface was employed [22]. The analysis demonstrated that the UR-VC application turned out to be more convenient, generic, intuitive, user-friendly, re-usable in collaborative assembly tasks and scalable then the previous gesture-based control system. While observing the participants’ behavior, we noted a number of interesting features. Even though some participants forgot or followed some steps in the instructions incorrectly, which caused unexpected situations (i.e., a participant forgot to put the fidget spinner frame in the plastic mold for the robot) during the assembly task, the participants still coped with it with a minimum of hints due to a fact that they quickly learned the basics of controlling the cobot using the UR-VC. To evaluate the participants’ behavior in an unexpected situation, they were arbitrarily asked to suspend the robot’s operation in the course of a task. Even though a proper button for suspending was not specified to them explicitly, all participants made a correct choice within the UR-VC interface. No difference was observed in evaluations provided by participants with and without experience in robotic manipulators. Emphasized by the participants deficiencies mainly concerned the system response time and ergonomic requirements for the operator’s workplace, e.g., suggesting different screen or emergency stop button locations with regard to the operator’s position. These comments highlighted additional ergonomic requirements, which should be considered in HRC interfaces’ development. Overall, both the qualitative and quantitative assessments from users of the URVC application were significantly higher than the assessments obtained by the pure gesture-based interface [22].

7 Conclusion In this study, we presented the results of experimental validation of a virtual interface for contactless control of the UR cobot within a joint assembly task. The results of experiments with 24 participants with and without robotic manipulation background indicated that people can effectively and rapidly learn a novel type of interaction with a robot in collaborative assembly tasks.

34

M. Mustafin et al.

The new method of non-contact control of the UR robot turned out to be significantly more convenient for the users than a pure gesture-based interface [22]. The experiments emphasized the requirements for fine-tuning of ergonomics related parameters of virtual control applications for cobots, which is a part of our further research. Acknowledgment. The reported study was funded by the Russian Science Foundation (RSF) and the Cabinet of Ministers of the Republic of Tatarstan according to the research project No. 22-21-20033.

References 1. Chuengwa, T.Y., Swanepoel, J.A., Kurien, A.M., Kanakana-Katumba, M.G., Djouani, K.: Research perspectives in collaborative assembly: a review. Robotics 12(2), 37 (2023) 2. Bisen, A., Payal, H.: Collaborative robots for industrial tasks: a review. Mater. Today: Proc. 52(Part 3), 500–504 (2022) 3. Rodríguez-Guerra, D., Sorrosal, G., Cabanes, I., Calleja, C.: Human-robot interaction review: challenges and solutions for modern industrial environments. IEEE Access 9, 108557–108578 (2021) 4. Villani, V., Pini, F., Leali, F., Secchi, C.: Survey on human–robot collaboration in industrial settings: safety, intuitive interfaces, and applications. Mechatronics 55, 248–266 (2018) 5. Galin, R., Meshcheryakov, R.: Review on human–robot interaction during collaboration in a shared workspace. In: Ronzhin, A., Rigoll, G., Meshcheryakov, R. (eds.) ICR 2019. LNCS, vol. 11659, pp. 63–74. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-26118-4_7 6. Galin, R.R., Shiroky, A.A., Magid, E.A., Mescheriakov, R.V., Mamchenko, M.V.: Effective functioning of a mixed heterogeneous team in a collaborative robotic system. Inform. Autom. 20(6), 1224–1253 (2021) 7. Mavridis, N.: A review of verbal and non-verbal human–robot interactive communication. Robot. Auton. Syst. 63(1), 22–35 (2015) 8. Gustavsson, P., Holm, M., Syberfeldt, A., Wang, L.: Human-robot collaboration – towards new metrics for selection of communication technologies. Proc. CIRP 72, 123–128 (2018) 9. Sultanov, R., Sulaiman, S., Li, H., Meshcheryakov, R., Magid, E.: A review on collaborative robots in industrial and service sectors. In: International Siberian Conference on Control and Communications (SIBCON), Tomsk, Russian Federation, pp. 1–7 (2022) 10. Dobrokvashina, A., Sulaiman, S., Zagirov, A., Chebotareva, E., Kuo-Hsien, H., Magid, E.: Human robot interaction in collaborative manufacturing scenarios: prospective cases. In: International Siberian Conference on Control and Communications (SIBCON), Tomsk, Russian Federation, pp. 1–6 (2022) 11. Cherubini, A., Navarro-Alarcon, D.: Sensor-based control for collaborative robots: fundamentals, challenges, and opportunities. Front. Neurorobot. 14, 113 (2021) 12. Hameed, A., Ordys, A., Mo˙zaryn, J., Sibilska-Mroziewicz, A.: Control system design and methods for collaborative robots: review. Appl. Sci. 13(1), 675 (2023) 13. Costa, G.M., Petry, M.R., Moreira, A.P.: Augmented reality for human–robot collaboration and cooperation in industrial applications: a systematic literature review. Sensors 22(7), 2725 (2022) 14. Lindblom, J., Alenljung, B., Billing, E.: Evaluating the user experience of human–robot interaction. In: Jost, C., et al. (eds.) Human-Robot Interaction. SSBN, vol. 12, pp. 231–256. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-42307-0_9

Experimental Validation of an Interface for a Human-Robot Interaction

35

15. De Simone, V., Di Pasquale, V., Giubileo, V., Miranda, S.: Human-robot collaboration: an analysis of worker’s performance. Proc. Comput. Sci. 200, 1540–1549 (2022) 16. Gualtieri, L., Rauch, E., Vidoni, R., Matt, D.T.: Safety, ergonomics and efficiency in humanrobot collaborative assembly: design guidelines and requirements. Proc. CIRP 91, 367–372 (2020) 17. Shukla, D., Erkent, Ö., Piater, J.: Learning semantics of gestural instructions for human-robot collaboration. Front. Neurorobot. 12, 7 (2018) 18. Quek, F.K.H.: Eyes in the interface. Image Vis. Comput. 13, 511–525 (1995) 19. Simão, M., Gibaru, O., Neto, P.: Online recognition of incomplete gesture data to interface collaborative robots. IEEE Trans. Ind. Electron. 66, 9372–9382 (2019) 20. Vysocky, A., Poštulka, T., Chlebek, J., Kot, T., Maslowski, J., Grushko, S.: Hand gesture interface for robot path definition in collaborative applications: implementation and comparative study. Sensors 23, 4219 (2023) 21. Hand Gesture Recognition Using Mediapipe. https://github.com/Kazuhito00/hand-gesturerecognition-using-mediapipe. Accessed 01 Apr 2023 22. Mustafin, M., Chebotareva, E., Li, H., Martínez-García, E.A., Magid, E.: Features of interaction between a human and a gestures-controlled collaborative robot in an assembly task: pilot experiments. In: Proceedings of International Conference on Artificial Life and Robotics, pp. 158–162 (2023)

Fail It till You Make It: Error Expectation in Complex-Plan Execution for Service Robots Luis Contreras1 , Jesus Savage2(B) , Stephany Ortuno-Chanelo2 , Marco Negrete2 , Arata Sakamaki1 , and Hiroyuki Okada1 1 Tamagawa University, Machida, Tokyo 194-8610, Japan 2 National Autonomous University of Mexico, Coyoacan, 04510 Mexico City, Mexico

[email protected]

Abstract. Domestic service robots (DSR) are devices aimed to carry out daily household chores. Recently, the range and difficulty of the activities they can perform are reaching amazing performances, frequently resulting in complex plans with several steps requiring many skills where errors are more likely to occur. In this paper, we introduce the concept of error expectation in complex plans for domestic service robots and propose a classification of DSR’s tasks based on the abilities required for their execution. We also propose a recovery system where a type of feedback is chosen based on error expectation and DSR task type. We test our proposal in the context of an object manipulation task and discuss how error expectation contributes to a good feedback choice. An action involving several robot skills, namely navigation, human and object recognition, natural language processing, etc., was illustrated by the “take” action. The video showing the complete execution of a complex command by a robot using error recovery is presented. Keywords: Service Robots · Action Planning · Error Expectation

1 Introduction Autonomous service robots need several skills – such as motion planning in dynamic environments, object recognition and manipulation, human-robot interaction, and realtime awareness – that result from the combination of a number of actions, sub-actions, and behaviours. These skills have been continuously evaluated in domestic environments and put into practice in competitions such as Robocup@Home [1], RoCKIn@Home [2], and World Robot Summit [3], where the difficulty of the tasks increases each year to challenge the state-of-the-art technique. Nowadays, the tasks a service robot has to perform involve the integration of many skills, as individual systems can’t provide all the information required to solve a given problem.

Supplementary Information The online version contains supplementary material available at https://doi.org/10.1007/978-3-031-43111-1_4. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 36–46, 2023. https://doi.org/10.1007/978-3-031-43111-1_4

Fail It till You Make It: Error Expectation in Complex-Plan Execution

37

In [4] it was shown that from any given spoken command, through a semantic reasoning module, it is possible to generate a series of actions that allow a service robot to perform single tasks such as “Put the crackers on the kitchen table”, “Tell me the name of the woman in the kitchen”, and “Pour some cereals in the bowl”. Then, a rule-based system is used for high-level reasoning in more complex tasks by combining a series of low-level behaviour methods, seen as reactive algorithms that solve local problems using a number of finite steps. One important outcome of this evaluation is that we observe a decrease in performance while the number of skills necessary to perform a specific task increases due to the conditional probabilities of success given by each sub-task. While it has been shown how a supervisor system improves the performance of the robot by making new plans in unforeseen situations by calling the action again, active perception in low-level behaviours helps to solve the task without changing the plan by actively updating the state of the robot and the environment when executing a sub-task and or reacting when an unexpected situation arises. Considering a full planning system, from a given command we obtain a robot plan – consisting of several steps, where each step is a combination of several skills and behaviours –, in this paper, we introduce error expectation during plan execution for robot recovery while performing such plan; we present our results in the object manipulation problem where error detection and fast error recovery, as well as continuous environment updates, result in a fundamental feature to complete a task with the best performance.

2 Related Work Recently, RLBench, a test platform for several manipulation applications, has been presented in [5] where the authors use a simulator with a fixed robot arm on a surface and an upper RGBD camera and an eye-in-hand monocular camera, and they feature 100 different tasks providing proprioceptive and visual observations. Similarly, in [6] the authors present Ravens, a simulated benchmark with ten manipulation tasks, mainly focusing on the “transportation” skills, i.e., tasks that executes a sequence of spatial displacements. Towards task generalization, CLIPort [7] propose an end-to-end framework that is capable of solving a variety of language-specified tabletop tasks from packing unseen objects to folding cloths, all without any explicit representations of object poses, instance segmentation, memory, symbolic states, or syntactic structures; this work combines the manipulation skills of Transporter [6] with the image representations and semantic understanding given by CLIP [8]. However, based on the works of [9] and [10] on scaling law trends, in [8] the authors raise several concerns, not only considering the size of the dataset and the amount of compute used for training, but also in the bias that deep learning models might introduce through generalization in large datasets; regarding domestic robots, this bias might result in a single way to perform a task without considering users’ will – task performance generalization in personal robots might blur individual intentions and expectations and cause not only annoyance and frustration but also accidents and serious injuries. More general works towards service robots are Habitat 2.0 [11], sDSPL [12], and BEHAVIOR [13]. In Habitat 2.0, the authors present a simulated home environment

38

L. Contreras et al.

where the robot can perform several tasks; however, no specific rules and regulations are provided. In this sense, in sDSPL, the authors propose the simulation Domestic Standard Platform League (sDSPL) as a benchmark to evaluate the performance of service robots while executing a task; they focused on standardization and evaluation of service robots while performing a general purpose task to close the gap between robot competitions and research. Finally, in BEHAVIOR the authors present their “Benchmark for Everyday Household Activities in Virtual, Interactive, and Ecological Environments”; they propose 100 indoor chores in realistic environments such as “Assembling gift baskets”, “Cleaning the bedroom”, “Installing a modem”, or “Sorting groceries”. They provide 500 human demonstrations in virtual reality and propose a number of a set of metrics to measure task progress and efficiency, absolute and relative to human demonstrators. To address this problem, we are working on complex action planning by combining general-purpose service robot models based on expert systems (as in [4]) with standard and data-driven skill models (e.g., CLIPort [7]) and then, evaluate them in real scenarios explicitly designed to test such systems, like robot competitions. Furthermore, although deep learning systems generalize well, they should serve to create local plans and their execution should still be performed considering human intentions and using active reasoning as in [14–16] to encompass the specificities of the current scenario and to handle unexpected behaviours.

3 Error Expectation in General Purpose Service Robot (GPSR) The presence of service robots in domestic environments is increasing recently and, in consequence, the familiarity of users and the difficulty of the tasks that might request. From developing individual skills to performing complex tasks, several platforms have been proposed recently aimed at solving general-purpose domestic tasks. However, although many platforms and tests have been proposed, they mainly focus on single tasks at a time, like “Take out the garbage”, “Storing groceries”, “Clean up”, “Welcoming Visitors”, and so on. Furthermore, many of those tests are performed by static robots, consisting mainly of an rgb(d) camera and a robotic arm manipulating objects on a surface. However, one key challenge towards autonomous service robots has been proposed in the RoboCup at Home competition, namely, the General Purpose Service Robot (GPSR) test, where a number of general-purpose commands are randomly generated using a [EE]GPSR Command Generator and grammars publicly available at [17]. In short, in this test, there are three categories, according to their difficulty. In category I, the service robots must solve easy tasks involving basic skills such as indoor navigation, grasping known objects, answering questions (from the predefined set of questions), etc. (e.g., “Bring me the apple juice from the counter”, “Tell me how many beverages are on the shelf”, “Tell me the name of the person at the door”). In category II, robots must solve tasks with a moderate degree of difficulty. This category involves following a human, indoor navigation in crowded environments, manipulation, and recognition of alike objects, finding a calling person (waving or shouting), etc. (e.g., “Tell me how many beverages on the shelf are red”, “Count the waiving people in the living room”, “Follow Anna at the entrance”). Finally, in category III, the robots must understand challenging

Fail It till You Make It: Error Expectation in Complex-Plan Execution

39

tasks dealing with incomplete information, environmental reasoning, feature detection, natural language processing, outdoor navigation, pouring, opening doors, etc. Examples are: “Pour some cereals in the bowl”, “Go to the bathroom (the bathroom door is closed)”, “Bring me the milk from the microwave (the milk is inside the microwave)”. Moreover, there is also an Extended GPSR (the EEGPSR) test, where the robot has to perform three simple actions, given as a single sentence, and validate the individual commands through HRI. Some examples are: – Go to the kitchen counter, take the coke, and bring it to me. – Bring the chips to Mary at the sofa, tell her the time, and follow her. – Find a person in the living room, guide them to the kitchen, and follow them. In [4], we generate a base set of robot skills, namely, Speech Recognition (SR), Navigation (NV), Person Recognition (PR), Object Recognition (OR), and Object Manipulation (OM), and we classify the [EE]GPSR command into several categories that require a subset of skills to be solved: – – – – – – – – – – – –

Follow person: SR, NV, PR Guide person: SR, NV, PR Gender person: SR, NV, PR Person pose: SR, NV, PR Name Person: SR, NV, PR How many people: SR, NV, PR Person instructions: SR, NV, PR How many objects: SR, NV, OR Feature object: SR, NV, OR Bring me object: SR, NV, OR, OM Place object: SR, NV, OR, OM Handover object: SR, NV, PR, OR, OM

4 Experiments and Results In the remaining of this work, we will focus on the object manipulation problem to illustrate how to introduce error expectation to handle unexpected behaviours while solving this problem, depending on the strategy to follow. 4.1 Object Recognition We propose a series of strategies for object recognition in human-made environments and evaluate the performance in the object recognition task; the proposed object-recognition strategies are illustrated in Fig. 1 and are described as follows. Random Distance As a baseline, we propose to detect objects from a fixed distance close to the location of the object determined by the navigation strategy, i.e., a search point that allows future object manipulation without collisions. A clear advantage of this strategy is that object

40

L. Contreras et al.

Fig. 1. Robot-object observation strategies, namely a) random distance, b) close distance, and c) hand distance.

manipulation follows immediately after object recognition a disadvantage is the object’s low resolution in the images. Close Distance Here, the robot gets the closest to the location of the object, sometimes by rotating its head by 90° so the hand does not occlude its visibility. An advantage of using this strategy is that the image resolution improves; however, several disadvantages surges such as limited field of view and limited configuration space and therefore different reference systems between the recognition and manipulation poses, increasing the performance time. Hand Distance In this strategy, the robot first performs object detection and manipulation and then recognizes the object in its hand. An advantage is the constant object resolution in the recognition step but at the cost of an increased manipulation time consumption per object. We run a series of experiments in a similar setup, where objects were located inside a shelf at two different levels. For the random distance, we located the robot at 80 cm from the edge within a small odometry error; then, for the close-distance case, objects were located in four positions within the shelf, namely bottom and middle levels in the left

Fail It till You Make It: Error Expectation in Complex-Plan Execution

41

and right side and the robot gets the closest to each object so the objects are still visible for the upper camera. Finally, in the case of hand distance, the robot first detects and grasps the objects and then recognizes them. We use the same out-of-the-box CNN for object detection for all strategies to objectively assess the performance of each strategy, as we are not interested in the recognition rate but in assessing if there is a performance variation if we change the robot behaviour. Table 1 shows the recognition rate after 20 trials per strategy. Table 1. Recognition rate using different robot-object interaction strategies after 20 trials. Interaction strategy Performance

Random Distance

Close Distance

Hand Distance

0.40

0.40

0.45

From the results, it can be observed that the hand-distance strategy slightly performs better, we assume that it’s due to the more constant point of view while in the other cases highly depends on the object’s initial pose. 4.2 Object Detection To assess the importance of error expectation in robot task execution, we evaluate it in the object manipulation task, where we compare several feedback methods while cleaning a room (Fig. 2) as follows. We use Feedback I as a baseline, where the robot detects the objects using the RGBD camera, and grasps an object to place it in a free spot on a given shelf. Finally, the robot goes back close to the objects and starts the detectionand-grasping process again. There is no tactile feedback, and therefore dropping is not detected, in consequence, the robot performs the placing routine even though it has no object in its hand. Furthermore, there is no time limit in this baseline, so the robot will perform the routine until no object is detected on the floor. Feedback II includes an attempt constraint, where the robot performs object detection and grasping a limited number of times. Then, Feedback III introduces a memory feature by detecting the objects only once and saving their positions for the remaining task; similarly, it aims to grasp an object a limited number of times. Feedback IV uses the force sensor to detect an unsuccessful grasping and, in case of a drop, the robot searches for the objects once more and aims to grasp them without performing any unnecessary delivery; again, the time limit is used but not the memory feature. Finally, in Feedback V, within an attempt’s limit, the robot searches for the objects and saves their position and then, it aims to grasp them. If a successful grasping is detected, the robot delivers the object and goes back to take the next object without searching for them again. However, when a dropping is detected, the robot searches for the remaining objects again and updates its knowledge. In our experimental setup, the robot has to place five objects lying on the floor on three different shelves, and a neutral shelf in case a designated shelf is full (i.e., no free

42

L. Contreras et al.

Fig. 2. In the Tidy up task, the robot has to clean the room by placing the mislocated objects in their corresponding shelves.

space detected), as shown in Fig. 2. We use common objects found in a house, such as soft and hard toys, cards, food, and drinks. For every feedback method, we perform ten trials (50 experiments in total with five objects per trial); some setup samples are shown in Fig. 3. Regarding the attempts’ constraint, for the five objects, the robot is allowed to drop them at most seven times. Results are shown in Table 2.

Fig. 3. Different samples of the objects used in the experiments.

It can be observed that, except for the one-time memory only feedback method (Feedback III), while the success performance is not significantly different, the time

Fail It till You Make It: Error Expectation in Complex-Plan Execution

43

Table 2. Average performance in the object manipulation task where the robot has to take objects on the floor to their correct location. The performance indicates the ratio of successfully delivered to the total number of objects in ten trials per method with five objects per trial. Manipulation feedback I None

II Attempts

III Memory + Attempts

IV Force + Attempts

V Force + Memory + Attempts

Performance

0.82 ± 0.19

0.76 ± 0.18

0.54 ± 0.28

0.80 ± 0.20

0.82 ± 0.15

Time consumption [min]

8.17 ± 3.99

7.56 ± 1.47

4.95 ± 0.57

6.39 ± 1.42

5.53 ± 0.91

performance does. We explain it as follows. Regarding the baseline (Feedback method I), a perfect performance was not reached as some objects were very reflective or, after dropping, they rolled out of the field of view, so the RGBD camera was not able to detect them. Besides, some objects were very slippery (such as glass jars) and therefore they required many attempts to be grasped (expressed as a high time variance because from none to several of these objects could be randomly included in a given trial). While introducing the attempts’ constraint (Feedback II) slightly reduces the performance, the variance in time is decreased, as objects that require many grasping attempts are the most difficult and an early stop prevents them to continue the process until those objects move out of the field of view. On the other hand, although the one-time memory method (Feedback III) reduces the time consumption drastically, as the objects have to be detected only once, the performance is reduced as well because the objects near or on the dropping point in general change their pose and therefore the memory results outdated. Then, tactile Feedback method III reduces time consumption as no false delivery has to be performed. Importantly, this method represents a naive approach where we save the environment state once to perform the task at hand; however, we see that, given some small errors in the navigation system, future attempts result in unsuccessful and therefore it proves the necessity of an active feedback strategy. Finally, the multimodal Feedback method V takes all the advantages, reaching a similar performance to the baseline but in a significantly shorter time, as no false deliveries are performed, and few object detection are done (i.e., only in the presence of a dropping the memory is updated), and an early stop is used for hard-to-grasp objects. Moreover, it is easy to calculate the probability of success in the object detection and manipulation problem as a joint probability of any of the components, depending on the strategy to follow; for example, given the results in Tables 1 and 2, taking any object using force feedback would result in: p = pskill = pmanipulation = 0.80.

44

L. Contreras et al.

Similarly, taking the correct object from a random distance and using force feedback, would result in the skill joint probability: p = pskill = pdetection × pmanipulation, p = 0.40 × 0.80 = 0.32. With these results, different strategies can be considered, such as finding a minimum or maximum number of trials to deliver the right object before giving up; also, it can be highlighted the weakest skills in the whole process and focus efforts to improve their performance. For example, the expectation E of the number of trials to the first occurrence of successful performance with skill probability p in a sequence of trials is: E = 1/p. Furthermore, if we model the probability of success and failure as a binomial distribution, the probability P of x successes in n trials given a skill probability p is defined as P(x) = n Cx · px · (1 − p)n−x . In specific, while executing a robot plan, we move to the next step after the first success in the previous one and, therefore, x = 1 and n C x = n, then P(x = 1) = n · p · (1 − p)n−x . With this expression, we can get the number n of trials required to successfully perform a skill with a desired success probability P given the skill performance probability (or joint probability) p. After those trials, we can say that the robot is unable to perform the task and request human assistance or stop the task execution. To summarize our approach, in an environment as in Fig. 4, from an action planer, given the following command: “Go to the living room, find Luis, and give him some chips”, we get a robot plan consisting of several actions, namely: – – – –

go to large table in living room; take chips; go to living room center in living room; deliver the object to Luis.

Where each of the actions requires one or several robot skills, including navigation, person and object recognition, natural language processing, and so on – in this work, we illustrate these concepts for the “take” action. In [18] you can observe a full performance of a complex command using error recovery.

Fail It till You Make It: Error Expectation in Complex-Plan Execution

45

Fig. 4. Experimental setup as in [3] for the GPSR test from the RoboCup at Home competition where the robot should attend a complex command.

5 Conclusions In this work, we aimed at solving general-purpose tasks by introducing error expectation in complex planning execution. In specific, we illustrate our approach in the object manipulation task, where we work towards improving the performance when using service robots by adding active reasoning; this includes testing several points of view, by moving the robot’s base and or head and by first grasping the object and then observe it at a fixed distance in the robot’s hand. The reason for that is that lately, it seems that robots are being seen as external agents that move around an environment and not as intelligent agents that can actually interact in it and even alter it to its best convenience (sometimes in a way a human can’t do) when performing (and if allowed by) a given task. We consider a plan as a series of steps, and each step can consist of one or several skills; each skill has a probability of success, and the joint probability would result in an average performance. By determining these probabilities, a robot can recover from unexpected behaviours or give up after a specific number of attempts. Acknowledgements. This work has been supported by the project, JPNP20006, commissioned by the New Energy and Industrial Technology Development Organization (NEDO) and by PAPIITDGAPA UNAM under Grant IG-101721.

46

L. Contreras et al.

References 1. Wisspeintner, T., Van Der Zant, T., Iocchi, L., Schiffer, S.: RoboCup@Home: scientific competition and benchmarking for domestic service robots. Interact. Stud. 10(3), 392–426 (2009) 2. Amigoni, F., et al.: Competitions for benchmarking: task and functionality scoring complete performance assessment. IEEE Robot. Autom. Mag. 22(3), 53–61 (2015) 3. Contreras, L., Yamamoto, T., Matsusaka, Y., Okada, H.: Towards general purpose service robots: world robot summit partner robot challenge. Adv. Robot. 36, 812–824 (2022). https:// doi.org/10.1080/01691864.2022.2109428 4. Savage, J., et al.: Semantic reasoning in service robots using expert systems. Robot. Auton. Syst. 114, 77–92 (2019) 5. James, S., Ma, Z., Arrojo, D.R., Davison, A.J.: RLBench: the robot learning benchmark & learning environment. IEEE Robot. Autom. Lett. 5(2), 3019–3026 (2020) 6. Zeng, A., et al.: Transporter networks: rearranging the visual world for robotic manipulation. In: Conference on Robot Learning, pp. 726–747 (2021) 7. Shridhar, M., Manuelli, L., Fox, D.: CLIPort: what and where pathways for robotic manipulation. In: Conference on Robot Learning, pp. 894–906 (2022) 8. Radford, A., et al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning, pp. 8748–8763 (2021) 9. Hestness, J., et al.: Deep learning scaling is predictable, empirically. arXiv preprint arXiv: 1712.00409 (2017) 10. Kaplan, J., et al.: Scaling laws for neural language models. arXiv preprint arXiv:2001.08361 (2020) 11. Szot, A., et al.: Habitat 2.0: training home assistants to rearrange their habitat. Adv. Neural Inf. Process. Syst. 34, 251–266 (2021) 12. Contreras, L., Matsusaka, Y., Yamamoto, T., Okada, H.: sDSPL-towards a benchmark for general-purpose task evaluation in domestic service robots. In: The 39th Annual Conference of the Robotics Society of Japan (2021) 13. Srivastava, S., et al.: Behavior: benchmark for everyday household activities in virtual, interactive, and ecological environments. In: Conference on Robot Learning, pp. 477–490 (2022) 14. Vazquez, E., Contreras, L., Okada, H., Iwahashi, N.: Multimodal human intention-driven robot motion control in collaborative tasks. In: 38-th Annual Conference of the Robotics Society of Japan (2020) 15. Contreras, L., Muto, Y., Okada, H., Sakamaki, A.: Robot-object interaction strategies for object recognition. In: 37th Annual Conference of the Robotics Society of Japan (2019) 16. Contreras, L., Mayol-Cuevas, W.: Towards CNN map representation and compression for camera relocalisation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 292–299 (2018) 17. RoboCup@Home Command Generator. https://github.com/kyordhel/GPSRCmdGen. Accessed 19 May 2023 18. Fail it till you make it: error expectation in complex-plan execution for service robots. https:// youtu.be/rhJ-dcahfqc. Accessed 19 May 2023

Moving Person Detection Based on Modified YOLOv5 Xin Shen(B)

, Guo Wu , and Vadim Lukyanov

Bauman Moscow State Technical University, 5/1, 2nd Baumanskaya st., Moscow 105005, Russia [email protected]

Abstract. Visual Dynamic SLAM (Simultaneous Localization and Mapping) is a fundamental technology for intelligent mobile systems, enabling applications in robotics, augmented reality, and self-driving cars. This paper presents a novel approach to improve the performance of Visual Dynamic SLAM by integrating the YOLOv5 object detection framework with attention mechanisms (CBAM) and a BiFPN (Bidirectional Feature Pyramid Network) structure. The dynamic feature points, which are located in the bounding box of the dynamic object, are removed in the tracking thread, and only the static feature points are used to estimate the position of the camera. The primary focus is on improving the detection performance of dynamic objects, particularly persons, and addressing challenges such as occlusion and small object detection. Overall, the integration of YOLOv5 with attention mechanisms and a BiFPN structure presents a significant advancement in visual dynamic SLAM. The proposed approach enhances the detection of persons, addresses challenges related to small objects and occlusion, and improves the overall performance of the system in dynamic environments. These findings demonstrate the effectiveness of the proposed methodology and its potential for real-world applications in various domains, including robotics, augmented reality, and self-driving cars. Keywords: CBAM Attention Mechanism · Bidirectional Feature Pyramid Network · Person Detection · Occlusion · Dynamic SLAM · YOLO Detector

1 Introduction In the last few decades, visual SLAM techniques have gained significant interest from both the computer vision and robotic communities. Many variants of these techniques have started to make an impact in a wide range of applications, including robot navigation and augmented reality. However, most existing visual SLAM techniques assume a static environment, limiting their applicability in real-world scenarios. Detecting moving persons accurately and efficiently is a crucial task in computer vision, such as the dynamic SLAM problem. Traditional object detection methods face challenges in dynamic environments, including occlusion and small object detection. Currently, person detection in computer vision faces a trade-off between accuracy and real-time performance. Existing approaches typically fall into two categories: twostage detectors that offer high accuracy but lack real-time capability, which is not suitable © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 47–58, 2023. https://doi.org/10.1007/978-3-031-43111-1_5

48

X. Shen et al.

for SLAM tasks, and one-stage detectors that are fast but suffer from reduced performance, particularly in cases of occlusion and detecting small-sized people, which is dangerous with potential chance of collision. These enhancements can focus on two key aspects. Firstly, incorporating advanced attention mechanisms such as spatial and channel attention can help the detector better capture and leverage context information, allowing it to handle occlusion scenarios more effectively. Secondly, by incorporating more advanced neck structure such as BiFPN [1], the detector can effectively capture both local details and global context, leading to improved performance in detecting small and distant persons. To address these limitations, this paper proposes a novel approach for moving person detection using a modified YOLOv5 framework that incorporates the attention mechanism CBAM (Convolutional Block Attention Module) [2] and the BiFPN (Bidirectional Feature Pyramid Network) structure. The proposed approach, based on human detection and the use of the YOLOv5 detector, has significant potential for application in robot navigation. This approach allows robots to efficiently detect people in their environment and make appropriate decisions based on the information they discover. The use of modified YOLOv5 model provides high accuracy and processing speed, which are important factors for real-time robots. The main contributions of this paper are listed as follows: • We integrate CBAM into YOLOv5, which can help the network to find region of interest in images that have large region coverage. • We propose a modified YOLOv5 framework that incorporates the BiFPN structure as the neck part for moving person detection. • We present comprehensive experimental results demonstrating the superior performance of our approach compared to existing methods in terms of accuracy, robustness, and small object detection capabilities. The remainder of this paper is organized as follows. Section 2 provides an overview of related work in object detection. Section 3 describes the proposed methodology in detail, explaining the modifications made to the YOLOv5 framework and the integration of CBAM and BiFPN. Section 4 presents the experimental setup, datasets introduction, and evaluation used to assess the performance of our model. Finally, in Sect. 5 concludes the paper and highlights future research directions in the field of moving person detection based on YOLOv5 with attention mechanisms and the BiFPN structure.

2 Related Works 2.1 Dynamic SLAM Based on Deep Learning Most state-of-the-art SLAM systems are not able to handle dynamic scenarios, as they were designed with a static environment assumption [3]. The Visual SLAM systems that deal with dynamic content in the scene usually treat it as noise and filter it. With the rapid development of deep learning technology in the field of computer vision, object recognition technology is being increasingly applied in the field of robot navigation. One approach called DynaSLAM [3] employed by Berta Bescos et al. involves utilizing deep learning to detect and localize dynamic objects, such as people in indoor

Moving Person Detection Based on Modified YOLOv5

49

situations. They have used the Mask R-CNN [4] model in their work, which unfortunately incurs a significant computational cost and cannot operate in real-time scenarios. It is worth noting that Mask R-CNN is a two-stage object detector [4], known to be slower compared to one-stage detectors like the YOLO family [5–8]. In our work, our model is modified from the famous version of YOLO series, YOLOv5s, until now, it is still one of the most famous and effective SOTA in object detection [9]. 2.2 Person Detection in Indoor Environments From the perspective of components, object detectors usually consist of two parts, an CNN-based backbone, used for image feature extraction, and the other part is detection head used to predict the class and bounding box for object. In addition, the object detectors developed in recent years often insert some layers between the backbone and the head, people usually call this part the neck of the detector. Backbone. The backbone that are often used include EfficientNet [10], ResNet [11], etc., rather than networks designed by ourselves. Because these networks have proven that they have strong feature extraction capabilities on classification and other issues. Neck. The neck is designed to make better use of the features extracted by the backbone. It reprocesses and rationally uses the feature maps extracted by Backbone at different stages. Usually, a neck consists of several bottom-up paths and several topdown paths. Neck is a key link in the target detection framework. Now commonly used path-aggregation blocks in neck are: FPN [12], PANet [13]. Head. As a classification network, the backbone cannot complete the positioning task, and the head is designed to be responsible for detecting the location and category of the object by the features maps extracted from the backbone. Deep learning already became the mainstream method to realize person detection in different situations, in [14] Chloe Eunhyang Kim et al. listed the most popular embedded deep learning methods for person detection, and pointed that neither of these models nail the tradeoff between speed and accuracy. In [15] Linxiang Zhao et al. proposed a new model based on YOLOv3 to enhance its performance on person detection, but the test dataset they have used are all in good visibility conditions, without significant occlusions, as expected in dynamic SLAM problems. In other work such as [16–18], they also began to try use some good techniques to enhance the YOLO neural network such as new loss function to guide the optimization of the class to which the object belongs, or maybe exchange some small module structures such as SPP to make it could obtain more feature-rich image information. But the occlusion problem or small object detection task still could not be handled so well. Some research [17] focus on other object detection showed us that to make a good and specific dataset for our own situation is quite beneficial for us. In [18] Jiahui Sun et al. have begun to use attention mechanism into YOLO model, but how to choose a best or most suitable attention mechanism for specific scenario still need to be confirmed.

50

X. Shen et al.

3 New Model Architecture and Methodology 3.1 Overview of YOLOv5 Network The YOLO (You Only Look Once) network model [5–8] is an algorithm that can be used for target detection. Compared to two-stage detectors, one-stage YOLO series algorithms have much greater advantages in running speed, faster and more efficient. YOLO provides the classes of the detected objects, 2D bounding boxes with their corresponding positions and a confidence number for each box. At present, the YOLOv5 is mainly divided into four network models, which are YOLOv5s, YOLOv5m, YOLOv5l, and YOLOv5x. Among them, the YOLOv5s network model is the base model with shallow depth, and the narrowest width of the feature map compared to the other three models. Among these five models, we selected the smallest and fastest pre-trained model YOLOv5s to meet our requirements for speed. 3.2 Convolutional Block Attention Module (CBAM) CBAM [2] is a simple but effective attention module. It is a lightweight module that can be integrated into most CNN architectures, and it can be trained in an end-to-end manner. The attention mechanism CBAM enhances the model’s ability to selectively focus on relevant information, which is beneficial for occlusion problem. Given a feature map, CBAM sequentially infers the attention map along two separate dimensions of channel and spatial, and then multiplies the attention map with the input feature map to perform adaptive feature refinement. The structure of the CBAM module is shown in the Figs. 1, 2 and 3.

Fig. 1. The structure of CBAM.

The CAM (Channel Attention Module) exploits the interdependencies between channels within a feature map, enabling the network to learn channel-wise importance weights. It achieves this by utilizing a global pooling operation to capture channel-wise statistics, followed by a series of fully connected layers that generate channel attention weights. Mc (F) = σ (MLP(AvgPool(F)) + MLP(MaxPool(F)))       c  c = σ W1 W0 Favg + W1 W0 Fmax ,

(1)

where σ denotes the sigmoid function, W0 ∈ RC/r×C , and W1 ∈ RC×C/r . Note that the MLP weight, W0 and W1 , are shared for both inputs and the ReLU activation function is followed by W0 .

Moving Person Detection Based on Modified YOLOv5

51

Fig. 2. The structure of CAM.

Fig. 3. The structure of SAM.

On the other hand, the SAM (Spatial Attention Module) captures spatial dependencies by examining the relationships between spatial locations within a feature map.   Ms (F) = σ f 7×7 ([AvgPool(F); MaxPool(F)])    s s , (2) = σ f 7×7 Favg ; Fmax where σ denotes the sigmoid function and f 7×7 represents a convolution operation with the filter size of 7 × 7. 3.3 Bidirectional Feature Pyramid Network (BiFPN) Furthermore, we introduce the BiFPN [1] structure into the modified YOLOv5 framework. The BiFPN enhances multi-scale feature fusion, which is essential for handling small object detection. By incorporating the BiFPN, the model gains the ability to capture context and spatial information from different scales, resulting in improved localization and detection performance. The proposal FPN [12] solves the problem of multi-scale in target detection and greatly improves the detection performance of small targets. Compared with the traditional FPN network, BiFPN adds skip connections between the input and output feature in the same layer [6]. Because of the same scales, adding skip connections can better extract and transfer feature information. The structure of BiFPN is shown in Fig. 4.    w1 · P6in + w2 · Resize P7in td (3) P6 = Conv w1 + w2 +     w1 · P6in + w2 · P6td + w3 · Resize P5out out P6 = Conv (4) w1 + w2 + w3 + 

52

X. Shen et al.

The parameter w is a learned parameter that distinguishes the importance of different features during the feature fusion process, somewhat similar to an attention mechanism. In summary, BiFPN can be described as an enhanced version of PANet, which incorporates repeated bidirectional cross-scale connections and a weighted feature fusion mechanism.

Fig. 4. Basic structure of BiFPN.

3.4 New Model Architecture As shown in Fig. 5, our model consists of three main parts: a backbone with incorporation with CBAM, a detection head, and the neck part which utilize the BiFPN structure. Backbone. An end-to-end network, capable of extracting semantic information directly from the input image, has proven to be highly effective in the field of object detection, as demonstrated by YOLOv5 [9]. Neck. Transitioning from PANet to BiFPN. This design aims to enhance information flow through bottom-up path augmentation.

4 Experimental Results 4.1 Datasets Preparation and Training Results In this section, the dataset preparation process for the experimental results in the paper is outlined. The COCO dataset [19] was utilized, although not in its entirety due to its extensive size. For the specific focus of indoor dynamic SLAM, only images with the class label “person” were considered as they represented our target class in this situation. To extract the relevant images from the COCO dataset, a Python script file was employed. This script was designed to identify and retrieve all images associated with

Moving Person Detection Based on Modified YOLOv5

53

Fig. 5. Structure of our proposed model.

the “person” class label. Then a subset of images that matched the research requirements was obtained. Then we randomly chose around 20,000 images from them to train our proposed YOLOv5 model. We have chosen pre-trained weights yolov5s.pt to help our training process. Before training, we specified the model hyperparameters: batch size 32, image size 640, initial learning rate 0.01 using SGD with 0.001 weight decay and a momentum of 0.9, and number of epochs is 300, for the rest we just used the default values, such as IOU threshold 0.45 and confidence threshold 0.25. Our training process was performed on an Ubuntu 20.04 with NVIDIA GeForce RTX 3060Ti GPU with 8 GB memory. Table 1 shows the standard indicators: precision, recall, [email protected] (average AP accuracy at IoU of at least 0.5) and [email protected]:.95 (average AP accuracy at threshold values IoU in the range from 0.5 to 0.95 in steps of 0.5). The number of frames that the model can process in one second (FPS) is also calculated separately. Table 1. Quantitative comparisons of training results from two models. Model

Precision

Recall

[email protected]

[email protected]:.95

FPS

YOLOv5s

0.755

0.695

0.763

0.498

303

YOLOv5s_modified

0.812

0.696

0.767

0.503

333

54

X. Shen et al.

4.2 Performance Evaluation on TUM RGB-D Dataset The TUM RGB-D dataset, initially introduced by the TUM Computer Vision Group in 2012, has become a widely utilized resource within the Simultaneous Localization and Mapping (SLAM) field [20]. This dataset was gathered using a Kinect camera, capturing various essential components such as depth images, RGB images, and ground truth data. We have tested our modified model on some typical images with dynamic situation from this dataset. The results show that our model have more robust ability in the occlusion situation. In Fig. 6, we can observe qualitative differences, where the person behind the table, who is relatively small, is heavily occluded. Only a small portion is visible, and the original YOLOv5 fails to detect it, yielding no results. However, our improved algorithm can detect it with confidence scores of 0.60 and 0.31. Clearly, our enhancements have been highly effective.

Fig. 6. Qualitative comparison of test set.

4.3 Performance Evaluation on Real-Time Situation In this section, we describe the experimental setup and methodology used to evaluate the real-time detection performance of our modified YOLOv5 model with CBAM and BiFPN. The objective of our experiment was to assess the model’s effectiveness in detecting the “person” object class using the Realsense D455 camera, which was showed in the Fig. 7.

Moving Person Detection Based on Modified YOLOv5

55

Fig. 7. Realsense D455 RGB camera.

Experimental Setup. We conducted our experiments in indoor environments, simulating dynamic SLAM scenarios. The Realsense D455 camera was employed to capture real-time video streams of the indoor scenes. During the experiments, the modified YOLOv5 model with CBAM and BiFPN was utilized to perform real-time object detection. The provided Fig. 8 in a table presents a “Qualitative comparison of test results” for seven different indoor scenes regarding person detection. The first two scenes depict a laboratory environment, with one instance featuring a small-sized target and the other displaying an occluded target. The results clearly demonstrate that our proposed model outperforms the initial YOLOv5s model. The confidence scores for the detections in the first scene improved from 0.56 and 0.36 to 0.72 and 0.81, respectively, while in the second scene, they increased from 0.66 and 0.31 to 0.80 and 0.87. The third comparison focuses on small targets, where the detection performance for distant individuals increased from 0.74 to 0.80. In the fourth comparison, our model exhibited superior detection capabilities in identifying a heavily occluded person in a squatting position. However, both methods produced one “false positive” result due to the presence of a clothes rack with a jacket on it, which lacks specific objects in our prepared training datasets (also the COCO dataset). Thus, the models mistakenly classified the rack as a person, an error that can also occur even to human observers. The fifth comparison further highlights our model’s superior person detection ability in such scenarios, as it achieved a high confidence score of 0.66 despite the failure of YOLOv5s. The sixth comparison addresses a specific occlusion scenario involving a glass door. While this may not pose a challenge for human observers, YOLOv5s failed to detect the person due to the combined factors of distance and small object size. In contrast, our optimized model achieved a score of 0.56. The seventh and final comparison also involves a dynamic target partially occluded by glass. Our model improved the confidence score from 0.61, as initially obtained by YOLOv5s, to 0.74. To summarize, our modified model consistently outperforms YOLOv5s in various occlusion scenarios and situations involving small targets. These results strongly indicate the success and effectiveness of our modifications. In conclusion, our experiments validate the suitability and effectiveness of our modified YOLOv5 model for real-time “person” object detection in indoor dynamic SLAM scenarios. The model’s accurate detection performance and efficient real-time inference make it a valuable tool for applications that require real-time mapping and localization of people in indoor environments.

56

X. Shen et al.

1

2

3

4

5

6

7

Fig. 8. Qualitative comparison of real world scenario.

Moving Person Detection Based on Modified YOLOv5

57

5 Conclusion and Future Work In this paper we introduced a modified YOLOv5 model by incorporating CBAM and BiFPN techniques, with the goal of improving person detection in indoor environments for the dynamic SLAM tasks. By incorporating new techniques into YOLOv5s model, which is one of the SOTA, we can overcome the limitations and achieve a better balance between accuracy and real-time performance. Our experimental results demonstrated that the modified YOLOv5 algorithm outperformed the original YOLOv5s model in our target scenario. One notable advantage of our modified model was its improved performance in handling occlusion problems and small object detection. The original YOLOv5s model often failed to detect people in cases where strict occlusion occurred. Further research will be aimed at extending the application of our model to other domains beyond indoor environments. And our future investigation will focus on applying the results of the improved model, integrating them into a SLAM project, as a front-end image-processing tool in the SLAM system.

References 1. Wang, Z., Tang, J., Wu, B., Hu, H.: BiFPN: efficient multi-scale backbone for object detection. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 13029–13038 (2020) 2. Woo, S., Park, J., Lee, J.Y., Kweon, I.S.: CBAM: convolutional block attention module. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11211, pp. 3–19. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01234-2_1 3. Bescos, B., Fácil, J.M., Civera, J., Neira, J.: DynaSLAM: tracking, mapping, and inpainting in dynamic scenes. IEEE Robot. Autom. Lett. 3(4), 4076–4083 (2018) 4. He, K., Gkioxari, G., Dollár, P., Girshick, R.: Mask R-CNN. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2961–2969 (2017) 5. Redmon, J., Farhadi, A.: YOLO9000: better, faster, stronger. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6517–6525 (2017) 6. Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 779–788 (2016) 7. Bochkovskiy, A., Wang, C., Liao, H.Y.M.: YOLOv4: optimal speed and accuracy of object detection. arXiv preprint arXiv:2004.10934 (2020) 8. Redmon, J., Farhadi, A.: YOLOv3: an incremental improvement. arXiv preprint arXiv:1804. 02767 (2018) 9. Thuan, D.: Evolution of Yolo algorithm and Yolov5: the state-of-the-art object detention algorithm (2021) 10. Tan, M., Pang, R., Le, Q.V.: EfficientDet: scalable and efficient object detection. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 10781–10790 (2020) 11. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778 (2016) 12. Lin, T.Y., Dollár, P., Girshick, R., He, K., Hariharan, B., Belongie, S.: Feature pyramid networks for object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2117–2125 (2017)

58

X. Shen et al.

13. Liu, S., Qi, L., Qin, H., Shi, J., Jia, J.: Path aggregation network for instance segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8759– 8768 (2018) 14. Kim, C.E., Oghaz, M.M.D., Fajtl, J., Argyriou, V., Remagnino, P.: A comparison of embedded deep learning methods for person detection. arXiv preprint arXiv:1812.03451 (2018) 15. Zhao, L., Yi, W.: A new deep learning architecture for person detection. In: 2019 IEEE 5th International Conference on Computer and Communications (ICCC), pp. 2118–2122. IEEE (2019) 16. Ahmad, T., Ma, Y., Yahya, M., Ahmad, B., Nazir, S., Haq, A.U.: Object detection through modified YOLO neural network. Sci. Program. 2020, 1–10 (2020) 17. Jia, W., et al.: Proceedings of the real-time automatic helmet detection of motorcyclists in urban traffic using improved YOLOv5 detector. IET Image Process. 15(14), 3623–3637 (2021) 18. Sun, J., Ge, H., Zhang, Z.: AS-YOLO: an improved YOLOv4 based on attention mechanism and SqueezeNet for person detection. In: 2021 IEEE 5th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), Chongqing, China, vol. 5, pp. 1451–1456 (2021) 19. Lin, T.Y., Maire, M., Belongie, S., Hays, J., Perona P.: Microsoft COCO: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8693, pp. 740–755. Springer, Cham (2014). https://doi.org/10.1007/978-3-319-10602-1_48 20. Sturm, J., Engelhard, N., Endres, F., Burgard, W., Cremers, D.: A benchmark for the evaluation of RGB-D SLAM systems. In: 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 573–580. IEEE (2012)

Autonomous Robot Navigation System as Part of a Human-Machine Team Based on Self-organization of Distributed Neurocognitive Architectures Inna Pshenokova(B)

, Kantemir Bzhikhatlov , Olga Nagoeva , Idar Mambetov , and Alim Unagasov

The Federal State Institution of Science Federal Scientific Center Kabardino-Balkarian Scientific Center of Russian Academy of Sciences, 37-a, I. Armand Street, Nalchik 360000, Russia [email protected]

Abstract. The relevance of this study is due to the solution of the problem of developing the basic principles and algorithms for providing adaptive settings for autonomous robots intelligent control systems as part of a human-machine team based on the general method of machine learning. To do this, the paper proposes to use a formalism based on multi-agent neurocognitive architectures. Implementation of the possibility of adaptation is considered on the example of performing the task of orientation and navigation of an autonomous robot in an unfamiliar environment. An autonomous robot navigation system based on self-organization of distributed neurocognitive architectures has been developed. A multi-agent neurocognitive architecture is presented, which forms an active map containing all the locative information necessary to ensure the orientation and navigation of an autonomous robot between loci. The use of a multi-agent architecture to provide the representation of locative information in the task of implementing an interface in human-machine team will make it possible to build an ontology responsible for representing the location of objects in the external environment, as well as provide interaction with the user in natural language, taking into account its semantics. Keywords: Artificial Intelligence · Intelligent Agent · Multi-Agent Neurocognitive Architectures · Autonomous Robot · Human-Machine Team

1 Introduction The pace of development of artificial intelligence systems, robotics, and virtual reality lead to the spread of the introduction of collaborative robotics. The use of such robotic systems capable of functioning in a group with people and autonomously interacting with the environment will significantly improve efficiency and safety in some areas of activity [1, 2]. The basis of modern approaches to the development of human-machine teams © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 59–69, 2023. https://doi.org/10.1007/978-3-031-43111-1_6

60

I. Pshenokova et al.

(HMT) is the idea of creating common mental models and a communicative environment for the members of this team. At the same time, the applicability of HMT depends on the effectiveness of complex interactions between a person and a robot, as well as between them and the environment [3]. To support these interactions, an autonomous robot must have knowledge, skills, and abilities commensurate with human capabilities. That is, by itself, the introduction of HMT does not always lead to increased productivity at the individual, team, or organizational level. The works [4–7] define the competencies that are important for the successful association of HMT participants. That is, the successful application of autonomous robots within the HMT requires more complex decisionmaking systems (compared to modern solutions such as chatbots, social robots or digital assistants). The interaction of a human and a robot or an intelligent agent consists in the coordination of complex actions, such as communication, joint decision-making, and control [8, 9], in order to successfully complete tasks with potentially changing goals in uncertain environmental conditions. In [4], three main skills are identified that are necessary for effective collective interaction: communication, coordination, and adaptability, and in [7, 10] the importance of these skills for HMT is emphasized. Therefore, to improve performance in HMT, the implementation of these skills in robots and intelligent agents is very important. The development of dialogue systems is very important for high-quality communication between a robot and a person. The main task is to develop algorithms that implement the consistency of actions and statements. Existing dialogue systems solve the problems of text and speech recognition well, but in HMT it is necessary to ensure the consistency of statements with perceived actions. It is very important to develop an interface that allows the operator to clearly understand the intentions, plans, and reasoning process of an intelligent agent. Interface implementations with such functionality are still not available [11], however, machines have gained the ability to exchange information with people in natural language and more effectively coordinate joint tasks. Such an interface includes the use of speech ordering [12], which is an integral part of human-machine communication. Features such as interleaving and the ability to recognize human language [13] can improve the efficiency of bidirectional communication between humans and machines, making HMT more productive [14]. Despite the development of the communication abilities of intelligent robots and agents, there are limitations in the applicability of such interfaces in real conditions. For example, machines using neural networks provide high computational performance, but, as a rule, have problems with the interpretability of solutions due to the use of distributed statistical representations [11]. When implementing interaction in HMT, it is necessary to ensure the possibility of effective information exchange between robots, and in such a way that people can understand it. That is, machines must be able to accurately model human understanding of information, which is closely related to the second transferable teamwork competency, coordination. While communication refers to the process of exchanging information, coordination in human teams refers to the organization of knowledge, skills, and behavior of team members to achieve a specific goal [4]. In [15], coordination is defined as the process

Autonomous Robot Navigation System

61

by which humans and machines manage “dependencies between activities”. If group coordination is effective, then information that is relevant to the task is communicated in a timely manner, while avoiding excessive communication. Thus, effective communication processes be necessary but not enough for effective HMT coordination. Reference [16] defines the basic requirements for effective coordination. A machine is considered an effective coordinator if it is reliable, directive, able to communicate its own and recognize the intentions of other team members. The main problem of HMT development is the ability of robots to participate in implicit coordination. Implicit coordination refers to the process of synchronizing the activities of team members based on assumptions about what each teammate is likely to do. While machines can detect certain implicit cues through facial expression recognition, they are limited in their ability to detect contextual cues. Therefore, it is very important to develop a system that can, by observing teammates, establish their goals, preferences, and capabilities, as well as model behavior in various situations. These kinds of capabilities would support implicit coordination, allowing the machine to anticipate human behavior and expectations, and then adapt its own behavior to match. To support contextual understanding and adaptation (“third wave” of artificial intelligence), approaches are needed that can integrate both “first wave” knowledge-based methods of artificial intelligence and “second wave” methods of statistical deep learning. Most machine learning systems work by identifying correlations between variables. In contrast, at the heart of the “third wave” causal and counterfactual models aim to understand causal relationships between variables. By modeling causality, machines can better support counterfactual inference; those. They can generalize from observed operating conditions to unobserved ones. Thus, to implement adaptation in robotic systems, it is necessary that machines can not only recognize the knowledge and behavior of their teammates, but also anticipate and respond to new knowledge and behavior when necessary. Therefore, the relevance of this study is due to the solution of this problem, which consists in developing the basic principles and algorithms for providing adaptive settings for intelligent control systems for autonomous robots as part of a human-machine team based on the general method of machine learning. To do this, we propose to use a formalism based on multi-agent neurocognitive architectures [17] and having a functional similarity to the neuromorphological structure of the brain, which makes it possible to realize the unique human cognitive abilities necessary for effective interaction. Implementation of the possibility of adaptation is considered on the example of performing the task of orientation and navigation of an autonomous robot in an unfamiliar environment. The aim of the study is to develop a navigation system for an autonomous robot based on the self-organization of distributed neurocognitive architectures.

2 Autonomous Robot Navigation System Based on Multi-agent Neurocognitive Architecture In HMT, human activity consists in monitoring the functioning of an autonomous robot or robotic system and setting current tasks in a natural language dialogue mode. At the same time, it is important that the operator’s interface provides adequate human perception of

62

I. Pshenokova et al.

the current scene, and his commands are correctly interpreted by the assistant robot. The navigation system plays a significant role in this, since the robot must independently assess the environment and plan its path, including in the presence of other moving objects in the working area (for example, people, vehicles, and robots). Consider a navigation system for an autonomous robot based on the self-organization of multi-agent neurocognitive architectures. The navigation model of an autonomous robot is understood as a system of internal representations of certain specific places of the robot’s functioning (locations), which allows you to independently form a route between any significant geographical locations (locus) in this location [18]. It is assumed that the navigation model will be formed dynamically as the robot becomes familiar with the external environment. The model (map) should include only significant loci, which include the locations of objects (objects) in the location, or places where any significant events occurred with the robot (or with its participation) [19–21]. In [17], the main algorithms and methods of the multi-agent neurocognitive approach to the creation of intelligent systems are described. In [22], the concept of an intelligent agent is introduced, which is a system based on a multi-agent neurocognitive architecture and consists of a set of software agents-neurons (agneurons) interacting with each other by exchanging messages and resources. Agents operate according to the objective function aimed at increasing energy, considered as a certain scalar value characterizing the ability of the software agent to survive in the environment of its operation. Each agent-neuron in the composition of an intelligent agent represents a certain concept or its value, therefore, the activity of this concept or value in the process of intelligent decision-making by an agent or robot means its “life”. If agneuron does not participate in the intellectual activity of the agent, its active energy decreases, which leads to its “death”. The implementation of the objective function is possible since agents use protocols for interaction, called “multi-agent contracts”. A multi-agent contract is a set of algorithms according to which agents transfer some of their available energy to each other in exchange for the information they have. The ability of an agent to enter contractual relations with agents of a certain type at the structural and functional levels is called valency [23–25]. The agent receives information from the external environment through the sensory system. Different types of neuron agents that are part of an intelligent agent have a knowledge base according to which they function. In this model, knowledge is a causal relationship in which the starting situation is associated with the final (desired) situation and the action that needs to be performed to move from the initial situation to the desired one is indicated. Each such dependency represents one rule in the agents’ knowledge base. The agent’s behavior, which is determined by the internal objective function, can be controlled by editing the rules in the knowledge base. Knowledge generated by various intelligent agents built based on various neurocognitive architectures can be combined as part of an intelligent agent, since the system allows recursion (nesting) of cognitive architectures and agents into each other. The possibility of representing locative information using a multi-agent recursive cognitive architecture is based on the functional specialization of agents-neurons. Based on the hypothesis about the organization of the invariant of intelligent decision-making

Autonomous Robot Navigation System

63

[17], the following types of software agents-neurons (agneurons) are distinguished in a multi-agent architecture. To obtain data from the onboard sensors of the robot, agneurons-sensors are introduced, Lidar, Ultrasonic and GPS agents (Fig. 1). Their main function is to receive incoming information about the environment and pass it on to the appropriate agneurons for further processing. It should be noted that the navigation sensor subsystem is not limited to these sensors and includes motor encoders and inertial sensors.

Fig. 1. Agneurons-sensors.

So, for example, data from the agneuron-sensor GPS is transmitted for processing to locative agneurons (Fig. 2), the valences of which determine contractual relations with other agents, based on the need to provide comprehensive locative information (representations of a place in space, objects located in it, and events that have occurred).

Fig. 2. Agneurons of the locative type.

The semantics of relationships between different objects is conceptualized by action agneurons (Fig. 3).

Fig. 3. Action type agneurons.

64

I. Pshenokova et al.

In the case of the formation of a navigation system for an autonomous robot, an agneuron of the “locate” action type conceptualizes the relationship between the locus and the model of the subject of communication (robot) (Fig. 4) in a multi-agent architecture.

Fig. 4. Agneuron subject.

Time in the multi-agent neurocognitive architecture is represented by time-type agneurons in the Unix format (Fig. 5).

Fig. 5. Timing type agneurons.

Relationships between loci, time, action, and subject are formed by event locative agents (Fig. 6). These agents conceptualize predicative constructs that describe facts about locative information in the form “I am located at *locus at *time”.

Fig. 6. Event locative agneurons.

The formation of an active locative connection between agneurons is carried out based on the interaction of agneurons of various types with each other in accordance with contracts concluded based on the valences of agents-neurons. Let us now consider how locus are formed in a multi-agent neurocognitive architecture. As noted above, locus are the locations of objects in the location, or places in

Autonomous Robot Navigation System

65

which any significant events occurred with the robot (or with its participation). Data on the location of objects relative to the mobile robot enters the architecture from LiDAR and ultrasonic sensors installed on the robot. The corresponding agneurons-sensors are responsible for processing this information. A conceptual object agneuron is created in the architecture (Fig. 7), information about the location of this object is stored by a locative agneuron, and information that this object is in a given place and at a given time is contained in the event agneuron “Object is *locus at *time”.

Fig. 7. Object type agneurons.

In order to find out which object is in front of it, the intelligent agent addresses the user. Figure 8 shows a functional representation of the context and content of the dialogue, in which the user explains to the intelligent agent that the object they both observe in the external environment is a “table” [26]. In this case, in a multi-agent neurocognitive architecture, the context of the interactive interaction of an intelligent agent with a user is registered by the following events:

Multi-agent connections are formed between events and agneurons that perform a functional representation of these events, aimed at reflecting the cause-and-effect relationships between events that describe the context of the situation, events that describe the subject (theme and rhema) of the utterance, and an event that describes the utterance itself. Such connections are hypothetically able to provide the construction of functional systems of semantic ontology, understanding and synthesis of natural language statements [26], as well as the integration of computer vision and language.

66

I. Pshenokova et al.

Fig. 8. Functional representation of the context and content of the dialogue in a multi-agent neurocognitive architecture.

Autonomous Robot Navigation System

67

3 Experiments Figure 9 shows the editor of multi-agent neurocognitive architectures. Event local agents are represented, which form relations between locus, time, action, and subject. The picture shows that these agents conceptualize predicative constructions describing the facts about local information in the form of “I am at *locus in *time” and “Object is at *locus in *time”. A chat window is presented in the lower left corner, in which you can see as agneurons exchange information with each other and the user.

Fig. 9. Multi-agent neurocognitive architecture of the autonomous robot navigation system.

Such a multi-agent neurocognitive architecture, generating local and related software agents on demand when interacting with the environment, forms an active card containing all the local information necessary for the navigation between the loci. Information is presented in the form of interconnected predicative constructions, the elements of which are messages from neuron agents that conceptualize various parameters of locative and locus information. Such a multi-agent architecture should ensure the construction of an ontology of a map of the area and ensure the operation of the system of navigation and orientation in space for robots within the framework of the HMT.

4 Conclusion An autonomous robot navigation system based on self-organization of distributed neurocognitive architectures has been developed. An algorithm for the formation of contractual links between agneurons is presented, which considers the context of the dialogue situation, the content of the specific state of the “intelligent agent – environment” system, the content and form of the statement.

68

I. Pshenokova et al.

A multi-agent neurocognitive architecture is presented, which forms an active map containing all the locative information necessary to ensure the orientation and navigation of an autonomous robot between loci. The use of a multi-agent architecture to provide the representation of locative information in the task of implementing an interface in a HMT will allow building an ontology responsible for representing the location of objects in the external environment, as well as providing interaction with the user in natural language, taking into account its semantics. Further development of the approach is aimed at the implementation of the natural language interface of the HMT, which is not inferior in efficiency to the interaction between people.

References 1. Davenport, T.H.: The AI advantage: How to Put the Artificial Intelligence Revolution to Work. MIT Press, Cambridge (2018) 2. Russell, S., Norvig, P.: Artificial Intelligence: A Modern Approach, 2nd edn. Pearson Education Inc., Prentice-Hall (2010) 3. Stowers, K., Oglesby, J., Sonesh, S., Leyva, K., Iwig, C., Salas, E.: A framework to guide the assessment of human–machine systems. Hum. Factors 59, 172–188 (2017) 4. Salas, E., Rosen, M.A., Burke, C.S., Goodwin, G.F.: The wisdom of collectives in organizations: an update of the teamwork competencies. In: Team Effectiveness in Complex Organizations. Cross-Disciplinary Perspectives and Approaches, pp. 73–114. Routledge, NY (2009) 5. Klein, G., Woods, D.D., Bradshaw, J.M., Hoffman, R.R., Feltovich, P.J.: Ten challenges for making automation a “team player” in joint human-agent activity. IEEE Intell. Syst. 19, 91–95 (2004) 6. Ososky, S., et al.: The importance of shared mental models and shared situation awareness for transforming robots from tools to teammates. Unmanned Syst. Technol. 14(8387), 397–408 (2012) 7. Seeber, I., et al.: Machines as teammates: a research agenda on AI in team collaboration. Inf. Manage. 57(2), 103174 (2020) 8. You, S., Robert, L.: Emotional attachment, performance, and viability in teams collaborating with embodied physical action (EPA) robots. J. Assoc. Inf. Syst. 19(5), 377–407 (2018) 9. Lemaignan, S., Warnier, M., Sisbot, E.A., Clodic, A., Alami, R.: Artificial cognition for social human–robot interaction: an implementation. Artif. Intell. 247, 45–69 (2017) 10. Stowers, K., Kasdaglis, N., Rupp, M.A., Newton, O.B., Chen, J.Y., Barnes, M.J.: The IMPACT of agent transparency on human performance. IEEE Trans. Hum. Mach. Syst. 50(3), 245–253 (2020) 11. Nam, C.S., Lyons, J.B. (ed.): Trust in Human-Robot Interaction. Academic Press, London (2020) 12. Chao, C., Thomaz, A.: Timed Petri nets for fluent turn-taking over multimodal interaction resources in human-robot collaboration. Int. J. Robot. Res. 35, 1330–1353 (2016) 13. Tellex, S., Gopalan, N., Kress-Gazit, H., Matuszek, C.: Robots that use language. Ann. Rev. Control Robot. Auton. Syst. 3, 25–55 (2020) 14. Chen, J.Y., Lakhmani, S.G., Stowers, K., Selkowitz, A.R., Wright, J.L., Barnes, M.: Situation awareness-based agent transparency and human-autonomy teaming effectiveness. Theor. Issues Ergon. Sci. 19, 259–282 (2018)

Autonomous Robot Navigation System

69

15. Malone, T.W., Crowston, K.: What is coordination theory and how can it help design cooperative work systems? In: Proceedings of the 1990 ACM Conference on Computer-Supported Cooperative Work, pp. 357–370. Association for Computing Machinery, NY (1990) 16. Klein, G., Feltovich, P.J., Bradshaw, J.M., Woods, D.D.: Common ground and coordination in joint activity. Organ. Simul. 53, 139–184 (2005) 17. Nagoev, Z.V.: Intelligence, or thinking in living and artificial systems. KBNTs RAS Publishing House, Nalchik (2013) 18. Nagoev, Z., Bzhikhatlov, K.C, Nagoeva, O., Anchekov, M., Atalikov, B.: Models of orientation and navigation of autonomous household robots based on multi-agent neurocognitive architectures. In: Perspective systems and control tasks. Materials of the XVI All-Russian scientific-practical conference and the XII youth school-seminar, pp. 154–161. IP Maruk M.R., Rostov-on-Don (2021) 19. Nagoev, Z., Pshenokova, I., Gurtueva, I., Bzhikhatlov, K.: A simulation model for the cognitive function of static objects recognition based on machine-learning multi-agent architectures. In: Samsonovich, A.V. (ed.) BICA 2019. AISC, vol. 948, pp. 370–378. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-25719-4_48 20. Nagoev, Z., Pshenokova, I., Nagoeva, O., Sundukov, Z.: Learning algorithm for an intelligent decision-making system based on multi-agent neurocognitive architectures. Cogn. Syst. Res. 66, 82–88 (2021) 21. Nagoev, Z., Nagoeva, O., Gurtueva, I.: Multi-agent neurocognitive models of semantics of spatial localization of events. Cogn. Syst. Res. 59, 91–102 (2020) 22. Nagoev, Z., Gurtueva, I., Malyshev, D., Sundukov, Z.: Multi-agent algorithm imitating formation of phonemic awareness. In: Samsonovich, A.V. (ed.) BICA 2019. AISC, vol. 948, pp. 364–369. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-25719-4_47 23. Nagoev, Z., Nagoeva, O., Gurtueva, I., Denisenko, V.: Multi-agent algorithms for building semantic representations of spatial information in a framework of neurocognitive architecture. In: Samsonovich, A.V. (ed.) BICA 2019. AISC, vol. 948, pp. 379–386. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-25719-4_49 24. Nagoev, Z., Nagoeva, O., Pshenokova, I., Gurtueva, I.: Multi-agent model of semantics of simple extended sentences describing static scenes. In: Ronzhin, A., Rigoll, G., Meshcheryakov, R. (eds.) ICR 2019. LNCS (LNAI), vol. 11659, pp. 245–259. Springer, Cham (2019). https:// doi.org/10.1007/978-3-030-26118-4_24 25. Nagoev, Z., Lyutikova, L., Gurtueva, I.: Model for automatic speech recognition using multiagent recursive cognitive architecture. Procedia Comput. Sci. 145, 386–392 (2018) 26. Nagoev, Z.V., Nagoeva, O.V.: Justification of symbols and multi-agent neurocognitive models of natural language semantics. KBNTs RAS Publishing House, Nalchik (2022)

3D-CNNs-Based Touchless Human-Machine Interface Ali Asgarov and Ali Parsayan(B) Process Automation Engineering Department, Baku Higher Oil School, 30, Khojaly Avenue, Baku 1025, Azerbaijan [email protected]

Abstract. Interacting with machines via hand gestures is a common way for people to communicate with robots. Human utilize gestures in a regular talk to convey meaning and emotions to one another. Gesture-based interactions are utilized in a wide range of applied to a wide range of fields, as telephones, TVs, monitors, video games, and other electronic devices. By technological improvements, gesture recognition is now a more realistic and appealing approach in the context of human interaction. In this research, the relevant experiments are conducted using numerous types of convolutional neural networks, including the proposed customized model, to see which ones performs the best. Because of the introduction of such Microsoft Kinect sensor, increased depth and vision sensing has been widely important for several purposes. Given its ability to measure ranges to objects at a fast frame rate, these types of sensors are widely being employed for 3D acquisitions, as well as for other purposes in robotics and machine learning. This research made use of the Kinect sensor and the use of an RGB-D camera and a 3D convolution neural network, which offer a novel approach for fingertips identification and hand gesture classification in real time that is both accurate and fast (3DCNN). Keywords: Hand Gesture Recognition · 3DCNN · HMI · HCI · Robots

1 Introduction As information technology progresses, human desire to communicate with robots and computers in a natural way rises in importance. Because of their limited flexibility, traditional human-computer interaction input devices such as mouse, keyboards, and remote controllers are no longer natural modes of communication. This isn’t a new phenomenon. They’ve been around for quite some time. Two of the most frequent ways for individuals to communicate with robots and computers are through voice commands and body language, both of which can be found in a wide range of applications. As a result, getting accurate results when utilizing automatic speech recognition in a noisy environment can be problematic due to the considerable variation in how people pronounce common terms. Additionally, the use of body language may be beneficial in the development of a human–computer interaction. In many circumstances, this sort of communication is more reliable than other types of communication. Hand gestures, body positioning, © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 70–80, 2023. https://doi.org/10.1007/978-3-031-43111-1_7

3D-CNNs-Based Touchless Human-Machine Interface

71

and facial expressions are just a few instances of language to consider. Gestures with one’s hands are the most effective method of communicating one’s ideas and feelings. In addition, they serve as a universal language for human to communicate with. There are various ways to say “hi”, such as waving your hand in front of someone’s face or saying it out loud [1–3]. It is essential for people to communicate with one another via hand gestures while they are together. In communicating with individuals around us, we make hand gestures that express crucial information and ideas to them. Without making any sort of gesture or expressing anything to someone or something else with our hands, it is hard to carry on a meaningful dialogue with them. Gestures with the hands are a frequent way of transmitting information to others. According to some projections, the touchscreen on mobile phones may be phased out in the near future and replaced with hand gestures. Systems for human–computer interaction (HCI) include things like sign language recognition, robot control, virtual mouse control, medical imaging data research, and immersive gaming. Human–vehicle interaction (HVI) and immersive gaming are also included in this category. As a result, computer vision experts have been attempting to identify hand motions in video for quite some time, particularly with older, less capable cameras (RGB cameras). A huge number of persons, shifting light levels, complicated backgrounds, varying user–camera distances, fluctuating user–camera distances, and background/foreground motions during hand tracking are all examples of situations in which many existing algorithms fail. A new version of Kinect RGB-D, Microsoft’s upgraded camera for gaming and other applications, has been released with a revised user interface and depth-sensing capabilities. Color and depth sensors are used to create synchronized photos with the Kinect sensor. In the beginning, users’ body motions were used to communicate with video games, and Microsoft used them as an input method for its Xbox gaming device. The computer vision community has extended the use of Kinect v1’s low-cost depth sensing technology well beyond the realm of gaming. The ability to interface with other devices or applications without the need to touch a controller has enabled human motion capture systems to be easily repurposed for a wide range of new applications, ranging from medical to robotics. These developments may now be of use to applications involving human motion analysis [1–3]. After being introduced to computer science in the 1990s, convolutional neural networks (CNNs) have made significant contributions to the field of human-computer interaction (HCI) [2]. With a large number of layers in CNNs, it is possible to learn features more quickly and efficiently. Data may be represented hierarchically by using deep learning frameworks, where characteristics are ordered from the lowest to the highest level, resulting in hierarchical representations of data. The ultimate goal in this research is to develop a depth video-based hand gesture interface, which is the subject of this study. In this experiment, fingertip detection and a 3D convolutional neural network are utilized to detect hand motions captured by a depth camera. Using multi-level characteristics and the right utilization of multi-level features, the network architecture under consideration may provide high-performance predictions while also exploiting multi-level characteristics. This research made use of the Kinect sensor and the use of an RGB-D camera and

72

A. Asgarov and A. Parsayan

a 3D convolution neural network, which offer a novel approach for fingertips identification and hand gesture classification in real time that is both accurate and fast (3DCNN) [1–3].

2 Related Works When it comes to human-to-human communication, hand gestures are extremely important [3]. This type of communication has resulted in the development of new concepts for human-computer interaction as a result of its high level of information transfer efficiency. In order for this to operate, the computer must be able to recognize the gestures made by the person who is running the machine. That is all that is necessary for the identification of hand gestures to work. When it comes to resolving these types of problems, feature extraction algorithms are the most commonly used method [2]. Using a specialized method, the hand image is matched up with a preset template image. It has been demonstrated that template matching is ineffective when dealing with a diverse range of environments, hand shapes, and actions. When it comes to classification algorithms, different feature extractors are unable to adapt to new datasets and changing scenarios effectively. Deep convolutional neural networks are excellent alternatives in these cases because of their robustness and invariance. Deep learning techniques are becoming increasingly popular when it comes to computer vision applications. By convolutional neural networks, the parallel nature of the computations enables them to be applied elegantly to the matrix representation of data when dealing with problems that can be represented visually. When employing multicolumn deep CNNs that leverage several parallel networks [4], it has been demonstrated that the recognition rates of single networks for a variety of photo classification tasks rise by 30 to 80%. Once again, convolutional neural networks (CNNs) were used by Neverova et al. to identify between 20 different Italian sign language movements by combining RGBD data from the hand area with upper-body skeletal movement [5]. While their system, on the other hand, was intended to be used only for indoor activities, Pablo Barros and colleagues developed a Multichannel Convolutional Neural Network (MCNN) that recognizes hand gestures and extracts implicit features from the architecture itself, a process known as implicit feature extraction, from the architecture [6]. To reach their results, the researchers examined data from two existing datasets of static hand motion photos. A robot was utilized to create the first dataset in a lab environment, using four different types of hand gestures. It was done in a similar manner with the second dataset. They used a new set of data to evaluate their technique, which comprised ten distinct hand gestures performed in natural, uncontrolled environments. Hand-gesture detection in cars using RGBD data was studied by many researchers. They discovered that the best values were achieved when histogram of gradient (HOG) characteristics were combined with an SVM classifier. Molchanov et al. combined data on hand movements gleaned from depth, color, and radar sensors while simultaneously training a CNN to make use of the data. They were able to correctly identify under a variety of lighting and environmental settings [5]. As a result, the actions detailed above provided the necessary framework for our research and inspired us to keep going.

3D-CNNs-Based Touchless Human-Machine Interface

73

3 Proposed Method Figure 1 depicts a high-level representation of the proposed system. Microsoft Kinect Sensor version 2 pictures with in-depth skeleton-joint information are used to extract the hand region of interest, and a border-tracing method is used to extract and characterize the contours of hands. Using the hand-contour coordinate model, the K-cosine algorithm detects the fingertip location, which is then translated into a gesture initialization to help identify hand movements. When a gesture is finally identified, a 3D convolutional neural network is used to do the process.

Hand Region Extraction

Fingertip Detection

Hand Gesture Spotting

Hand Gesture Recognition

Kinect v2

Fig. 1. The proposed system’s block diagram.

3.1 Extraction of the Hand Region The extraction procedure for the hand region is depicted in Fig. 2, and we’ll go over it in more detail in the following paragraphs. The depth images of the hand region of interest and the middle of the palm that are provided by the Kinect V2 skeletal tracker are then converted to binary images. In order to extract and describe the hand outlines, a method called border tracing is utilized. Each of the user’s body parts is detected using a Microsoft Kinect V2 sensor, which uses depth pictures to map each of the learnt body parts to the depth images as the user moves. Therefore, the camera may acquire skeleton-joint information for 25 joints, such as the hip, spine, head, shoulder, hand, foot, and thumb. Kinect skeletal tracker depth images are used to accurately extract the hand region of interest (HRI) and palm center. In order to get rid of the noise in the hand region, the Weiner filter is utilized in conjunction with morphological processing. The findings are then transferred to the binary image using the thresholds that were specified using the depth signals provided by the Kinect. The Moore–Neighbor approach may be used to produce hand contours from binary photographs of the hand regions. These hand contours can then be used to draw the hand. Utilizing this method allows for the performance of image contour extraction, also known as area extraction. As a consequence of carrying out these steps, we now have a collection of hand contour pixels. These data are used to measure the curvature of the surface.

74

A. Asgarov and A. Parsayan Depth Images No Skeleton Information Yes Hand Detection and Segmentation

Contour Extraction

Fig. 2. Hand region extraction flowchart.

3.2 Moore–Neighbor Approach The Moore Neighboring Algorithm (Fig. 3) is what is used to carry out the process of edge detection. The average and the highest possible value for each pixel are calculated using this procedure. The use of cellular automata [2] is a well-known method that enhances the overall quality and functionality of noisy images. Modified Moore-Neighborhood (P) of an 8-pixel pixel, shares a vertex or edge with the other 8-pixel pixels. A white pixel will be found by P if and only if it finds one in its near surroundings. There are no white pixels in the immediate neighborhood, thus the operation continues. Setting a suitable end point is crucial when utilizing MooreNeighbor tracing.

Fig. 3. The Moore Neighboring window consists of 9 cells, core cell and 8 stranding cells.

3.3 Detecting Hand Gestures Target-Person Locking When there are a large number of people in the scene, the individual whose movements are being monitored will have control of the mouse. As can be seen in Fig. 4, the Kinect

3D-CNNs-Based Touchless Human-Machine Interface

75

V2 sensor was used in this research to extract 25 skeletal joints from up to six people at the same time. These joints include the head, the neck, the left and right hands, the base of the spine, and so on. Because of this, the procedure that was just detailed enables the system to locate the fingers of anywhere from one to six separate users. On the other hand, in order to control the mouse cursor, we must first determine who the intended recipient is. In order to solve this problem while hand tracking was being done, a user-locking mechanism was used.

Fig. 4. Kinect Sensor.

The first algorithm presents an illustration of how to put this plan into effect. Using the Kinect skeletal tracker’s head and right-hand joint coordinates, the target person is determined based on the distance between their heads and hands, as illustrated in following algorithm. Ten frames into the video, the user extends his hand over his head. However, determining the beginning and ending points of a hand gesture in real-time hand gesture detection is still difficult. The Target Person Locking Algorithm is: 1. starts. 2. Require the User ID, the x1, y1 coordinates of the head, and the x, y1 coordinates of the right hand joint (x13, y13). 3. Firstly, get the distance between y1 and y13 using the formula. 4. Secondly, y13 > y1 in 10 frames, else process target user. 5. Otherwise, redo first 6. ends, begin. A region in which a person utilizing a Kinect device may move the mouse cursor with their hands instead of using the mouse. This idea has the benefit of being able to be carried out on a number of screen sizes and resolutions, which is a distinct advantage. A simple glimpse at a virtual screen is all that’s needed to take control of the movements. At this point, the depth resolution of the Kinect V2 sensor is taken into account in order to establish the resolution of the virtual screen, which comes out to 512 × 424 pixels.

4 Simulations, Experimentations and Discussion of the Results Hardware Specifications. Kinect Camera: Officially, the Xbox 360 console is referred to as Kinect (14). A partnership between Microsoft and Prime Sense led to the creation of

76

A. Asgarov and A. Parsayan

the application in June of that year. Microsoft Windows operating system consist of this application was introduced in February 2012. Its introduction in 2009 had a significant influence on the Computer Vision and Computer Graphics fields. As a result of this Microsoft innovation, gamers may use speech and gestures to engage in a wide variety of ways in the game. Since its first release, the Kinect has grown in popularity across a variety of industries, including video games, theater, robotics, and natural interaction. In order to verify the efficacy of the proposed method, a dataset consisting of hand gestures that had been manually categorized was developed. The hand gesture films included in the collection were created through the use of fingertip identification derived from depth videos. According to the depiction, there are 5250 videos of each move made by fifty different people. Implementation. On a computer equipped with an Intel I7 10th Generation 10800U Six-Core Processor operating at 2.4 GHz, 12 gigabytes of random-access memory (RAM), and a GeForce MX250 (GPU), the proposed system was constructed. Python was the programming language which is used to build the system from the ground up. There were 30 frames taken per second as part of the tracking procedure. The 3DCNN model was implemented with the help of the deep learning frameworks Keras and Scikit-learn. Analysis of the Experiment. From the original data, 65% of it was used for training, 15% for validation, and 20% for testing in hand gesture recognition. Ensemble Decision Tree, Support Vector Machine (SVM) [7] and two-dimensional convolution neural network [8] were utilized to evaluate our proposed method for identifying hand gestures in videos (2DCNN). Convolutional neural networks (CNNs) were used in the network of the 2DCNN instead of 3DCNN, which employs 3D convolutions. 2DCNN input was similar to the standard ML model (EDT). The Scikit learn 0.20 module was used to build an SVM nonlinear multiclass approach with a kernel based on the intersection of histograms. Evaluation Metrics. The algorithms were compared using three criteria. The accuracy rate was included to account for situations in which the number of samples in each class is exactly equal, the number of accurate predictions to the total number of input samples is known as the predictive accuracy ratio (EPR), and the training and testing times of the deep learning model which were the second and third measures. Online training with data collection was termed a scenario. For each dataset, 100 training epochs was ran with the goal of improving classification accuracy. This procedure was tested five times, and the average of the findings was used to compute the testing duration. Discussion of the Results. The accuracy of the classification performed by 3DCNN model on both the training and validation sets is demonstrated in Fig. 5. The plots provide evidence that the model provides a satisfactory match to the issue. By the use of prepared hand gesture dataset, the results of the proposed method were compared with the reviewed methods results which are considered to be state of the art in Table 1. In terms of accuracy rate, the overall performance of Deep Learning algorithms was superior to that of the traditional Machine Learning models. In comparison, the accuracy achieved by the Ensemble Decision Tree, SVM, and CNN models are 10%,

3D-CNNs-Based Touchless Human-Machine Interface

77

Fig. 5. Accuracy rate of the classification performed by 3DCNN model on both the training and validation sets.

20%, 27%, and 30% lower than that achieved by 3DCNN. The SVM model achieved a score of 63%, while 2DCNN model achieved 65%. On the prepared hand gesture dataset, 3DCNN network achieved superior results in terms of video classification compared to SVM, EDT (Ensemble Decision Tree), KNN and 2DCNN. It’s probable that this discovery has a rational explanation, and that explanation is the flexibility of a decision boundary supplied by learning numerous frames in 3DCNN. 3DCNN took longer time to train and test than either the Ensemble Decision Tree, KNN, SVM or 2DCNN for convolutional neural networks. This fact showed that neural networks may be employed in real-time applications, since all designs were able to identify the hand motion in less than 7 s. Table 1. Comparison of different models results. Models

Time for Training

Time for Testing

Accuracy

Ensemble Decision Tree

25.3 m

6.4 s

83.3%

KNN

29.4 m

5.2 s

73.4%

SVM

23.5 m

5.7 s

63.4%

2DCNN

40.5 m

4.1 s

65.3%

3DCNN

97.3 m

7.3 s

93.2%

Table 2 depicts the proposed method hand motions recognition rate, which relies on ensemble learning. As a collection of 20 3DCNN models, the ensemble learning approach outperformed the single 3DCNN in video classification, with an accuracy rate of 97.4%, according to the study. The individual 3DCNN models in this dataset were shown to be considerably less accurate than the ensemble of 5 3DCNN models. But, when the number of models is raised to 10 or 20, there is no visible change. In contrast, ensemble learning necessitates a lengthy training period because of the high processing costs associated with a large number of models.

78

A. Asgarov and A. Parsayan Table 2. Comparison of ensemble approach for 3DCNN.

Models

Accuracy (%)

3DCNN

93.2

Ensemble with 5 3DCNN networks

94.5

Ensemble with 10 3DCNN networks

95.6

Ensemble with 15 3DCNN networks

96.1

Ensemble with 20 3DCNN networks

97.5

Table 3 is an illustration of the confusion matrix produced by 3DCNN using an ensemble of 20 models. As it was to be expected, zooming in and zooming out (ZI) had the lowest accuracy rate, but swiping left, swiping right, and selecting “meaningless” (M) all had the highest accuracy rate. Because of the rapid hand movement, it was difficult to detect the difference between the ‘drag’ gesture and other motions. This was because the ‘drag’ gesture was quicker than other gestures, making it more difficult to differentiate between the two. Table 3. Confusion Matrix. Target

Selected

Acc

S_L

S_R

S_U

S_D

Z_I

Z_O

M

S_L

150

0

0

0

0

0

1

0.991

S_R

1

145

0

0

0

0

1

0.981

S_U

1

0

145

0

2

0

2

0.972

S_D

0

0

0

146

1

1

2

0.973

Z_I

0

0

0

0

132

15

3

0.954

Z_O

4

4

1

5

4

129

4

0.945

M

1

0

0

0

0

1

145

0.991 0.975

5 Conclusion In order to identify and recognize fingertip gestures, this research used a mix of a geometry algorithm and a deep learning method given in this paper. The proposed method was able to accurately predict gestures, but it also had the potential to be used in real-world contexts. Fingertip detection with better threshold segmentation using depth information and the k-cosine curvature technique is both discussed in this work in an attempt to overcome the distance restriction in gesture identification.

3D-CNNs-Based Touchless Human-Machine Interface

79

In addition to functioning effectively in a variety of lighting conditions and with various backdrop patterns, the suggested approach is capable of accurately detecting hand movements even at a distance of several meters, and it can even recognize the hand motions of many persons. As the further work, the performance of the suggested method can be improved by carrying out a holistic search in an effort to optimize each of the hyperparameters individually. Also, the results of the experiments showed that this methodology is a potential method for real-time hand-gesture interfaces. Also, the proposed system can be improved to accommodate additional hand motions in the future and use it in more practical applications.

References 1. Hasan, H., Abdul-Kareem, S.: Human–computer interaction using vision-based hand gesture recognition systems: a survey. Neural Comput. Appl. 25, 251–261 (2014) 2. Klompmaker, F., Nebe, K., Fast, A.: dSensingNI: a framework for advanced tangible interaction using a depth camera. In: Proceedings of the Sixth International Conference on Tangible, Embedded and Embodied Interaction, pp. 217–224 (2012) 3. Gallo, L., Placitelli, A.P., Ciampi, M.: Controller-free exploration of medical image data: experiencing the Kinect. In: Proceedings of the 2011 24th International Symposium on Computer-Based Medical Systems (CBMS), pp. 1–6 (2011) 4. Krejov, P., Bowden, R.: Multi-touchless: real-time fingertip detection and tracking using geodesic maxima. In: Proceeding of 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), pp. 1–7 (2013) 5. Freeman, W.T., Roth, M.: Orientation histograms for hand gesture recognition. In: International Workshop on Automatic Face and Gesture Recognition, vol. 12, pp. 296–301 (1995) 6. Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556 (2014) 7. Fossati, A., Gall, J., Grabner, H., Ren, X., Konolige, K.: Consumer Depth Cameras for Computer Vision: Research Topics and Applications. Springer, Heidelberg (2012) 8. Dao, T.T., Tannous, H., Pouletaut, P., Gamet, D., Istrate, D., Tho, M.H.B.: Interactive and connected rehabilitation systems for e-Health. Irbm 37(5–6), 289–296 (2016) 9. Reza, M.N., Hossain, M.S., Ahmad, M.: Real time mouse cursor control based on bare finger movement using webcam to improve HCI. In: 2015 International Conference on Electrical Engineering and Information Communication Technology (ICEEICT), pp. 1–5. IEEE (2015) 10. Pisharady, P.K., Saerbeck, M.: Recent methods and databases in vision-based hand gesture recognition: a review. Comput. Vis. Image Underst. 141, 152–165 (2015) 11. Amin, M.A., Yan, H.: Sign language finger alphabet recognition from Gabor-PCA representation of hand gestures. In: 2007 International Conference on Machine Learning and Cybernetics, vol. 4, pp. 2218–2223. IEEE (2007) 12. Sharma, A., Mittal, A., Singh, S., Awatramani, V.: Hand gesture recognition using image processing and feature extraction techniques. Procedia Comput. Sci. 173, 181–190 (2020) 13. Lian, S., Hu, W., Wang, K.: Automatic user state recognition for hand gesture based low-cost television control system. IEEE Trans. Consum. Electron. 60(1), 107–115 (2014) 14. Sharp, T., et al.: Accurate, robust, and flexible real-time hand tracking. In: Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems, pp. 3633–3642 (2015)

80

A. Asgarov and A. Parsayan

15. Fossati, A., Gall, J., Grabner, H., Ren, X., Konolige, K.: Consumer Depth Cameras for Computer Vision: Research Topics and Applications. Springer, Heıdelberg (2012) 16. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778 (2016) 17. Bagdanov, A.D., Del Bimbo, A., Seidenari, L., Usai, L.: Real-time hand status recognition from RGB-D imagery. In: Proceedings of the 21st International Conference on Pattern Recognition (ICPR2012), pp. 2456–2459. IEEE (2012) 18. Tsai, T.H., Huang, C.C., Zhang, K.L.: Embedded virtual mouse system by using hand gesture recognition. In: 2015 IEEE International Conference on Consumer Electronics-Taiwan, pp. 352–353. IEEE (2015) 19. Wang, R.Y., Popovi´c, J.: Real-time hand-tracking with a color glove. ACM Trans. Graph. (TOG) 28(3), 1–8 (2009) 20. Wang, P., Li, W., Ogunbona, P., Wan, J., Escalera, S.: RGB-D-based human motion recognition with deep learning: a survey. Comput. Vis. Image Underst. 171, 118–139 (2018) 21. Grif, H.-S., Farcas, C.C.: Mouse cursor control system based on hand gesture. Procedia Technol. 22, 657–661 (2016) 22. Andersen, M.R., et al.: Kinect depth sensor evaluation for computer vision applications. Aarhus University, pp. 1–37 (2012) 23. Shou, Z., Wang, D., Chang, S.-F.: Temporal action localization in untrimmed videos via multi-stage CNNs. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1049–1058 (2016) 24. Molchanov, P., Yang, X., Gupta, S., Kim, K., Tyree, S., Kautz, J.: Online detection and classification of dynamic hand gestures with recurrent 3D convolutional neural network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4207– 4215 (2016)

Development of a Mechanism for Recognizing the Emotional State Based on the Unconscious Movements of the Subject Yaroslava Gorbunova1,2(B)

and Gleb Kiselev1,2

1 Artificial Intelligence Research Institute, Federal Research Center “Computer Science and

Control” of Russian Academy of Sciences, 44/2, Vavilova st., 119333 Moscow, Russia [email protected] 2 Peoples’ Friendship University of Russia named after Patrice Lumumba, 6, Miklukho-Maklaya st., Moscow 117198, Russia

Abstract. The task of assessing emotional state is not easy for a human and especially challenged for an automated system. It requires not only detection but also complex analysis of various factors. The effectiveness of existing algorithms based on such modalities as text, audio, video, physiological characteristics, etc., depends on the subject’s race, language, and other affiliations, which makes it difficult to study in case of lack or complete absence of such data. The paper deals with the creation of a mechanism capable of detecting the emotional state of a person based on his interaction with a computer via a mouse. To solve the problem, we collected a dataset using a web application developed in the Python programming language. The application is focused on collecting data about the subject’s cursor movement such as distance travelled, maximum deviation from the line that connects the start and end points of the movement, the time the subject interacted with the computer mouse during the session and the maximum speed of the cursor movement. Trained classifiers for emotion analysis based on human control of the computer mouse on the created dataset and analyzed the results. Keywords: Emotion Analysis · Emotional Coloring · Computer Mouse · Cursor Movements · Dataset Creation · Machine Learning

1 Introduction The task of recognizing and detecting emotions is relevant for decades. Knowledge of the emotions experienced by a subject can be useful in many fields, e.g., marketing, education, robotics, banking, advertising, and others. People’s moods can be recognized through many distinctive abilities. The most common are facial expressions, gestures, voice volume and interruption, and physiological characteristics (sweating, heart rate, skin conductivity, etc.). A human is capable of recognizing emotion from a person’s facial expression with a high degree of accuracy, but for an automated system, it is a highly complex task. A computer needs to take many factors into account when identifying moods, and there are other problems as well. Cultural differences are also important © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 81–92, 2023. https://doi.org/10.1007/978-3-031-43111-1_8

82

Y. Gorbunova and G. Kiselev

in analysis because similar facial expressions in different cultures can sometimes mean the opposite of emotion, leading to a deliberately false assessment. Furthermore, facial expressions may not reflect a person’s real feelings, or may reflect false ones. Therefore, there is a need to create a new way of identifying and analyzing emotions that can be applied to different ethnic and linguistic groups, avoiding the difficulties of existing mechanisms. The diversity of data types is leading to new and improved analysis methods. Today, there are many approaches to detect the psycho-emotional state of a person, capable of analyzing speech recordings, digitized speech recognition results, textual data, images, and video materials. Textual information can serve as the main source of data about a person’s state. The application of machine learning in text analysis leads to an accuracy of over 90% [1]. Also, high quality of emotion classification can be achieved through the study of audio signal [2, 3] and combining the modalities, audio, and video [4]. Video-based emotion analysis also makes it possible to study the relationship between cognitive function and emotion in humans, a particularly acute issue in the elderly [5]. Video streaming also makes it possible to assess the emotional state of a car driver [6], which will help to avoid accidents and keep all road users safe from dangerous situations. For more detailed research in emotion recognition, there are datasets that include several modalities at once [7], such as audio, video, oculography, speech, motion capture data, and others. The emotions evoked in a person, intentionally or not, leave a trace in their physiology. In this way, information on afferent muscle excitation [8], skin conductance [9], functional magnetic resonance imaging (fMRI) data, respiration rate [10] can be used to determine a subject’s emotional state, sadness, happiness, joy, and others. To analyze emotional states, it is possible to use different modalities, which is a wide field for researchers.

2 The Need for a New Modality Modalities for recognizing emotional states such as text, audio, video, or characteristics of human physiology are quite informative sources of knowledge about a subject, but each provides advantages and disadvantages in the accumulation, transformation, and processing of data. There are more than 7000 languages in the world, many of which have dialects. However, only a seventh can be machine-processed, making it difficult to analyze the emotions of the rest of the language space. In addition, textual and audio modalities are not always available for processing, nor are video, photo and physiological characteristics. This becomes the reason for seeking out and analyzing new sources of data. One of them is unconscious actions of a subject while interacting with a computer, in particular is with a computer mouse. During this interaction, a person experiences many emotions acquired during work and from the environment, and at the same time makes hundreds of cursor movements, which can be indicative of the subject’s internal state. The creation of an emotion recognition mechanism based on unconscious activity is driven by the need to monitor the emotional well-being of the population to identify problematic situations in the early stages of development. The implementation of the mechanism will

Development of a Mechanism for Recognizing the Emotional State

83

solve the problem of mental ill-health of the subject (for example, emotional burnout, fatigue, depression, etc.) by identifying bad moods, irritability at the initial stage of their manifestation through computer mouse movements. However, the lack of freely available data for studying the problem entails the need to accumulate such a set. It will help to identify and assess the subject’s emotions regardless of language, race or other affiliations. A person’s movements are constantly changing according to the senses they are experiencing. If we consider computer mouse movement, the subject continuously processes data about the desired hand location and its difference from the actual location, while generating the necessary motor commands to achieve the goal [11]. Neurological movement disorders, such as Parkinson’s and Tourette’s diseases, prove the assumption that cursor movements in a choice task are influenced by emotion [12]. Experienced arousal in online shopping influences the duration of movements and quantitative changes in computer mouse speed [13]. Decision making and change in emotion can also be encountered when using not only a computer mouse, keyboard, and other things, but also when using a smartphone. The gadget can be applied to study changes in levels of joy while interacting with it [14]. An important aspect of the analysis is to identify and define the indicators to look for. To obtain a plausible assessment of a subject’s emotional state, it is necessary to understand what indicators to look for. A study of the impact of negative emotions on cursor speed and distance travelled can help in this regard [15]. Besides the distance travelled and the mouse cursor speed, the use of such spatial characteristics of the trajectory as AUC (area under the curve), and zigzag (the number of trajectory deviations from a straight line connecting the start and the end points) are considered [12]. The analysis of emotional state based on the subject’s unconscious actions, such as mouse movement, interaction with a smartphone, etc., is an innovative approach for this task. The versatility of the mechanism lies in its independence from the subject’s race, language, and other affiliations, which makes the modality most suitable for investigating psycho-emotional states in the absence or deficiency of such data as text, audio, video, and others. Both spatial characteristics such as distance travelled, curve deviation and temporal characteristics such as total mouse interaction time per session can be considered as properties of cursor movement for the task. This allows us to analyze how well the emotion scores are based on computer mouse activity, and to extend the dataset with additional modalities to achieve better results.

3 Creation of a Mechanism for Recognizing the Emotional State of the Subject Based on His Unconscious Actions 3.1 Designing a Mechanism for Recognizing the Emotional State of the Subject The first step in implementing a mechanism for recognizing a person’s psycho-emotional state is to design the mechanism. For successful design it is necessary to understand who will interact with the system, which blocks are needed to implement the analysis, and how the blocks will be linked together. The structure of the emotion recognition mechanism is shown in Fig. 1.

84

Y. Gorbunova and G. Kiselev

Fig. 1. Structure of the mechanism for recognizing the emotional state of the subject.

The person who will communicate directly with the system is the user who controls the mouse when performing tasks. The emotion recognition mechanism consists of two main units: data collecting and processing. Data collection for analysis in the current study is done manually, this is due to the lack of freely available datasets that provide information on human-computer interaction by controlling the mouse cursor. An algorithm for creating an in-house dataset, including the coordinates of the subject’s cursor position and an estimate of his or her emotional state, developed and implemented. This mechanism will be described in more detail in Sect. 3.2. After the necessary data is received from the user, it is fed for processing to the next unit, where the estimation of the emotional state of the user interacting with the system is performed based on the spatial characteristics of the mouse cursor position and the accumulated experience of the mechanism. A detailed description of the data processing algorithm is presented in Sect. 3.2. The output of the mechanism is the classification of the user as a subject with a positive, negative, or neutral emotional state. Based on the description of the human emotion recognition mechanism, it can be deduced that the input data for emotion analysis are the spatial characteristics of the mouse cursor position, the output data are the resulting assessment for the user. Such inputs have nothing to do with the subject’s language, appearance, or other factors, confirming the universality of the emotional state analysis mechanism. In the following chapters, the data collection and processing algorithm and the training of neural networks for classifying participants are described in detail. 3.2 Collect Spatial Characteristics of the Mouse Cursor Position and Create Own Dataset The task of analyzing the emotional state of a user is one of the classification family, where, based on a training sample, the model identifies relationships between dependent and independent variables to enable further accurate prediction of results based on the data obtained for the first time. The classification task using information technology exists for decades, helping people to perform analysis in very different areas of life. A training dataset is a necessary part of the classification task.

Development of a Mechanism for Recognizing the Emotional State

85

The problem of emotion analysis by computer mouse movements is new and accumulating and canonizing such data is a time-consuming process. Therefore, such training sets are not publicly available, making analysis in this area difficult. To implement collection of data about a subject’s cursor location during interactions with a computer, it is necessary to design and recreate a user interface that is free of distractions so that the evaluation of emotional state can be more reliable. It is also necessary to organize the data collection itself and the mechanism of interaction between the handler program and the user interface. This work is done in the data collection unit, the structure of which is shown in Fig. 2. As a platform for research participants to interact with the interface as well as to save the coordinates of the cursor location in parallel, it is decided to create a web application, which is the block of the engine responsible for data collection. It is a structure consisting of the following elements: 1. A database into which the details of each participant are placed. 2. A server, written in the Python programming language, responsible for processing client-side requests, accumulating and transforming received data, and providing feedback. 3. Web-interface is an interface between a client (the page with which a computer user interacts) and a server, designed for the stable functioning of a web application. The interface is written using the FastAPI web framework [16] for Python.

Fig. 2. Structure of the “Data collecting” block of the emotion recognition mechanism of the subject.

The participant never interacts directly with the server, it is always done by sending GET, POST and other requests going through the web interface and getting a response. These responses can be either web pages that are displayed to the user, status codes, success messages, etc. A web application is a complex structure that combines a client and server side, as well as a web interface. First, let’s look at what the client side is like. In the first step of interacting with the web application the user sees a welcome page with introductory instructions and instructions for further actions. The second page is a small survey to understand whose data will be processed in the future.

86

Y. Gorbunova and G. Kiselev

When the user completes registration in the system (by pressing the “Continue” button), a page is made available with an image that has the potential to induce a person to experience a particular emotion. These images are taken from the open affective image database OASIS [17]. There is no limit to the amount of time a participant can view the page. The participant is then asked to rate their psycho-emotional state on an emotion scale: positive, rather positive, neutral, rather negative, negative. After viewing the affective image and self-assessment, a simple choice task (association task) [12, 18] becomes available (Fig. 3a), the meaning of which is to choose from two offered figures the one that, in the participant’s opinion, is more consistent with the one given. A perfect choice is indicated by pressing the Right or Left button. An example of cursor movement during a simple choice task is Fig. 3b. The whole test is a cycle of an affective image, a self-assessment survey and three association tasks. In total each participant completed 10 cycles, respectively viewing 10 images, completing 10 self-assessments, and solving 30 simple choice tasks. Thanks also to time recording, it can be said that, on average, it took the participants no more than 7 min to solve the complete test.

Fig. 3. a) Example of an association task to record the location of a subject’s mouse cursor. b) An example of a solution to an association problem. The red line is a hypothetical trajectory and is not shown in the real experiment.

While the participants solve the choice tasks, each new cursor location is recorded in the background, namely the x and y coordinates. The scheme of interaction between the parties, client, and server, is shown in Fig. 4. Each time a participant moves the computer mouse, the client interface sends data such as the identification number (ID) and spatial characteristics of the cursor to the server through a POST request via the web interface. In response, the server sends a request processing report message. In this way, the information about the movement of the cursor on each task page is sent to the server, where it is placed in the database and waits for further processing. The next stage after the accumulation of information from the participants is its systematization and transformation into a unified data set. According to [12, 15, 18, 19], attention ceases to be directed and becomes stimulated when a person is affected

Development of a Mechanism for Recognizing the Emotional State

87

Fig. 4. Interaction between client and data collection unit server in a mechanism for analyzing emotional state based on unconscious subject movements.

by emotion, with the subject becoming more distracted by environmental stimuli. When viewed from the perspective of neuroscience, the brain’s capacity for concentration is also affected by emotions, particularly negative ones. Therefore, temporal, spatio-temporal and two spatial characteristics are chosen as metrics to assess emotional state from mouse cursor movements. Temporal is the time of user’s interaction with the page of the simplest choice (association) task. The spatio-temporal characteristic is the maximum speed at which the computer mouse moves. The spatial characteristics are the distance travelled by the cursor and the value of the maximum deviation from the straight line connecting the start and end points of the movement. A program developed in Python is used to create a single dataset. The distance travelled is found by Eq. (1) as the sum of the Euclidean distance between two points, according to Eq. (2). D=  d (ai , ai+1 ) =

n−1 i=1

d (ai , ai+1 )

    (y) (y) 2 (x) (x) 2 ai − ai+1 + ai − ai+1

(1)

(2)

Peak mouse speed is calculated as the difference of the Euclidean distance between points and the time it is travelled. The maximum deviation of the cursor path from the straight line connecting the start and end points is calculated as the greatest minimum distance (perpendicular) from the point to the straight line. After finding the four characteristics for each completed association task (assuming that the participant indicated a choice by pressing the appropriate button), all data are placed in one file, each line of which carries: id is participant number; gender is value 1 for male, 2 for female; emotion is self-rated emotional state (1-positive, 2-negative, 3-neutral); Euclidean distance - distance travelled by cursor (pixels); max straight line distance is maximum deviation of the cursor from the straight line connecting the start and end points; peak velocity is peak cursor speed (pixel/nanosecond); time is interaction time with the simplest selection task (nanoseconds). The final file contains 3139 lines of information.

88

Y. Gorbunova and G. Kiselev

3.3 Research and Data Processing The study of temporal, spatio-temporal and spatial characteristics is the key to the analysis of the emotional state based on the unconscious actions of the subject. The tables below present the statistical values, maximum, minimum and median, respectively, for the four characteristics investigated. The values are specified for males, females, and the whole group of participants. This will reveal the correlations, if any, between the values and the gender of the participants. From the analysis of the maximum values of the characteristics (Table 1), the maximum values for men are observed in the maximum deviation of the movement trajectory from the straight line connecting the start and end points (856.21 versus 823.48 for women), peak speed (816 × 10–6 vs. 612 × 10–6 for women), and interaction time with the simplest choice task (575.040 × 108 vs. 382.679 × 108 for women). The maximum value of distance travelled is observed for women (14620.5 vs. 10955.7 for men). Table 1. Maximum values for spatial and temporal characteristics of the mouse cursor. Characteristic

Maximum values Men

Women

Shared

Distance travelled (pixels)

10955.7

14620.5

14620.5

Maximum deviation (pixels)

856.21

823.48

856.21

Peak speed (pixels/ns)

816 × 10–6

612 × 10–6

816 × 10–6

Time (ns)

575.040 × 108

382.679 × 108

575.040 × 108

Analysis of the minimum values (Table 2) shows that the lowest values for women are observed in the maximum deviation of the trajectory from a straight line (5.87 vs. 7.26 for men), time to solve the simplest choice problem (3.671 × 108 vs. 4.858 × 108 for men). The minimum value of the traveled distance is observed for men (477.9 vs. 532.3 for women). The peak speed is independent of the participant’s gender and amounted to 1 × 10–6 . Table 2. Minimum values for spatial and temporal characteristics of the mouse cursor. Characteristic

Minimum values Men

Women

Shared

Distance travelled (pixels)

477.9

532.3

477.9

Maximum deviation (pixels)

7.26

5.87

5.87

Peak speed (pixels/ns)

1 × 10–6

1 × 10–6

1 × 10–6

Time (ns)

4.858 × 108

3.671 × 108

3.671 × 108

The median peak speed values (Table 3) for males and females are closest compared to the rest of the characteristics - 5.02 × 10–6 and 5.73 × 10–6 respectively (5.28 × 10–6

Development of a Mechanism for Recognizing the Emotional State

89

for the full group of participants). The same can be said for the association task solution time - 18.19 × 108 for men and 18.20 × 108 for women (18.19 × 108 for the full group of participants). Table 3. Median values for spatial and temporal characteristics of the mouse cursor. Characteristic

Median values Men

Women

Shared

Distance travelled (pixels)

796.5

698.8

761.3

Maximum deviation (pixels)

134.05

Peak speed (pixels/ns)

5.02 × 10–6

Time (ns)

89.17

115.32

5.73 × 10–6

18.19 × 108

18.20 × 108

5.28 × 10–6 18.19 × 108

From the analysis of the statistical characteristics, it can be concluded that the values are not strongly dependent on the gender of the participants, the observed difference in performance is not significant. Table 4 presents the lowest and highest values of time, peak speed, spatial characteristics for the three subject states: positive, negative, neutral. Table 4. Maximum and minimum values of characteristics for different emotional states of the subject. Characteristic

Emotional state Positive

Negative

Neutral

Max

Min

Max

Min

Max

Min

Distance travelled (pixels)

6838.9

538.6

10467.6

532.29

14620.5

477.9

Max. Deviation (pixels)

755.35

6.48

856.21

5.87

668.91

7.21

Peak speed (×10–6 pixels/ns)

816

1

505

1

420

1

The distance travelled by the cursor reaches a maximum and a minimum when the condition is neutral. The situation is similar for the maximum deviation of the trajectory of movement from a straight line only for negative emotions. Peak speed has the highest value for a positive emotional state; the minimum is independent of the mood of the participant. The longest and shortest time to solve the simplest choice (association) task corresponds to positive emotions. The analysis shows that emotions affect subjects differently, and one cannot assess emotional state based only on the statistical characteristics of cursor movements. Two neural networks using the Python programming language are trained to analyze emotions based on the unconscious movements of the subject, mouse movement: a multilayer perceptron using the TensorFlow library [20], and the OneVsRestClassifier from

90

Y. Gorbunova and G. Kiselev

the Scikit-learn library [21]. The task of emotion recognition based on subject actions is a multi-class classification [22] because there are three classes: positive, negative, and neutral emotional states. The multi-layer perceptron (MLP) is a class of simple directcoupled neural networks using supervised learning. The OneVsRestClassifier network is based on the principle of selecting a classifier for each class and then matching the class to others for each classifier, thus making the approach interpretive. The emotion analysis task solved using seven approaches to compare performance: in six cases, different pairs of features are taken as independent variables (two features out of four are chosen), and in the seventh case, all features are taken as independent variables. In each approach, the data set is split into a training sample and a test sample. For MLP neural network analysis of emotion, the dependent variables (emotional state value) transformed using One-Hot coding, where the result is a binary data array whose size equals the number of features, with one in the column corresponding to the feature number and zeros in the others. The neural network consists of a normalizing layer adapted for the corresponding independent variables; two hidden full-connected layers (Dense) with 64 neurons each and a ReLU activation function [23]; an output full-connected layer (Dense) with a Softmax activation function [24] and three neurons according to the number of classes. The model trained over one hundred epochs. To investigate the relationship between emotion and mouse movements using the OneVsRestClassifier classifier, no transformation of dependent variables is required. Reference vector classification is used in the construction of the model [25]. Table 5 summarizes the results of seven studies for each classifier. Table 5. Training accuracy of multi-class classification models for different independent variables. Independent variables for the dataset

Classifiers MLP

OneVsRestClassifier

Euclidean distance + max straight line distance

0.4121

0.2663

Euclidean distance + peak velocity

0.4116

0.3174

Euclidean distance + time

0.4151

0.2663

max straight line distance + peak velocity

0.4087

0.3173

max straight line distance + time

0.4166

0.3174

peak velocity + time

0.4161

0.3173

All four attributes

0.4161

0.2663

From the results obtained, we can conclude that a pair of features is the maximum deviation of the movement trajectory from the straight line connecting the start and end points and the time to solve the simplest choice (association) problem is the most informative for analyzing the emotional state based on the unconscious movements of the subject, as both classifiers showed the maximum accuracy index for the pair of features. The most inefficient pair to investigate with an MLP network is the maximum deviation

Development of a Mechanism for Recognizing the Emotional State

91

of the trajectory from a straight line and the peak speed, for the OneVsRestClassifier model these are three pairs: distance travelled and maximum deviation from a straight line, distance travelled and interaction time, a set of four features.

4 Conclusion The technologies of emotion detection based on the subject’s unconscious actions, computer mouse control, and interaction with a smartphone considered. The lack of freely available data for a study on this topic is revealed. A dataset containing temporal, spatial, and spatio-temporal characteristics of the cursor collected. These are the interaction time with the computer mouse during the session, the distance travelled by the cursor, its maximum deviation from the straight line connecting the start and end points of movement, and the peak speed of movement respectively. A web application in Python programming language using FastAPI web framework developed to compile the dataset. The final dataset contains 3139 lines of information. Such data helps to explore a subject’s emotional state regardless of their race, language, or other affiliations. The statistical values of the cursor characteristics, maximum, minimum and median, for men and women maximum and minimum values for each emotional state, positive, negative, neutral analyzed. From the analysis of the statistical characteristics, it is deduced the values do not depend on the gender of the participants. The MLP and OneVsRestClassifier from the TensorFlow and Scikit-learn libraries trained to identify the relationship between emotional state and cursor movement, respectively, with a 41% accuracy in determining emotional state. The results indicate there is a relationship between the emotions experienced and unconscious movements of the subject. Acknowledgments. This study was conducted within the framework of the scientific program of the National Center for Physics and Mathematics, section №9 “Artificial intelligence and big data in technical, industrial, natural and social systems”.

References 1. Sun, C., Huang, L., Qiu, X.: Utilizing BERT for Aspect-Based Sentiment Analysis via Constructing Auxiliary Sentence. CoRR, abs/1903.09588 (2019) 2. Caschera, M., Grifoni, P., Ferri, F.: Emotion classification from speech and text in videos using a multimodal approach. Multimodal Technol. Interact. 6(4), 28 (2022) 3. Yoon, S., Byun, S., Jung, K.: Multimodal speech emotion recognition using audio and text. In: 2018 IEEE Spoken Language Technology Workshop (SLT), pp. 112–118 (2018) 4. Zhao, J., et al.: M3ED: Multi-modal multi-scene multi-label emotional dialogue database, arXiv preprint arXiv:2205.10237 (2022) 5. Fei, Z., Yang, E., Yu, L., Li, X., Zhou, H., Zhou, W.: A novel deep neural network-based emotion analysis system for automatic detection of mild cognitive impairment in the elderly. Neurocomputing 468, 306–316 (2022) 6. Chand, H.V., Karthikeyan, J.: CNN based driver drowsiness detection system using emotion analysis. Intell. Autom. Soft Comput. 31(2), 717–728 (2022) 7. Perepelkina, O., Kazimirova, E., Konstantinova, M.: RAMAS: Russian Multimodal Corpus of Dyadic Interaction for studying emotion recognition. PeerJ Preprints 6, e26688v1 (2018)

92

Y. Gorbunova and G. Kiselev

8. Ackerley, R., Aimonetti, J.M., Ribot-Ciscar, E.: Emotions alter muscle proprioceptive coding of movements in humans. Sci. Rep. 7(1), 8465 (2017) 9. Troy, A.S., Shallcross, A.J., Brunner, A., Friedman, R., Jones, M.C.: Cognitive reappraisal and acceptance: effects on emotion, physiology, and perceived cognitive costs. Emotion 18(1), 58 (2018) 10. Goldin, P., Moodie, C., Gross, J.: Acceptance versus reappraisal: behavioral, autonomic, and neural effects. Cogn. Affect. Behav. Neurosci. 19, 927–944 (2019) 11. Smith, K.A., Morrison, S., Henderson, A.M., Erb, C.D.: Moving beyond response times with accessible measures of manual dynamics. Sci. Rep. 12(1), 19065 (2022) 12. Yamauchi, T., Xiao, K.: Reading emotion from mouse cursor motions: affective computing approach. Cogn. Sci. 8, 771–819 (2018) 13. Yang, L., Qin, S.: A review of emotion recognition methods from keystroke, mouse, and touchscreen dynamics. IEEE Access 9, 162197–162213 (2021) 14. Tag, B., Sarsenbayeva, Z., Cox, A.L., Wadley, G., Goncalves, J., Kostakos, V.: Emotion trajectories in smartphone use: towards recognizing emotion regulation in-the-wild. Int. J. Hum Comput Stud. 166, 102872 (2022) 15. Hibbeln, M., Jenkins, J.L., Schneider, C., Valacich, J.S., Weinmann, M.: How is your user feeling? Inferring emotion through human-computer interaction devices. MIS Q. 41(1), 1–22 (2017) 16. FastAPI Documentation. https://fastapi.tiangolo.com/. Accessed 28 May 2023 17. Kurdi, B., Lozano, S., Banaji, M.: Introducing the open affective standardized image set (OASIS). Behav. Res. Methods 49, 457–470 (2017) 18. Yamauchi, T., Leontyev, A., Razavi, M.: Assessing emotion by mouse-cursor tracking: theoretical and empirical rationales. In: 2019 8th International Conference on Affective Computing and Intelligent Interaction (ACII), pp. 89–95 (2019) 19. Ihbour, S., Anarghou, H., Boulhana, A., Najimi, M., Chigr, F.: Mental health among students with neurodevelopment disorders: case of dyslexic children and adolescents. Dement. Neuropsychologia 15(4), 533–540 (2021) 20. How to create an MLP classifier with TensorFlow 2 and Keras. https://github.com/christ ianversloot/machine-learning-articles/blob/main/how-to-create-a-basic-mlp-classifier-withthe-keras-sequential-api.md. Accessed 28 May 2023 21. One-vs-the-rest (OvR) multiclass strategy. https://scikit-learn.org/stable/modules/generated/ sklearn.multiclass.OneVsRestClassifier.html. Accessed 28 May 2023 22. Ghosal, D., Majumder, N., Poria, S., Chhaya, N., Gelbukh, A.: DialogueGCN: a graph convolutional neural network for emotion recognition in conversation. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 154–164. Association for Computational Linguistics (2019) 23. Nair, V., Hinton, G.: Rectified linear units improve restricted Boltzmann machines Vinod Nair. In: 27th International Conference on International Conference on Machine Learning, vol. 27, pp. 807–814 (2010) 24. Goodfellow, I., Bengio, Y., Courville, A.: Deep Learning. The MIT Press, Cambridge (2016) 25. Support Vector Machines. https://scikit-learn.org/stable/modules/svm.html. Accessed 28 May 2023

Development of a Device for Post-traumatic Ankle Rehabilitation Andrey Knyazev1(B)

, Sergey Jatsun1

, Andrey Fedorov1 , and Jamil Safarov2

1 Southwest State University, 94, 50 Let Oktyabrya st., Kursk 305040, Russia

[email protected] 2 Azerbaijan Technical University, 25, Hüseyn Cavid pr., Baku AZ1073, Azerbaijan

Abstract. This article discusses a device for active-passive mechanotherapy of the ankle joint. The device is based on a controllable mobile platform equipped with force-moment sensors, on which the patient’s foot is mounted by means of cuffs, and the platform rotation angles are controlled by linear motion sensors. The platform of the device is designed in such a way that the rotation axis of the platform always coincides with the centre of the ankle joint. For this purpose, a parallel kinematics mechanism is used, which is based on three linear electric drives. The control system of the device provides both active and passive movement of the platform. For realization of the control algorithm of the mobile platform movement, a mathematical model is developed, which allows establishing connections between angular motions of the mobile platform and linear drives of the parallel mechanism. Models of reaction forces of the platform support on the patient’s foot during operation of the device are also described. A functional control diagram of the device is presented, and the modes of operation of the device are described. Keywords: Ankle Mechanotherapy · Force Control Sensors · Mathematical model · Parallel Kinematics Mechanism · Functional Diagram

1 Introduction Among injuries of lower extremities, the most widespread are those of distal part of lower leg and ankle, which, according to literature data, make up from 12.0 to 20.0% of all fractures of locomotor apparatus. In 12–39.8% of cases unsatisfactory results of treatment are observed, and the long-term disability lasts from 4 to 8 months. One of the most common injuries sustained by humans is damage to the ankle joint in sports, domestic, industrial exercises, and as a result of car accidents. Falling from heights with a landing on the feet, including parachuting, also often results in injuries to the ankle joint (AJ). Statistically, more than half of all lower limb injuries and about 40% of joint injuries are ankle injuries. We also know that 54% of ankle fractures and dislocations occur at a young age, when the ability to work is important. After an injury, there is a long process of treatment and rehabilitation, and after surgery, rehabilitation can be complicated by prolonged stiffness of the joint. In many countries, work is underway to develop devices and devices that allow for the post-traumatic rehabilitation of the © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 93–102, 2023. https://doi.org/10.1007/978-3-031-43111-1_9

94

A. Knyazev et al.

individual using AJ passive mechanotherapy devices. This approach makes it possible to perform foot movements according to an individual rehabilitation program (IRP) set by the doctor. The FLEX-02, A3 Ankle CPM, Kinetec Breva ankle, and ARTROMOT SP3 are widespread [1–3]. At the same time, due to the influence of indefinitely variable parameters of the AJ muscular system, it is difficult to ensure the required accuracy of the patient’s foot movement along the trajectory set by the doctor, which reduces the effectiveness of the rehabilitation process. Therefore, the creation of such devices requires an in-depth study of the theory of human-device interaction, creation of man-machine interfaces, mathematical models, and control algorithms that provide the specified quality indicators. Thus, the issues of developing and creating robotic devices for post-traumatic ankle rehabilitation that provide a given movement of the foot under unpredictable changes in physiological parameters are relevant [4–6]. The aim of the study is to improve the effectiveness of the rehabilitation process with an active-passive mechanotherapy and rehabilitation device that provides a given precision through adaptive control of foot movement, taking into account the individual characteristics of the patient’s AJ.

2 Tasks of Ankle Rehabilitation The main objectives of ankle rehabilitation during the recovery period: – – – –

reverse the processes of muscle atrophy and destructive changes in the vessels; return mobility to the joint; prevent stagnation of fluids in the reconstructed limb; increase the motor activity of the joint.

In the early stages of rehabilitation, it is necessary to perform simple movements of the foot in small ranges at a slow pace. Movements such as plantar flexion and dorsiflexion as well as pronation and supination. The exercises are shown in Fig. 1. Once the rehabilitator has noticed an improvement in the ankle’s motor function, the range of motion can be gradually increased.

Fig. 1. Exercises for early rehabilitation.

Development of a Device for Post-Traumatic Ankle Rehabilitation

95

All therapeutic and rehabilitation measures are determined by the severity of the injury as well as the individual characteristics of the patient. After the cast is removed, it is important to put stress on the ankle joint gradually. The exercises are shown in Fig. 2.

Fig. 2. Exercises for the final stage of rehabilitation.

In the final stage of recovery, you should start gradually combining these two movements by making circular movements with the foot. Also, continue to perform the operations described in the first phase, but with larger rotation angles until the ankle is able to perform the movements within the usual ranges of motion. The normal ranges of motion of the foot are shown in Fig. 3 [7, 8].

Fig. 3. Range of motion of the healthy ankle.

3 The Circuit Diagram of the Active-Passive Mechanotherapy Device The circuit diagram of the active-passive mechanotherapy device for the ankle, shown in Fig. 4, is a parallel manipulator equipped with linear actuators [9].

96

A. Knyazev et al.

Fig. 4. Spatial kinematic diagram of the appliance.

Figure 4 shows the following designations: 1 is the linear actuator 1 (l1 ); 2 is the linear actuator 2 (l2 ); 3 is the linear actuator 3 (l3 ); 4 is the controlled mobile platform; 5 is the patient’s foot; 6 is the platform in initial position; 7 is the virtual joint. By changing the lengths of the actuators, it is possible to change the position of the controlled mobile platform (CMP) and consequently the patient’s foot. The structure of the device includes the patient’s foot, mobile platform, electric actuators, power frame sensors, control system. The electric actuators of the manipulator are connected to the body by means of joints A1 , B1 , C1 . Actuators are connected to the platform by means of joints A, B, C. The joints A, B, C, A1 , B1 are two-coordinate, while C1 is a singlecoordinate joint, capable of changing its position in the sagittal plane [10, 11]. In order for the device to work correctly, the center of rotation of the CMP “virtual joint” O1 , defined by the intersection of the joint axes (A, B, C), must be at the point O2 of the ankle joint center. During rehabilitation measures, point O2 remains stationary and all platform movements occur around this point. The radius vectors determining the position of points O1 and O2 must be equal r o1 = r o2 . At the same time, under realworld conditions, there is a tolerance between the vectors, determined by the formula: r = r o1 − r o2 . We will assume that the trajectories of the mobile platform points satisfying the  = ε ≤ 0.01, are valid trajectories. condition |r| r  O1

A parallel kinematics mechanism (PKM) [12, 13] is used to implement the “virtual hinge” of the platform and ensure the intersection of the mobile platform rotation axes in the center (tibia and talus contact zone) of the AJ, based on three linear actuators. The ankle joint is formed by the tibia and talus bones. The articular surfaces of the lower leg bones and their ankles are encompassed by the Talus block in a fork-like fashion. The ankle joint is block-shaped. In this joint, flexion (movement towards the plantar surface of the foot) and extension (movement towards the rear surface of the foot) are possible around the transverse axis that runs through the talus block. The subtalar joint is responsible for pronation (rotation of the foot with the sole turning outwards) and

Development of a Device for Post-Traumatic Ankle Rehabilitation

97

supination (rotation of the foot with the sole turning inwards) of the foot. The human foot moves in three planes, the rotation around the vertical axis is caused by the rotation of the lower leg, therefore when working with active-passive mechanotherapy for ankle joint device (APMAJ) you should take this fact into account and not to fix the lower leg, otherwise you can get new injuries.

4 Diagrams for Measuring Leg and Platform Interaction Forces The principle of operation of the force meter is based on monitoring the relative movement of the upper and lower module elements of the mobile platform at four points. The device has sensors that measure the force between the patient’s foot and the mobile platform at these points. In active mode the patient provides the foot movement and the APMAJ platform replicates this movement. Impedance control is used for this. The mobile platform is attached to the patient’s foot via a system of elastic connections, the drive platform is equipped with angle sensors and electrically controlled actuators and ensures a defined movement of the foot. When a relative movement of one platform with respect to the other occurs, a deformation of the elastic elements of the measuring device occurs, which is registered by the movement sensor [14–16]. During operation of the device, reaction forces from the platform arise between the patient’s foot and the platform. To account for these forces, it is necessary to introduce a model to describe the interactions between the foot and the platform [17]. The foot acts on the platform as a distributed load, so replace it with two concentrated forces. Assume that the foot is in contact with the platform by the heel and metatarsal bones, together they form a bony system, which is connected to the plantar side of the foot by muscular tissues. Figure 5 shows a diagram of the interaction of the foot and platform in the sagittal plane. The following symbols are included in the diagram: 1 is ankle joint; 2 is metatarsal bones; 3 is heel bone; 4 is tibia; 5 is platform; 6 is force-moment interaction sensors.

Fig. 5. Diagram of foot and platform interaction in the sagittal plane.

Let’s write down the equation of moments for the patient’s foot in the sagittal plane: Jbs ψ¨ = MR + M (G) − MC ,

(1)

98

A. Knyazev et al.

where: J bs is the moment of inertia of the bone system; M R is the moment created by the support reaction forces; M(G) is the moment created by the gravity of the bone system; M C is the resistance moment of leg muscles. Let’s represent the muscle tissues as a Kelvin-Feugt model, with viscous and elastic components arranged in parallel. Given this, for sagittal plane, we will find forces R1 , R2 by formulas:   (2) R1 = c1 (ψ − ψ1 ) + μ1 ψ˙ − ψ˙ 1 ,   R2 = c2 (ψ − ψ1 ) + μ2 ψ˙ − ψ˙ 1 ,

(3)

where c is the reduced muscle stiffness coefficient; μ is the reduced muscle viscosity coefficient; (ψ − ψ1 ) is the relative strain of muscle tissues; (ψ˙ − ψ˙ 1 ) is the relative strain rate of the muscle tissues.

5 Functional Diagram of the Control System The functional diagram of the APMAJ robotic device is shown in Fig. 6. The diagram consists of a hardware and software system, a physiological parameter monitoring unit, a rehabilitation program selection unit, a signal processing unit, a mobile platform, electric drives, amplifiers, a control unit, a sensor system for measuring linear and angular movements and reactions between the patient’s foot and the APMAJ mobile platform in the sagittal and frontal planes [18]. The choice of rehabilitation mode determines the program of movements of the patient’s foot, formalized in the form of parametric equations λ(t)∗ = (φ(t)∗ , θ (t)∗ , ψ(t)∗ )T . If the conditions are met λ(t)∗ < λ(t)∗0 , where λ(t)∗0 is the range of permissible rotation angles. The on-board computer (signal processing unit) solves the inverse kinematics task (IKT) and determines the length change laws of the linear actuators L¯ ∗ = (L∗1 , L∗2 , L∗3 )T . ∗ Further signals are proportional L = (L∗1 , L∗2 , L∗3 )T are fed to the appropriate comparison units (comparators), where they are compared with the real values L = (L1 , L2 , L3 )T . The presence of feedbacks makes it possible to determine the deviation of the real position of the platform from the set position in the form of a vector λ and force interaction vector P. The regulators receive an error value L¯ = (L1 , L2 , L3 )T for each controlled variable. These values are converted by the control algorithm into control voltages U = (U1 , U2 , U3 )T . The force-moment sensing system makes it possible to evaluate changes in the reaction value over time and detect the moment of spastic effects, muscle contractures and automatically change the laws of foot movement in order to eliminate the patient’s trauma and pain syndrome. In accordance with the ACS diagram, the actuators are controlled to provide the specified movement of the patient’s foot. The feedback channels are the data recorded by the swing angle sensors on the respective axes, as well as the force-momentum sensor values. The LQR-optimization strategy is used to optimize the settings of the controllers to give the required quality performance of the APMAJ control

Development of a Device for Post-Traumatic Ankle Rehabilitation

99

Fig. 6. Functional diagram of the APMAJ robotic appliance.

system. If an abnormal situation occurs (reaction force limits are exceeded), the ACS algorithm allows you to change the exercise pace, perform the complex movements required for safe operation of the device and stop if necessary, thus ensuring safe patient rehabilitation [19].

6 Description of the Unit’s Operating Modes The rehabilitation unit has several modes of operation. Before starting the rehabilitation procedure, the rehabilitation therapist must select the operating mode parameters, using a program on a personal computer. The operation of the device has been simulated in MATLAB/Simulink environment, information about the operation modes is presented below. Plantar - Dorsiflexion. This robotic mode allows a rotational movement of the patient’s foot in the sagittal plane. The range of rotation angle is 42° ≤ ψ ≤ 23°. The results of the simulation mode are shown in Fig. 7. Pronation - supination. This mode of operation allows a rotational movement of the patient’s foot in the frontal plane. The range of rotation angle is 40° ≤ ϕ ≤ 40°. The results of the simulation are shown in Fig. 8. Any rehabilitation mode should be set according to the recommendations of the rehabilitation therapist. In this mode, the person’s foot can perform both combined and circular movements of the foot. The mode chosen by the user can include changing direction, speed of movement and number of repetitions (Fig. 9). The effectiveness of rehabilitation interventions depends largely on factors such as gender, age, duration of disability, educational level, and occupational affiliation of the patient. The Barthel and Rankin scales record the condition without offering the patient understandable benchmarks for the dynamics of the rehabilitation process, do not allow for accessible formulation of rehabilitation goals and motivation of the patient. Thus, these scales are appropriate for a screening assessment of the overall rehabilitation process. RMI, HAI and the KX test are recognized as informative techniques that

100

A. Knyazev et al.

Fig. 7. Mode of operation - plantar – dorsiflexion.

Fig. 8. Mode of operation - Pronation – supination.

Fig. 9. User-configured operating mode.

Development of a Device for Post-Traumatic Ankle Rehabilitation

101

are sensitive to changes in the patient’s condition in relation to both muscle function and functioning. The clarity of objectives and graded assessment allows rehabilitation goals to be formulated in an accessible format, which increases patient motivation and satisfaction with treatment. It is important to correctly determine the predicted end result of rehabilitation in each individual case, as this is the basis for the subsequent comprehensive assessment of the effectiveness of rehabilitation. At the same time, objective and informative criteria for the effectiveness of medical rehabilitation in various forms of occupational diseases have not been sufficiently developed and put into practice today. Meanwhile, data from domestic and foreign sources show that the levels of rehabilitation indicators are closely interrelated with the indicators of disability, mortality, morbidity, organization and quality of medical care and are directly related to qualitative and quantitative criteria that characterize public health.

7 Conclusion On the basis of the research carried out in the article, the following scientific and practical results have been obtained: – Computational schemes have been given, on the basis of which mathematical models describing kinematic and dynamic interactions of the device and the patient’s leg have been constructed. – The functional diagram of the device reflecting the structure of automatic control system is given and described. – The operating modes of the device are described. A prototype device is under development. Acknowledgment. The work was supported by grant of Russian Science Foundation № 22–2100464 – “Development of models and control algorithms for biotechnical walking systems”.

References 1. Pechurin, A., Fedorov, A., Jatsun, A., Jatsun, S.: Mathematical modelling of human walking in a rehabilitation exoskeleton using video gait analysis method. In: Proceedings of Southwestern State University, vol. 25, no. 3, pp. 27–40 (2022) 2. Karlov, A., Postolny, A., Fedorov, A., Jatsun, S.: Modeling an exoskeleton with a hybrid linear gravity compensator. In: Proceedings of Southwestern State University, vol. 24, no. 3, pp. 66–78 (2020) 3. Dmitriev, V., Fedorov, A.: Analysis of industrial exoskeleton qualitative performance based on a set of criteria. Issues in the methodology of natural and technical sciences: contemporary context, pp. 131–135 (2019) 4. Antipov, V., Karlov, A., Fedorov, A.: Energy distribution in the human-exoskeleton system. Issues in the methodology of natural science and engineering: contemporary context, pp. 109– 112 (2019)

102

A. Knyazev et al.

5. Pechurin, A.S., Jatsun, S.F., Fedorov, A.V., Jatsun, A.S.: Studying the two-legged walking system with video capture methods. In: Chugo, D., Tokhi, M.O., Silva, M.F., Nakamura, T., Goher, K. (eds.) CLAWAR 2021. LNNS, vol. 324, pp. 3–12. Springer, Cham (2022). https:// doi.org/10.1007/978-3-030-86294-7_1 6. Jatsun, S., Yatsun, A., Fedorov, A., Saveleva, E.: Simulation of static walking in an exoskeleton. In: Ronzhin, A., Shishlakov, V. (eds.) Electromechanics and Robotics. SIST, vol. 232, pp. 49–60. Springer, Singapore (2022). https://doi.org/10.1007/978-981-16-2814-6_5 7. Knyazev, A., Jatsun, A., Fedorov, A.: Mathematical modeling of the biomechanical rehabilitation system of foot exoskeleton in frontal and sagittal planes. In: Ronzhin, A., Pshikhopov, V. (eds) Frontiers in Robotics and Electromechanics. Smart Innovation, Systems and Technologies, vol. 329, pp. 19–32. Springer, Cham (2023). https://doi.org/10.1007/978-981-197685-8_2 8. Knyazev, A., Yatsun, S., Fedorov, A.: Control of a Device for mechanotherapy of the ankle joint. Biomed. Eng. 56, 392–396 (2023) 9. Knyazev, A., Jatsun, S., Fedorov, A.: Algorithm of personalized adjustment of the activepassive mechanotherapy device for the ankle joint. In: International Conference on Industrial Engineering, Applications and Manufacturing, pp. 661–666 (2023) 10. Hassan, M., Khajepour, A.: Optimization of actuator forces in cable-based parallel manipulators using convex analysis. IEEE Trans. Rob. 24, 736–740 (2008) 11. Alvarez-Perez, G., Garcia-Murillo, A., Jesús Cervantes-Sánchez, J.: Robot-assisted ankle rehabilitation: a review, disability and rehabilitation. Assist Technol. 15(4), 3e94-408 (2020) 12. Antonellis, P., Galle, S., Clercq, D., Malcolm, P.: Altering gait variability with an ankle exoskeleton. PLoS One 13(10), 0205088 (2018) 13. Jamwal, P., Hussain, S., Xie, S.: Restage design analysis and multicriteria optimization of a parallel ankle rehabilitation robot using genetic algorithm. IEEE Trans. Autom. Sci. Eng. 12(4), 1433–1446 (2014) 14. Tsoi, H., Xie, Q.: Design and control of a parallel robot for ankle rehabiltation. In: 15th international conference on mechatronics and machine vision in practice, pp. 515–520 (2008) 15. Vallés, M., Cazalilla, J., Valera, Á.: A 3-PRS parallel manipulator for ankle rehabilitation: towards a low-cost robotic rehabilitation. Robotica 35, 1939–1957 (2017) 16. Zeng, X., Zhu, G., Zhang, M.: Reviewing clinical effectiveness of active training strategies of platform-based ankle rehabilitation robots. Healthcare Eng. 2018, 1–12 (2018) 17. Zhang, M., McDaid, A., Veale, A.: Adaptive robot with trajectory tracking control of a parallel ankle rehabilitation joint-space force distribution. IEEE Access 7, 812–820 (2019) 18. Shevko, D.: Adaptive management in the conditions of undefinition. Sci. Rev. Tech. Sci. 2, 75–77 (2017) 19. Yatsun, A., Karlov, A., Malchikov, A., Jatsun, S.: Investigation of the dynamical characteristics of the lower-limbs exoskeleton actuators. In: MATEC Web of Conferences. EDP Sciences, vol. 161, pp. 3–8 (2018)

Evaluation of EEG Data for Zonal Affiliation of Brain Waves by Leads in a Robot Control Task Daniyar Wolf , Yaroslav Turovsky , Anastasia Iskhakova(B) and Roman Meshcheryakov

,

V.A. Trapeznikov Institute of Control Sciences of Russian Academy of Sciences, 65, Profsoyuznaya Street, Moscow 117997, Russia [email protected]

Abstract. The task of creating a neural interface for controlling a robotic system by means of an oculographic interface and bioelectric signals, is considered. The article highlights the results of scientific experimental research aimed at the evaluation of the representativeness of bioelectrical signals obtained by electroencephalography (EEG). The basic hypothesis is formulated and tested with the help of artificial neural network technology. The authors consider an experiment on the formation of steady-state visually evoked potentials in a group of people with the subsequent creation of an applied database. They describe an original approach for extracting representative features from the EEG signal. With the help of deep machine learning technology the representativeness of the data under study is evaluated. The main conclusions are formulated and the hypothesis that each brain lead reproduces unique waves which are characteristic of each brain zone is confirmed. The proposed model of a symmetric multilayer multi-adaptive direct propagation neuron can find its application in solving problems related to the processing of EEG signals. Based on the results of this study, the authors suggest that data on the bioelectrical activity of the brain can be uniquely identified, and thus used as control signals for various robotic devices. Keywords: Robot Control · Brain-Computer Interface · Bioelectrical Signals · Electroencephalography · Steady-State Visual Evoked Potentials · Machine Learning

1 Introduction The development of modern interdisciplinary approaches at the junction of information technology and physiology has led in recent decades to the appearance of a large number of human-computer communication devices. Managing robotic systems poses new challenges, including the development and improvement of control methods. For example, controlling copter movement using bioelectric operator signals. In addition to the movement of robotic systems, such interfaces can be applied to a wide range of tasks related to the rehabilitation of patients with neurological and traumatological © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 103–115, 2023. https://doi.org/10.1007/978-3-031-43111-1_10

104

D. Wolf et al.

profiles, and improvement of control of various devices: from a personal computer to aviation systems. That is, in this case, the benefits to humans are multifaceted - both a new level of control, combining the benefits of the human brain, and helping humans themselves to analyze and interpret their brain signals in cases of rehabilitation. The most common tasks currently being solved are improving their hardware, developing new algorithms for processing received signals and transforming them into commands to effector devices, and searching for new physiological phenomena that can be used as the basis for commands transmitted through new interfaces. The relevance of developments in the field of systems that could help to organize interaction between a human being and external devices through brain signals or brain– computer interface (BCI) is confirmed by the fact that at the moment there is already a number of studies dedicated to this topic. One of the studies in 2020 was conducted by Dalin Yang, Trung-Hau Nguyen, and Wan-Young Chung of Busan National University in South Korea. Their work aimed to create their own neurocomputer interface. The researchers proposed a simplified synchronized hybrid system for multiple command control of electroencephalograph signals in the motor cortex. The proposed system can issue 38 control commands, for which the user only needs to focus on the stimulus and blink his eyes [1]. Another study was conducted by researchers at the National Chin-Yi University of Technology in Taiwan. In their research paper, they draw attention to the significance of BCI for people suffering from motor neuron disease (MND) who are unable to move independently. The paper mainly proposed a brain-computer interface (BCI) based on a wireless electroencephalogram and DC motor drive circuit to control electric wheelchairs via Bluetooth interface for paralyzed patients [2]. S. Rihana, P. Damien, T. Moujaess aimed their work at detecting eye blink signals from electroencephalography (EEG) signals. The researchers collected data, described methods used for pre-processing of EEG signals and also for the classification of eye blink signals using a probabilistic neural network as a binary classifier. Their ultimate goal was to apply the resulting database to a neurorehabilitation application for patients with movement disorders [3]. One of the ambitious projects in this direction is the Neuralink project by Elon Musk, as a result of which a neuro-technology capable of connecting the human brain with a computer is to be created. As of today, Neuralink presents a miniature chip with a diameter of about 23 mm and a height of 8 mm. The chip can have a maximum of 1024 connections to the brain, flexible filaments for invasive connection of chips to the brain, efficient algorithms for signal processing, and immediately a robotic surgeon to perform invasive procedures of filament implantation [4]. Despite significant scientific advances in technological development, the human brain remains the most powerful, unique and fastest computing system. In this regard, developments and research allowing a human being to interact directly with technology excluding any physiological activity (as BCI) are becoming more and more widespread and relevant. Nevertheless, in the above-mentioned works, the topics of obtaining and preparing statistical information, as well as their theoretical evaluation, are poorly covered. Undoubtedly, the use of hybrid technologies (elements and technologies of “artificial

Evaluation of EEG Data for Zonal Affiliation of Brain Waves

105

intelligence”) to obtain scientific and engineering results is currently the current world practice. The main hypothesis of this study is that, supposedly, during successive cycles of formation of steady-state visually evoked potentials (SSVEP), the brain reproduces unique waves consisting in changes of frequency-time characteristics of registered signals, typical for each head zone (leads). In other words, presumably, it is possible to identify with a certain degree of probability a particular EEG signal belongs to which brain lead. However, before making this kind of estimation, it is necessary to perform some transformations of the collected EEG data. Thus, this study is aimed at obtaining informative features from the EEG signals with their subsequent machine classification. The authors in the article [5] evaluated for brain fatigue during sessions of the SSVEP. The statistical conclusions that were obtained as a result of this experiment served as a reason to continue research in this direction. Thus, it is necessary to conduct a study to confirm the hypothesis that the waves recorded by the electroencephalograph sensors from the head leads are subject to grouping and can be applied in machine learning tasks, and subsequent machine classification in various scientific and engineering tasks. And, as a consequence, the results of this study will advance research into the identification of EEG signals of the human brain. The task of controlling robots without using physical influence on the device – with the use of human brain signals (as if thoughts) can be solved using the data from this study.

2 Collecting and Preparing Statistical Data EEG-type bio-signals are difficult to interpret and use directly in the control task. In order to formalize and further standardize in the solution of this problem, an experiment was conducted. The scheme of interaction looks quite ordinary (Fig. 1). But an important component is a block BCI, which is the intellectual part of the processing of complex formalized signal of the person to the formal signal to the device. To improve the efficiency and quality of interpretation of human control signals, it is necessary to study the signals, to try to identify them, to select the best attributes for their recognition. By doing so, the BCI unit will become progressively more effective.

Fig. 1. Scheme of human-robot interaction via BCI.

The aim was to test the previously presented hypothesis, whether the signal could be claimed to have been caused by one or another zone of the brain.

106

D. Wolf et al.

For the study, we created a database with EEG SSVEPs obtained from SSVEP sessions from 30 subjects of both sexes aged from 17 to 23 (12 women and 18 men) with no neurological or psychiatric pathologies. Before the experiment, the participants did not take any psychotropic drugs and had normal or corrected to normal vision. The electroencephalogram data were recorded using Neuron-Spectr-4VP device (Neurosoft, Russia) at O1, O2, Oz, P3, P4 and Pz leads with a sampling frequency of 5000 Hz, with the cutoff filter on and with the high and low frequencies filter off. Photostimulation was performed at frequencies 1, 8, and 14 Hz. The duration of each SSVEP session was 15 s. Based on the available input dataset, the dynamics of frequency-time characteristics of SSVEP was determined. When the retina is excited by flashes of frequency ranging from 3.5 to 75 Hz, the brain generates electrical activity with the frequency of flashes [6–10]. After data accumulation, the electroencephalograms were filtered by a Sixth Order Butterworth Low Pass Filter with 2 (low cut) to 35 (high cut) Hz bandwidth (Fig. 2).

Fig. 2. The SSVEP signals from the “P4” lead: blue graph – before filtering; green graph – after filtering (Color figure online).

Next, the signals were grouped according to the photostimulation frequencies (1, 8, and 14 Hz) and the groups of leads “O1”, “O2”, “Oz”, “P3”, “P4”, and “Pz” for each subject. Given the sampling rate, three data matrices were obtained after grouping: N = 30 × M = 75000 × V = 6 elements, where: N is the ordinal number of the subject; M is the number of samples in the signal;

Evaluation of EEG Data for Zonal Affiliation of Brain Waves

107

V is the number of signals in the group for each subject (also let this value denote the set consisting of {“O1”, “O2”, “Oz”, “P3”, “P4”, “Pz”}). If only the first and the last 25000 signs are selected, two matrices could be obtained containing information about the signals for the first and the last 5 s. However, this approach of data preparation for both machine learning and machine classification produces a very large vector of informative features – M, and a very small vector with the samples for each signal (examples) – N. At a minimum, the high dimensionality of the features is associated with very high computational costs, and to a greater extent, a large number of trainable parameters can lead to a complex classification model, which will have a high variance and will not allow to qualitatively identify the desired features. It is not difficult to see that there are sufficiently large sets on M for set V. To reduce the dimensionality of V, we performed resampling of each signal in groups up to M = 25,000 while maintaining the data structure. Data augmentation was accomplished by “splitting” the feature M into windows of 1260 counts in increments of 625 counts, giving a total of 1140 samples for each lead. As a result of the above work, applied data were obtained which is a structured record of the electrical activity of the human brain as a result of the experiment with SSVEP (Fig. 3).

Fig. 3. The structure of the EEG input data with SSVEP for each subject in the frequency groups of 8 and 14 Hz. Similarly for 1 Hz.

A number of studies [11–14] noted the individual psychophysical and psychoemotional features of each subject – there is no guarantee that the signs are unambiguously separable in time for the entire group of subjects. To solve this problem, we applied deep machine learning for the further machine classification based on classical methods [15].

3 Obtaining Representative Features To integrate the elements of the artificial neural network (ANN), a specialized computer program was developed [16]. Taking into account the obtained data structure, a special autoencoder for EEG coding was developed.

108

D. Wolf et al.

An autoencoder (AE) is a kind of artificial neural network the purpose of which is to restore input information (data) at the output. AE performs only two tasks: 1) compression of the input data into a representation of the hidden space (or latent vector), also known as the information bottleneck; 2) recovery of the output data based on the resulting representation. The autoencoding process itself is an unsupervised machine learning algorithm since the feature extraction algorithm is determined by deep machine learning. The concept of information bottleneck (IB) was introduced by Tishby et al. [17]. This notion was introduced together with the hypothesis that IB can extract relevant information by compressing the amount of information that passes through the entire neural network using previously learned input data compression. Traditionally, AE has been used in dimensionality reduction tasks to study informative features. And, only relatively recently, AE models with hidden parameters which use the concept of prior and posterior distribution, such as variational AE, began to be used to build generative models that can generate new data. This is accomplished by compressing information in an information bottleneck such that only important attributes are extracted from the entire data set, and these extracted attributes (representations) can be used to generate new data. From a formal point of view, the mathematical model of AE can be expressed as follows: φ : χ →F ψ :F→ χ

(1)

φ, ψ = argminχ − (φ ◦ ψ)χ2 , φ,ψ

coder function; χ is the source data; F is the latent space; ψ is the decoder function. Essentially, the neural network is divided into two segments: encoder and decoder. The encoder function (1), denoted as φ, maps the original data χ into the hidden space F which is present in IB. The decoder function, denoted as ψ, maps the hidden space F into the IB at the output. The output in this case coincides with the input function. Thus, the original image is reconstructed after some generalized nonlinear compression. The coding network can be represented by a standard neural network function gated through the activation function, where z is the latent dimensionality: z = σ (Wx + b).

(2)

Similarly, the decoding network can be represented in the same way but with different weight and offset – b, and with potentially used activation functions: x = σ  (W  x + b ). Using (2) and (3), the loss function will take the form:     x, x = x − x 2 = x − σ  W  (σ (Wx + b)) + b 2 .

(3)

(4)

Evaluation of EEG Data for Zonal Affiliation of Brain Waves

109

The loss function (4) is used to train the neural network using the standard backpropagation procedure. Since the input and output are the same data, and as already indicated above, the AE machine learning algorithm is unsupervised, the AE deep machine learning process is self-supervised learning. The goal of self-supervised learning for AE is to select encoder and decoder functions such that the IB dimensionality is minimal to encode the input signal but sufficient to recover it at the output. If the number of perceptrons in the IB layer is small enough, the ability to reconstruct the input information on the output will be limited and very different from the original (high losses). If too many perceptrons are used, it makes no sense to use compression at all. AE can start the task of reconstructing the original signal (a negative effect in the current study) without extracting useful information about the data distribution, if 1. the dimensionality of the hidden representation is the same as the dimensionality of the input; 2. the dimensionality of the hidden representation is greater than the dimensionality of the input; 3. a too large amount of data is provided for AE. In these cases, even a linear encoder and a linear decoder simply copy the input data to the output without learning anything useful about the distribution. Summarizing the theoretical aspects outlined above, we can conclude that in an AE network functionally the left part is the encoding part and the right part is the decoding part. Ideally, any AE architecture can be organized by setting the IB dimensions based on the complexity of the simulated distribution. Therefore, the goal of deep machine learning of the AE is not to copy the input data at the output of the neural network but to train AE to reconstruct the input data so that the informative bottleneck highlights (as if it recognizes) only useful information. Today, the application of AE is actively popularized in the tasks of image reconstruction and basic coloring, data compression, conversion of grayscale images into color images, creation of images with higher resolution, etc. The problem, however, is that the various AE models work only with data that is similar to the data on which they have been trained – applied data. The final implementation of the model of the obtained AE – “encoder-decoder” – is a symmetric multilayer multi-adaptive forward propagation neuron [18–20]. The encoder can map attributes with a dimension of 1260 elements into attributes with a dimension of 24 elements. However, this dimensionality was not obtained at once. The relevance of the received features was also estimated by means of a deep neural network in the decoder part which allows one to reconstruct the original signal (Fig. 4). Figure 4 shows the final AE network consisting of 9 layers, of which 7 layers with perceptrons are hidden (hidden1 – hidden7). The layers are reduced to the 4th hidden layer (5th layer for the whole network) and also increase towards the output layer, forming symmetric neural network architecture. Except for the 5th output layer IB, each output of the inner layers is provided by the activation function – Relu. It is worth noting that if a linear activation function is applied in each layer, then the hidden parameters present in IB correspond directly to the main components from PCA (principal components approach).

110

D. Wolf et al.

Fig. 4. Graph of the symmetric multilayer multi-adaptive forward propagation neuron providing functions of the encoder and the decoder.

AE training was reproduced until the AE was incomplete. Based on the results of each cycle of deep learning, the representation of the hidden space was evaluated by learning the informative features in the direction of decreasing IB size. As a result of each deep learning cycle of the network, a periodic test data run of N = 582 × M = 1260 × V = 6 was performed. The coding results were then subjected to machine classification. The input dimensionality of the machine classification data was as follows: N = 582 × M = p × V = 6, where p is the empirically chosen unknown value [21].

4 Evaluation of the Representativeness of the EEG Data Before performing machine classification the neural network was evaluated for the presence of significant features in the resulting data set. A random forest – an ensemble technique – was used to evaluate significant features [20–23]. By applying the random forest we were able to estimate the importance of the features as an average reduction in “contamination” [23, 24] calculated from all decision trees in the forest. This allowed us to make no assumptions about whether or not the data obtained were linearly separable. By training a forest of 500 trees on the resulting dataset, 24 features were obtained in order of their relative importance (Fig. 5). Histograms A, B in Fig. 5 show the distributions of features according to their importance. Histogram A shows the significance of features x7 and x16 for stimulation at 8 Hz, and histogram B shows x12 and x13 as significant features. Certainly, the features x22, x9, x6, x14, x0, and x18 for histogram A are also significant. The same can be said for features x11, x0, x7, x14, x15, x18, x16, as well as x23 of histogram B. However, machine evaluations were made based on the first two significant ones.

Evaluation of EEG Data for Zonal Affiliation of Brain Waves

111

Fig. 5. Histograms of the relative importance of the features in the dataset for the leads: A) O2; B) P4.

The k-nearest neighbor algorithm [25] was used for machine classification. For the tests, 873 non-training samples were used (cross-validation of 20% of the training sample). The results of the machine classification for the test sample are shown in Tables 1 and 2. Table 1. Results of the machine classification of encoded EEGs with SSVEP using the k-nearest neighbor algorithm. The stimulation frequency was 8 Hz (accuracy of 88%). Leads

Precision

Recall

F1-score

Support

O1

0.88

0.97

0.92

873

O2

0.90

0.57

0.70

873

Oz

0.92

0.84

0.88

873

P3

0.75

0.94

0.83

873

P4

0.98

0.99

0.99

873

Pz

0.91

0.97

0.94

873

Tables 1 and 2 contain estimates of the classifier’s ability to distinguish encoded EEG signals from one another (distinguish zonal leads). This is indicated quantitatively by the evaluation precision. The evaluation recall demonstrates the quality of the classifier algorithm to detect a series of corresponding coded EEG with SSVEP in general. The characteristic f1-score confirms the necessary medium-harmonic balance between the precision and recall evaluations, especially since the classes are almost balanced. The parameter support indicates the number of test samples. The reliability of the obtained results of machine classification can be estimated by the Confusion matrices shown in Fig. 6. In Fig. 7 we can observe the borders and clusters within them which correspond to the six leads: O1, O2, Oz, P3, P4 i Pz. We can also observe that the hidden space contains “gaps”; this is equivalent to missing data in the teacher-trained task because AE was not

112

D. Wolf et al.

Table 2. Results of machine classification of encoded EEG with SSVEP by k-nearest neighbor algorithm. The stimulation frequency was 14 Hz (accuracy of 80%). Recall

F1-score

Support

0.82

0.85

0.84

873

O2

0.72

0.68

0.70

873

Oz

0.82

0.82

0.82

873

P3

0.80

0.86

0.83

873

P4

0.97

0.92

0.94

873

Pz

0.68

0.67

0.67

873

True labels

Precision

O1

O1 O2 Oz P3 P2 Pz

True labels

Leads

O1 O2 Oz P3 P2 Pz

8 Hz stimulation 0 25 0 497 31 225 4 737 49 38 12 822 3 0 0 10 0 5 O2 Oz P3 Predicted labels 14 Hz stimulation 744 2 125 0 0 596 1 72 151 0 713 0 0 28 0 754 12 17 34 7 0 180 0 111 O1 O2 Oz P3 Predicted labels 848 28 82 0 0 3 O1

0 5 1 0 870 9 P2

0 87 0 1 0 846 Pz

2 16 9 1 803 0 P2

0 188 0 90 0 582 Pz

Fig. 6. Confusion matrices on leads O1, O2, Oz, P3, P4 and Pz.

trained on these features of the hidden space. Another problem is the separability of the gaps; some leads are well separated but there are also areas where there are “embeddings” of some features into others, which makes it difficult to separate the unique features of these leads. Nevertheless, we can conclude that the obtained features have a good enough separability, and therefore can be applied in further machine learning for the task of EEG signals classification. Based on the results of this study we can conclude that each brain lead reproduces unique waves specific to each area of the head. Moreover, the results of the experimental study show that there is a certain generalization of signals by their leads in a part of samples from the general population of statistical data from the subjects.

Evaluation of EEG Data for Zonal Affiliation of Brain Waves

113

Fig. 7. Borders of the solutions for leads O1, O2, Oz, P3, P4 and Pz obtained on the basis of the signs under study: A) for significant signs x7 and x16 (stimulation with 8 Hz); B) for significant signs x12 and x13 (stimulation with 14 Hz).

5 Conclusion Statistical information was collected to confirm the main hypothesis. Obtaining statistical data was achieved by creating technical conditions and organizational arrangements for a scientific experiment with a group of 30 subjects of both sexes. The experiment was provided with specialized medical measuring equipment of high-resolution – electroencephalograph. The achievability of the set task and the obtaining of 85% of the expected results were based on the preliminary estimation of relevance and representativeness of the received data as a result of an experiment on the gathering of bioelectrical activity of a brain, where the frequency-time response was taken as the sets of primary features. Thus, the representativeness of the experimental data was evaluated before proceeding to deep machine learning and the machine classification task. The proposed model of symmetric multilayer multi-adaptive forward propagation neuron which provides encoder and decoder functions can find its application in solving problems related to EEG signal processing. Based on the results of this study, the authors suggest that the data on the bioelectrical activity of the brain recorded by EEG can be represented as a multidimensional random variable, where the center position is also the mathematical expectation of its projections on the axes of the principal components. Acknowledgements. The study was financially supported by the Russian Science Foundation under scientific project No. 23-19-00664.

References 1. Yang, D., Nguyen, T.H., Chung, W.Y.: A bipolar-channel hybrid brain-computer interface system for home automation control utilizing steady-state visually evoked potential and eyeblink signals. Sensors (Basel) 20(19), 5474 (2020). https://doi.org/10.3390/s20195474.

114

D. Wolf et al.

2. Lin, J.-S., Yang, W.-C.: Wireless brain-computer interface for electric wheelchairs with EEG and eye-blinking signals. Int. J. Innovative Comput., Inf. Control 8, 6011–6024 (2012) 3. Rihana, S., Damien, P., Moujaess, T.: EEG-Eye blink detection system for brain computer interface. In: Pons, J.L., Torricelli, D., Pajaro, M. (eds.) Converging Clinical and Engineering Research on Neurorehabilitation, pp. 603–608. Springer, Heidelberg (2013). https://doi.org/ 10.1007/978-3-642-34546-3_98 4. Musk, E.: An integrated brain-machine interface platform with thousands of channels. BioRxiv preprint, https://www.biorxiv.org/content/10.1101/703801v4. Last accessed 31 May 2023. https://doi.org/10.1101/703801 5. Turovsky, Y., Wolf, D., Meshcheryakov, R., Iskhakova, A.: Dynamics of frequency characteristics of visually evoked potentials of electroencephalography during the work with brain-computer interfaces. In: Mahadeva Prasanna, S.R., Alexey Karpov, K., Samudravijaya, S.S., Agrawal, (eds.) Speech and Computer: 24th International Conference, SPECOM 2022, Gurugram, India, November 14–16, 2022, Proceedings, pp. 676–687. Springer International Publishing, Cham (2022). https://doi.org/10.1007/978-3-031-20980-2_57 6. Tao, T., Yi, X., Xiaorong, G., Shangkai, G.: Chirp-modulated visual evoked potential as a generalization of steady state visual evoked potential. J. Neural Eng. 9(1), 016008 (2011). https://doi.org/10.1088/1741-2560/9/1/016008 7. Kwak, N.-S., Müller, K.-R., Lee, S.-W.: Toward exoskeleton control based on steady state visual evoked potentials. In: 2014 International Winter Workshop on Brain-Computer Interface (BCI 2014), pp. 1–2. Gangwon, Korea (2014). https://doi.org/10.1109/iww-BCI.2014. 6782571 8. Balnyt˙e, R., Uloziene, I., Rastenyt˙e, D., Vaitkus, A., Malcien˙e, L., Lauˇckait˙e, K.: Diagnostic value of conventional visual evoked potentials applied to patients with multiple sclerosis. Medicina 47(5), 263–269 (2011) 9. Markand, Omkar N.: Visual evoked potentials. In: Clinical Evoked Potentials, pp. 83–137. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-36955-2_3 10. Chaudhary, U., Birbaumer, N., Curado, M.R.: Brain-machine interface (BMI) in paralysis. Ann. Phys. Rehabil. Med. 58(1), 9–13 (2015). https://doi.org/10.1016/j.rehab.2014.11.002 11. Aminoff, M., Goodin, D.: Visual evoked potentials. J. Clin. Neurophysiol.: Official Publ. Am. Electroencephalographic Soc. 11, 493–499 (1994). https://doi.org/10.1097/00004691199409000-00004 12. Taylor, M., McCulloch, D.: Visual evoked potentials in infants and children. J. Clin. Neurophysiol.: Official Publ. American Electroencephalographic Soc. 9, 357–372 (1992). https:// doi.org/10.1097/00004691-199207010-00004 13. Liasis, A.: Visual evoked potentials. Acta Ophthalmol. 94 (2016). https://doi.org/10.1111/j. 1755-3768.2016.0215 14. Carter, J.: Visual evoked potentials. Clinical Neurophysiology, 311–322 (2011). https://doi. org/10.1093/med/9780195385113.003.0022 15. Kwak, N.-S., Müller, K.-R., Lee, S.-W.: A convolutional neural network for steady state visual evoked potential classification under ambulatory environment. PLoS ONE 12(2), 1–20 (2017). https://doi.org/10.1371/journal.pone.0172578 16. Wolf, D.A., Turovsky, Y.A., Meshcheryakov, R.V., Iskhakov, A.Y., Iskhakova, A.O.: EEG signal auto encoder, computer software, https://www1.fips.ru/iiss/document.xhtml?faces-red irect=true&id=d4eb144baee4f995556af206cde9da36. Last accessed 31 May 2023. (In Russ.) 17. Naftali, T., Pereira, F.C., Bialek, W.: The information bottleneck method. In: Proceedings of the 37th Allerton Conference on Communication, Control and Computation, https://www.res earchgate.net/publication/2844514_The_Information_Bottleneck_Method. Last accessed 31 May 2023

Evaluation of EEG Data for Zonal Affiliation of Brain Waves

115

18. Nguyen, H., Bottone, S., Kim, K., Chiang, M., Poor, H.V.: Adversarial Neural Networks for Error Correcting Codes (preprint), https://www.researchgate.net/publication/357267696_ Adversarial_Neural_Networks_for_Error_Correcting_Codes. Last accessed 31 May 2023 19. Kose, U., Deperlioglu, O., Alzubi, J., Patrut, B.: Diagnosing parkinson by using deep autoencoder neural network. In: Deep Learning for Medical Decision Support Systems. SCI, vol. 909, pp. 73–93. Springer, Singapore (2021). https://doi.org/10.1007/978-981-15-6325-6_5 20. Mirjalili, V., Raschka, S., Namboodiri, A., Ross, A.: Semi-adversarial networks: convolutional autoencoders for imparting privacy to face images. In: 2018 International Conference on Biometrics (ICB), pp. 82–89. IEEE, Gold Coast, QLD, Australia (2018). https://doi.org/10. 1109/ICB2018.2018.00023 21. Meshcheryakov, R.V., Wolf, D.A., Turovsky, Y.A.: An autocoder of the electrical activity of the human brain. Bulletin of the South Ural State University, Series Mathematics. Mechanics. Physics 15(1), 34–42 (2023). https://doi.org/10.14529/mmph230104. (In Russ.) 22. Bicego, M., Escolano, F.: On learning random forests for random forest-clustering. In: 2020 25th International Conference on Pattern Recognition (ICPR), pp. 3451–3458. IEEE, Milan, Italy (2021). https://doi.org/10.1109/ICPR48806.2021.9412014 23. Olson, M.: Essays on Random Forest Ensembles, https://repository.upenn.edu/ dissertations/AAI10786136/. Last accessed 31 May 2023 24. Nayyar, A., Mahapatra, B.: Effective classification and handling of incoming data packets in mobile Ad Hoc networks (MANETs) using random forest ensemble technique (RF/ET). In: Sharma, N., Chakrabarti, A., Balas, V.E. (eds.) Data Management, Analytics and Innovation. AISC, vol. 1016, pp. 431–444. Springer, Singapore (2020). https://doi.org/10.1007/978-98113-9364-8_31 25. Fahim, A.: K and starting means for k-means algorithm. J. Comput. Sci. 55, 101445 (2021). https://doi.org/10.1016/j.jocs.2021.101445

Comparison of ROS Local Planners for a Holonomic Robot in Gazebo Simulator Artem Apurin1(B)

, Bulat Abbyasov1 , Edgar A. Martínez-García2 and Evgeni Magid1,3

,

1 Intelligent Robotics Department, Institute of Information Technology and Intelligent Systems,

Kazan Federal University, 35, Kremlyovskaya street, Kazan 420111, Russia [email protected] 2 Institute of Engineering and Technology, Department of Industrial Engineering and Manufacturing, Autonomous University of Ciudad Juarez, Manuel Díaz H. No. 518-B Zona Pronaf Condominio, Chihuahua, 32315 Cd Juárez, Mexico 3 Tikhonov Moscow Institute of Electronics and Mathematics, HSE University, 34, Tallinn street, Moscow 123458, Russia

Abstract. A safe robot navigation in a dynamic environment is an essential part of an autonomous exploration path planning. A path planning part of a navigation involves global and local planners. While a global planner finds an optimal path with a prior knowledge of an environment and static obstacles, a local planner recalculates the path to avoid dynamic obstacles. The main goal of a local planning is adjusting an initial plan produced by a global planner in an online fashion. It is a crucial step to ensure a robot operation in dynamic environments because in real world scenarios an environment usually contains people and thus, a dynamic obstacles avoidance must respond quickly and recalculate an actual route. Holonomic robotic platforms are robotic vehicles that use omni-wheels to move in any direction, at any angle, without an additional rotation. These robotic platforms are ideal for working zones with a limited space access. This paper provides a comparison of ROS local planners that support omni-wheel mobile robots: Trajectory Rollout, DWA, EBand, and TEB. The algorithms were compared using a path length, a travelling time and a number of obstacle collisions. Gazebo simulator was used for modeling virtual scenes with dynamic obstacles. Keywords: Mobile Robot · Mecanum Wheel · Local Planner · ROS · Gazebo

1 Introduction Nowadays, mobile robotics provides new opportunities for developing novel robotic systems. A wide range of wheels of various sizes, different design types and materials used allow to integrate mobile robotic platforms into many areas of a human life. Common mobile robot applications include industrial automation [1], transportation [2], medical care [3], emergency rescue operations [4], and other areas. Mobile robots are featured by different motion systems. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 116–126, 2023. https://doi.org/10.1007/978-3-031-43111-1_11

Comparison of ROS Local Planners

117

There are two different types of mobile robots drive systems: a holonomic and a nonholonomic. For a wheeled robot, a non-holonomic drive system is a robot configuration limited by a number of wheels or their orientation. Holonomic drive systems have more than two degrees of freedom, which provide more freedom and flexibility of motion. The main benefit of a holonomic drive system is an ability to travel in any desired direction at any specified orientation without additional rotations with regard to Z-axis (yaw) of a series of intermediate motions (e.g., a typical car parking procedure). To perform such locomotion a robotic platform uses a special design of wheels called mecanum or omnidirectional [5]. Omnidirectional wheels increase a robot mobility and are used in tasks where a high maneuverability is required. Omnidirectional robotic platforms are ideal for working zones with a limited space access and cluttered environments, e.g., for scheduling pick-up and delivery tasks in hospitals [6]. Performing safe robot navigation is a general issue faced by a robot operating in a real environment [7]. Real world environments usually contain people and other dynamic obstacles. A real-time path planning is an essential part of an autonomous exploration. An obstacle avoidance capability used by a path planning approach must detect obstacles quickly and replan an actual route [8]. A path planning part of a robot navigation involves global and local planners [9]. While a global planner finds an optimal path with a prior knowledge of an environment and static obstacles, a local planner recalculates the path to avoid dynamic obstacles. The main goal of a local planning approach is adjusting a plan produced by a global planner in an online fashion. This paper presents a comparison of ROS local planners supporting a holonomic drive system: Trajectory Rollout [10], DWA [11], EBand [12] and TEB [13]. These local planner algorithms were selected because they are most popular for ROS environment, easily pluggable and support a holonomic motion. The main contribution of the paper is a benchmark to discover the most suitable ROS local planner for a holonomic system used within a dynamic environment. Virtual experiments were conducted in Gazebo simulator [14] using static and dynamic obstacles.

2 System Setup 2.1 Mecanum Wheel Robot A virtual model of a modular multifunctional robotic omni-wheeled mobile platform ArtBul [15] was used for experiments. Mecanum wheel models were created in Blender software [16]. To reduce complex collision calculations and increase a real time factor (RTF), low-poly models were used for a roller collision part. A 3D model of the mecanum wheel is shown in Fig. 1. Each roller has its own joint and can be freely rotated along the Z-axis of its frame. A ROS plugin was developed to control the robot in Gazebo simulation [17] by publishing messages with linear velocities along the X and Y axes and an angular velocity along the Z-axis to a robot command topic. To detect collisions gazebo_ros_bumper plugin [18] was used. The mobile platform with a laser range finder (LRF) and an enabled bumper plugin is depicted in Fig. 2.

118

A. Apurin et al.

Fig. 1. 3D model of a mecanum wheel in Gazebo: a red, green and blue arrows’ set denotes a coordinate frame of each roller.

Fig. 2. ArtBul mobile robot in Gazebo: blue rays visualize LRF beams.

2.2 Virtual Environments Simulation provides a significant support in early stage testing. A 3D modeling can be used to produce a necessary 3D digital representation of real objects with a varying difficulty. Modern modeling tools are often used for designing virtual environments [19]. Testing local planners requires a special navigation map called an occupancy grid map (OGM). The OGM is a 2D binary map that consists of cells. The OGM encodes occupancy data where white pixels represent free cells, black pixels are occupied cells and gray pixels are not yet explored. In our test cases, OGMs should not contain any information about obstacles because a goal of a local planner is a real-time path planning processing.

Comparison of ROS Local Planners

119

Two different virtual worlds in Gazebo were created to benchmark ROS local planner algorithms. The first world had 20 × 6 × 3 m dimensions and contained static obstacles (Fig. 3, top). An OGM of the first virtual environment is shown in Fig. 3, bottom. The second world of 10 × 10 × 3 m dimensions contained a single dynamic obstacle – a cube with a 1 m length side (Fig. 4). We created two motion patterns that the cube uses while moving: along X-axis (Fig. 4, left top) and Y-axis (Fig. 4, left bottom). An OGM of the second virtual world is shown in Fig. 4, right.

Fig. 3. A 3D virtual environment filled with static obstacles: cubes of varying sizes, cuboids and cylinders (top). The corresponding 2D OGM with static obstacles excluded (bottom).

3 ROS Local Planners A motion control plays an important role in an autonomous navigation. ROS local planners use sensory information to perceive a current robot state and generates feasible trajectories that the robot is allowed to follow. Avoiding any dynamic or static obstacles that may (or may not) be included in a given global map is a responsibility of a ROS local planner. ROS local planners use sensory data from various sensors such as LRF sensors or ultrasound sensors, and various devices to plan an optimal trajectory [20]. All local planners use the same values for coinciding parameters and the same global and local costmap configurations. For all local planners limitations were set as follows: a linear acceleration was limited to 2.5 m/s2 , an angular acceleration to 3.2 rad/s2 , a linear speed to 0.5 m/s, and an angular speed to 1 rad/s. A controller tolerance in yaw/rotation (yaw_goal_tolerance) was set as 0.05 rad, a controller tolerance in the X and Y distance (xy_goal_tolerance) as 0.1 m.

120

A. Apurin et al.

Fig. 4. (Left) 3D virtual environment with a dynamic obstacle: a cube moves along X-axis (top) and Y-axis (bottom). (Right) The corresponding 2D OGM with static obstacles excluded.

Trajectory Rollout. ROS package base_local_planner provides implementations of the Dynamic Window and Trajectory Rollout approaches to a local control. Base_local_planner is a basic ROS local planner that provides an application programming interface for other local planners. In order to use Algorithm Trajectory Rollout, the parameter dwa should be set to false. For mecanum wheel robots holonomic_robot parameter should be set to true. DWA. ROS package dwa_local_planner is a modular DWA implementation with more flexible y-axis variables for holonomic robots than base_local_planner’s DWA. DWA discretely samples a robot’s control space, performs a forward simulation for each sample, evaluates and filters each trajectory in the local costmap and finally selects a highest-scoring trajectory. Eband. ROS package eband_local_planner implements the Elastic Band (EBand) method. An elastic band is a deformable collision free path generated by a global path incorporating information about obstacles proximity. A main drawback of eband_local_planner is that the ROS-based method implementation does not support an obstacle avoidance for moving obstacles. TEB. ROS package teb_local_planner implements the Timed-Elastic-Band (TEB) method for an online trajectory optimization. A difference between TEB and EBand is that a local trajectory is optimized not by external forces, but by applying a cost function. For a holonomic robot min_turning_radius parameter should be set to 0 and weight_kinematics_nh parameter to 1.

Comparison of ROS Local Planners

121

4 Performance Comparison Three experiments with different scenarios were conducted to identify a most suitable local planner for a mecanum wheeled robot. In the first experiment, a starting position of the robot was set to (0; −8) and a goal was set to (0; 8). A 2D occupancy grid map did not contain static obstacles. Next, several static obstacles were added to the world after a global map had been built. In the second experiment, the robot started at (−3.5; 0) and targeted to (3.5; 0). A cube with sides of 1 m moved linearly without an acceleration along a trajectory from point (2; 0) to point (−2; 0) and backwards, with 0.4 m/s linear velocity. In the third experiment, the cube moved from (0; 2) to (0; −2) and backwards. A reference trajectory depicted in Fig. 5 represents a suggested optimal path. Distance-optimal robot trajectories in the world with static objects are shown in Fig. 6.

Fig. 5. Distance optimal robot trajectory.

Experiments showed that the Trajectory local planner never generated linear velocities along the Y-axis, and the omnidirectional robot moved as a differential wheeled robot with any planner settings. The Eband local planner trajectory was the smoothest, but this planner cannot handle dynamic obstacles and does not perform an online trajectory replanning. The drawbacks of the Eband are that the algorithm uses a global costmap updated dynamically, does not publish a response after reaching a target point and a task execution time is unmeasurable. Table 1 demonstrates experimental results of local planners evaluated in the first world. Max T, Min T and Avg T stand for a maximum, minimum and average task execution time, respectively. Max D, Min D and Avg D denote a maximum, minimum and average path length, respectively. Success (Suc) column depicts how many times the robot reached the goal without obstacle collisions. Success with collision (SwC) column depicts how many times the robot reached the goal with at least one obstacle collision. Failed (F) column depicts how many times the robot failed to reach the goal. The TEB local planner showed the lowest minimum time and the lowest average time to complete the task. In one case, the robot with the TEB collided with an obstacle because the TEB heavily loaded the PC system and a frequency of publishing velocities to a command topic decreased. The DWA achieved the lowest minimum and the lowest average path length.

122

A. Apurin et al.

Fig. 6. Robot trajectories built by the local planners (in the world with static obstacles).

Table 1. ROS local planners’ evaluation results (in the world with static obstacles). Planner

Max T

Min T

Avg T

Max D

Min D

Avg D

Suc

SwC

F

Trajectory

188.6

64.8

100.9

23.2

17.9

19.6

50

0

0

DWA

67.1

39.6

43

18.9

17.2

17.7

50

0

0

EBand

-

-

-

19.2

18.1

18.9

50

0

0

TEB

91.6

32.4

38.3

43.4

17.9

19

49

1

0

In the experiment with the moving along the X-axis cube, all planners generated approximately the same trajectory (Fig. 7). Table 2 demonstrates experiments results of the local planners evaluated in the second world with the dynamic obstacle (a pattern of motion along X-axis). In this case, the TEB also showed the lowest minimum and the lowest average time required for a successful task completion. The Trajectory local planner showed the worst result within 50 experiments with only 7 successful and 15 (completely) failed. The EBand was successful in all 50 cases.

Comparison of ROS Local Planners

123

Table 2. ROS local planners evaluation results (the pattern of motion along X-axis). Planner

Max T

Min T

Avg T

Max D

Min D

Avg D

Trajectory

222.5

23.4

40.8

15.4

2.7

8.6

DWA

41.2

18.6

22.1

13.3

1.7

8

EBand

-

-

-

9

8.3

8.6

TEB

31.41

16.2

20.8

16.9

8.5

11

Suc

SwC

F

7

28

15

24

23

3

50

0

0

47

2

1

Fig. 7. Robot trajectories built by the local planners (the pattern of motion along X-axis).

In the case of the moving along the Y-axis cube, the DWA and the Trajectory Rollout showed a similar behavior (Fig. 8). When the cube appeared in the local costmap, the robot stopped and attempted to select an optimal movement trajectory. When the cube left the local costmap, the robot moved forward. The Eband and the TEB rebuilt the trajectory and continued the motion. Table 3 demonstrates experimental results of the local planners evaluated in the second world with a dynamic obstacle (a pattern of motion along Y-axis). The TEB generated a maximum robot velocity. This method demonstrated the minimum task

124

A. Apurin et al.

execution time and the least number of collisions with the obstacle. The worst result was shown by the DWA, which completely failed the task in 32 cases out of 50. Table 3. ROS local planners evaluation results (the pattern of motion along Y-axis). Planner

Max T

Min T

Avg T

Max D

Min D

Avg D

Suc

SwC

F

Trajectory

75

15.6

30.1

8.9

4.1

6.7

DWA

78

17.2

32.4

14.4

3.4

6.3

21

9

20

7

11

32

EBand

-

-

-

60

19.6

56.5

22

1

27

TEB

21.4

15

17.3

11

5.8

9

30

19

1

Fig. 8. Robot trajectories built by the local planners (the pattern of motion along Y-axis).

5 Conclusion This paper presented a comparison of ROS local planners for mecanum wheeled robots (Trajectory Rollout, DWA, EBand, and TEB) and provided a benchmark that allowed to experimentally determine a recommended ROS local planner for a wheeled holonomic

Comparison of ROS Local Planners

125

system. Virtual experiments were conducted in Gazebo simulator, which was used for modeling virtual environments with static and dynamic obstacles. Three types of virtual environments were employed with 50 virtual experiments within each environment. The algorithms were compared using a path length, a travelling time and a number of obstacle collisions. The virtual experiments showed that the Trajectory Rollout local planner did not generate linear velocities along the Y-axis in all configurations; this planner demonstrated the worst task execution time and the longest trajectory path lengths. The DWA local planner handled static obstacles effectively, but performed poorly with dynamic obstacles. The EBand local planner did not rebuild a local motion trajectory when it worked on a non-renewable costmap of an explored environment; therefore, the use of this planner is possible only when a global costmap is dynamically updated, in which case the robot’s trajectory will be rebuilt by a global planner. The TEB local planner achieved a significantly better performance in terms of a task execution time and showed the least number of obstacle collisions. Therefore, the TEB local planner could be recommended for dynamic scenes since it demonstrated the best results within a dynamic environment. Acknowledgements. The reported study was funded by the Russian Science Foundation (RSF) and the Cabinet of Ministers of the Republic of Tatarstan according to the research project No. 22-21-20033.

References 1. Koubaa, A., et al.: Introduction to mobile robot path planning. In: Robot Path Planning and Cooperation. SCI, vol. 772, pp. 3–12. Springer, Cham (2018). https://doi.org/10.1007/978-3319-77042-0_1 2. Gul, F., Rahiman, W., Nazli Alhady, S.S.: A comprehensive study for robot navigation techniques. Cogent Eng. 6(1), 1–25 (2019) 3. Hai, N.D.X., Nam, L.H.T., Thinh, N.T.: Remote healthcare for the elderly, patients by tele-presence robot. In: 2019 International Conference on System Science and Engineering (ICSSE), pp. 506–510 (2019) 4. Murphy, R.R.: Humans and robots in off-normal applications and emergencies. In: Chen, J. (ed.) Advances in Human Factors in Robots and Unmanned Systems: Proceedings of the AHFE 2019 International Conference on Human Factors in Robots and Unmanned Systems, July 24–28, 2019, Washington D.C., USA, pp. 171–180. Springer International Publishing, Cham (2020). https://doi.org/10.1007/978-3-030-20467-9_16 5. Taheri, H., Zhao, C.X.: Omnidirectional mobile robots, mechanisms and navigation approaches. Mech. Mach. Theor. 153, 103958 (2020) 6. Safin, R., Lavrenov, R., Hsia, K.H., Maslak, E., Schiefermeier-Mach, N., Magid, E.: Modelling a turtlebot3 based delivery system for a smart hospital in gazebo. In: 2021 International Siberian Conference on Control and Communications (SIBCON), pp. 1–6 (2021) 7. Pimentel, F., Aquino, P.: Performance evaluation of ROS local trajectory planning algorithms to social navigation. In: 2019 Latin American Robotics Symposium (LARS), 2019 Brazilian Symposium on Robotics (SBR) and 2019 Workshop on Robotics in Education (WRE), pp. 156–161 (2019)

126

A. Apurin et al.

8. Vagale, A., Oucheikh, R., Bye, R.T., Osen, O.L., Fossen, T.I.: Path planning and collision avoidance for autonomous surface vehicles I: a review. J. Marine Sci. Technol. 26(4), 1292– 1306 (2021). https://doi.org/10.1007/s00773-020-00787-6 9. Cybulski, B., Wegierska, A., Granosik, G.: Accuracy comparison of navigation local planners on ROS-based mobile robot. In: 2019 12th International Workshop on Robot Motion and Control (RoMoCo), pp. 104–111 (2019) 10. Gerkey, B. P., Konolige, K.: Planning and control in unstructured terrain. In: ICRA Workshop on Path Planning on Costmaps (2008) 11. Fox, D., Burgard, W., Thrun, S.: The dynamic window approach to collision avoidance. IEEE Robot. Autom. Mag. 4(1), 23–33 (1997) 12. Gehrig, S.K., Stein, F.J.: Elastic bands to enhance vehicle following. In: 2001 IEEE Intelligent Transportation Systems, pp. 597–602 (2001) 13. Keller, M., Hoffmann, F., Hass, C., Bertram, T., Seewald, A.: Planning of optimal collision avoidance trajectories with timed elastic bands. IFAC Proc. Volum. 47(3), 9822–9827 (2014) 14. Koenig, N., Howard, A.: Design and use paradigms for gazebo, an open-source multi-robot simulator. In: 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 2149–2154 (2004) 15. Apurin, A., et al.: LIRS-ArtBul: design, modelling and construction of an omnidirectional chassis for a modular multipurpose robotic platform. In: Interactive Collaborative Robotics: 7th International Conference (ICR 2022), pp. 16–18 (2022) 16. Gschwandtner, M., Kwitt, R., Uhl, A., Pree, W.: BlenSor: blender sensor simulation toolbox. In: Bebis, G., et al. (eds.) Advances in Visual Computing: 7th International Symposium, ISVC 2011, Las Vegas, NV, USA, September 26-28, 2011. Proceedings, Part II, pp. 199–208. Springer Berlin Heidelberg, Berlin, Heidelberg (2011). https://doi.org/10.1007/978-3-64224031-7_20 17. Apurin, A., Abbyasov, B., Dobrokvashina, A., Bai, Y., Svinin, M., Magid, E.: Omniwheel chassis’ model and plugin for gazebo simulator. Proc. Int. Conf. Artif. Life Robot. 28, 170–173 (2023). https://doi.org/10.5954/ICAROB.2023.OS6-7 18. Takaya, K., Asai, T., Kroumov, V., Smarandache, F.: Simulation environment for mobile robots testing using ROS and Gazebo. In: 2016 20th International Conference on System Theory, Control and Computing (ICSTCC), pp. 96–101 (2016) 19. Iskhakova, A., Abbyasov, B., Mironchuk, T., Tsoy, T., Svinin, M., Magid, E.: LIRS-MazeGen: an easy-to-use blender extension for modeling maze-like environments for gazebo simulator. In: Ronzhin, A., Pshikhopov, V. (eds.) Frontiers in Robotics and Electromechanics, pp. 147–161. Springer Nature Singapore, Singapore (2023). https://doi.org/10.1007/978-98119-7685-8_10 20. Valera, Á., Valero, F., Vallés, M., Besa, A., Mata, V., Llopis-Albert, C.: Navigation of autonomous light vehicles using an optimal trajectory planning algorithm. Sustainability 13(3), 1233 (2021)

Movement Along the Trajectory of a Home Quadruped Robot Dmitry Dobrynin(B) Federal Research Center “Computer Science and Control” of Russian Academy of Sciences, 44/2, Vavilova Street, 119333 Moscow, Russia [email protected]

Abstract. Home walking robots imitating pets have a high appeal due to increased maneuverability in a cramped home environment. Planning the movements of a home robot is an important component of the control system of a quadruped walking robot. The article deals with the problem of following the trajectory of a walking robot, which relates to motion planning. The article presents a model of a robot and a mathematical model of its legs. Two methods of approximation of the trajectory of motion are proposed in the article are piecewise linear approximation and approximation by arcs of a circle. The use of piecewise linear approximation makes it possible to solve the problem using simple robot movements. The use of approximation by arcs of a circle allows you to build a universal gait for a walking robot. The simulation of robot movements using two types of approximations is carried out. The article presents experimental modeling data. It is shown that the average speed of a walking robot with piecewise linear approximation is significantly lower than the speed of the robot moving in a straight line. The article presents the conclusions drawn from the results of experiments. Keywords: Quadruped Robot · Motion Planning · Walking Robot

1 Introduction Walking robots are currently an important area of research in the field of robotics. An attractive feature of walking robots is the ability to adapt to difficult terrain. For example, home walking robots have more maneuverability than wheeled robots. An important area of research in the development of walking robots is the planning of robot movement in a complex environment. Choosing a suitable trajectory planning algorithm helps to ensure safe and efficient point-to-point navigation, and the optimal algorithm depends on the geometry of the robot, as well as computational constraints. Some algorithms consider the robot as a point object, the structure of which is not important for solving the problem. Other algorithms take into account the peculiarities of robot movement, including static holonomic and dynamic non-holonomic-bounded systems. An overview of planning methods is given in [1, 2]. Movement along the trajectory of a walking robot can be attributed to the task of planning robot movements. Motion planning is aimed at generating interactive trajectories in the workspace when robots interact with a dynamic environment, therefore, when © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 127–136, 2023. https://doi.org/10.1007/978-3-031-43111-1_12

128

D. Dobrynin

planning movement, it is necessary to take into account the peculiarities of kinetics, speed and posture of robots and dynamic objects nearby. At this level of motion planning, the robot body and its individual parts move relative to surrounding objects. In this case, the robot cannot be considered as a point object, since the size of the individual parts of the robot greatly affects the nature of movement as a whole. An overview of motion planning methods is given in [3]. Research in the field of four-legged robots has been actively going on since the 1980s. Initially, gaits based on the kinematic model of the robot’s leg were used, similar to the gaits of insects and animals [4]. Matsuoka [5] proposed a CPG method to simulate rhythmic movement. CPG-based gait control uses a simplified single-layer feedback model. This method is widely used by many researchers [6–8]. CPG allows you to generate periodic gaits using a small number of parameters. Modern research in the field of walking robots is aimed at solving the problems of moving through difficult terrain [9–11]. With this method of movement, the robot must choose a fulcrum for each step-in order to optimally solve the problems of patency and stability. Such a robot gait is called a free gait. McGee and Frank [12] were engaged in the analysis of free gait when studying the movement of a robot over difficult terrain. To calculate a free gait, various methods are used is optimization of the center of gravity position to ensure a given stability [13], training of multilayer neural networks [14], etc. A good overview of some methods can be found in [15]. Free gait involves calculating each step of the robot, which requires a lot of computing power [16]. Home robots usually do not have powerful computers on board, so they need methods of movement that have low computational complexity. It is possible to combine the robot’s gaits, depending on the terrain on which the robot moves [17]. In this paper, we propose two ways to approximate the trajectory of the robot is straight lines and circular arcs. The approximation of the circle arcs allows you to build a universal gait for a home robot that requires small computing resources.

2 Model of Home Robot A home walking robot is similar to a dog (Fig. 1a). The robot has four legs of a simple design. Each leg can move with three degrees of freedom. The robot leg (Fig. 1b) has three drives M1, M2 and M3, which allow you to move the touch point of the foot in three dimensions. The touch point can move in some area relative to the robot body, as shown in Fig. 1a. The dimensions of this area are determined by the geometric parameters of the robot leg and the height of the robot body relative to the surface. The lower the robot body is located to the surface, the larger the area of possible movements of the touch point. In Fig. 2 presents a mathematical model of a walking robot’s leg. The initial parameters for calculating the angles of the drives θ1 , θ2 and θ3 are the coordinates of the fulcrum (X, Y, Z) with respect to the beginning of the leg. The calculation of the drive angles is similar to the equations presented in [18]. To simplify the calculations, we will assume that the robot body retains a horizontal position during movement. Since the home robot must move on flat surfaces, this

Movement Along the Trajectory of a Home Quadruped Robot

129

M1

M2

M3 foot feet contact point a)

b)

Fig. 1. a) Robot model; b) Model of robot’s leg.

M1

M1 M2 L2

θ2

H L4

θ3

M3

θ1

H

Z

Fy

L3 1 O

2

h

O

X LS

Y

Fig. 2. Mathematical model of robot’s leg.

assumption is quite justified. Y θ1 (Y , Z) = arc t g , Z L2 + L24 − L23 X θ2 (X , H ) = arccos 2 − arcsin , 2L2 L4 L4 L2 + L23 − L24 , θ3 (X , H ) = arccos 2 2L2 L3   L4 = X 2 + H 2 , H = Z 2 + Y 2 .

(1)

Here L 2 and L 3 are the lengths of the robot leg links, M1 , M2 and M3 are the robot leg drives, O is projection of the beginning of the foot on the support surface, X and Y are the displacement of the point of contact of the foot from the projection of the beginning of the foot in the plane of the support surface,

130

D. Dobrynin

Z is the height of the beginning of the foot from the surface, H is the distance from the beginning of the leg to the point of contact in the plane of the leg, h is the height of the leg lift above the surface, Ls is step length. The stability of the home robot can be ensured by the choice of gait [18]. The robot will be statically stable if the horizontal projection of its center of gravity lies inside the reference polygon. The reference polygon is formed by the points of contact of the robot’s legs with the reference surface. When performing the steps, the support phase and the leg transfer phase are selected in such a way that the stability condition is always met.

3 Movement Along the Trajectory The movement of a walking robot along the trajectory is one of the mandatory tasks when moving a robot in a difficult environment. Trajectories usually include linear and rounded sections. We will use two options: approximation of the target trajectory by straight lines and approximation by arcs of a circle. 3.1 Piecewise Linear Approximation One of the possible solutions is piecewise linear approximation of a given trajectory. In this case, the robot must move in steps along the trajectory of movement (Fig. 3a). Note that for this type of robot movement, two types of robot movement are necessary: moving straight and turning in place. Due to the fact that the turn requires the robot to stop, the general nature of the movement turns out to be intermittent. Consider the movement of the robot from point A to point C (Fig. 3a). At point A, the robot turns, then moves in a straight line to point B. After stopping, a turn is made, and then moving to point C.

'R

B

1

n*Ls R

A

a)

C

1

2

2

b)

r

c)

Fig. 3. a) Piecewise linear approximation of the trajectory; b) Moving the foot support points for linear movement; c) Moving the support points of the legs when turning.

Movement Along the Trajectory of a Home Quadruped Robot

131

The length of the robot’s elementary movement is a multiple of the robot’s step Ls. Equation (2) shows that the positioning error of the robot relative to the trajectory R is smaller the larger the turning radius R. As the step length Ls decreases, the error R also decreases:    nLs 2 R =1− 1− (2) R 2R To move in a straight line, the robot can use different types of gaits [18]. The point of contact of the robot’s foot with the surface thus performs a rectilinear movement relative to the robot body (Fig. 3b). Point 1 corresponds to the beginning of the reference phase (Fig. 2). Point 2 corresponds to the end of the reference phase. The length of the leg movement from point 1 to point 2 is equal to the step of the robot Ls. The trajectories of the reference points of all robot legs are the same. The rotation of the robot body by a certain angle is carried out in steps (Fig. 3c). In this case, the trajectories of the reference points of the robot’s legs lie on a circle of radius r. As can be seen from Fig. 3c, the trajectories of the reference points of the robot’s legs are different and represent the arcs of a circle with a common center. The advantages of piecewise linear movement along the trajectory include the simplicity of implementation, since the individual movements of the walking robot are simple. The disadvantage of piecewise linear movement along the trajectory is the low speed of movement. Stopping the robot to make a turn does not allow you to develop a high speed of movement. The accuracy of the trajectory repetition can be adjusted by changing the value of the robot’s step length Ls. 3.2 Gait with Variable Turning Radius (VTR Gate) To improve the nature of the robot’s movement along the trajectory, it is necessary to move the walking robot along the arc of a circle. From Fig. 3c, it can be seen that if the center of rotation is moved outside the body, then the nature of the movement of the support points of the legs becomes similar to the case of rectilinear movement. The design of the robot’s leg has three degrees of freedom, so it is possible to realize a circular trajectory of movement of the support points of the legs. The limitations in this case are the size of the leg movement area and compliance with the stability criteria of the robot. Let’s define such a gait as a VTR gate is a gait with a variable turning radius. This gait is periodic and has versatility. In Fig. 4a shows the principle of moving the robot around a circle with a radius of Rc. Each leg of the robot moves in such a way that the trajectory of its reference point (from 1 to 2) lies on a circle. The centers of all circles must match. The angular velocity of the movement of the reference points on the surface must also be the same. Under these conditions, the center of the robot will move in a circle around the common center C0 with an angular velocity ω. Note that each leg of the robot requires its own trajectory of movement of the reference point. When moving, the robot must rearrange its legs in accordance with the

132

D. Dobrynin

'R 1 Z

1 2

Cr 2 1

1

2

2

Rc

C0

R

A

B

R1 Rc R2

a)

b)

Fig. 4. a) Gate with variable turning radius; b) Robot positioning error.

selected type of gait. The reference points for the legs external to the center of the circumference move at high speeds. As the radius of the circle increases, the trajectories of the reference points increasingly approach straight lines. In the extreme case, when the radius of the circle is infinite, the trajectories of the reference points are straight lines. This case corresponds to the movement of the robot in a straight line. By changing the position of the center of the circle, you can change the orientation of the robot body during movement. The positioning error of the robot (Fig. 4b) in this case depends on the difference between the radius of rotation of the robot and the target trajectory. The mismatch of the centers of rotation also increases this error. VTR gate replaces several types of movements of a walking robot is movement in a straight line, rotation, and sideways movement. Note that when moving the pivot point back, such a gait can provide lateral movement of the robot body. When moving by a walking robot using the VTR gate, there is no need to stop for turning. Therefore, the speed of movement of the walking robot will be higher than when using piecewise linear movement.

4 Experiments For experimental verification, a robotic simulator was used (Fig. 5). The robot moved in a circle using two types of approximations are piecewise linear and VTR gate. The diameter of the circle = 280 cm, the linear speed of the robot was 10 and 20 cm/s. The robot’s step length is 5 cm. When using piecewise linear approximation, the robot performed one or more steps in a straight line, then stopped and made a turn. The rotation was performed by rotating

Movement Along the Trajectory of a Home Quadruped Robot

133

Fig. 5. Robot on the test polygon.

the robot body when all the legs of the robot were in the reference phase [18]. The turn execution time is about 2 s. When using the VTR gate, the robot moved along a circle of constant radius. Correction of the position of the center of the circle and the turning radius was not carried out. When moving the robot around the polygon, the average deviation of the robot from the specified trajectory and the average speed of movement were estimated. The results of the experiments are shown in Table 1. Table 1. Experimental results on the test polygon. V linear, cm/s

Piecewise linear approximation n*Ls

R, cm calculated

R, cm measured

V avg cm/s

10

5

0.022

4

1.9

10

0.09

5.5

3.3

20

VTR gate

15

0.2

7

4.3

5

0.022

6.5

2.2

10

0.09

8.5

3.9

15

0.2

10.5

5.5

R, cm measured

V avg cm/s

8.5

9.2

12.5

18.1

Let us analyze the experimental results for the case of piecewise linear approximation. For a low speed of movement of the robot (10 cm/s) with a small step, the highest accuracy of movement is achieved (deviation of about 2 cm). Note that the measured

134

D. Dobrynin

deviation (R measured) is much larger than the calculated deviation (R calculated), which was calculated by the formula (2). This is due to the fact that instabilities during the execution of the robot step greatly affect the accuracy of movement. The average speed of movement when using linear approximation is lower than the possible speed of movement in a straight line. This is due to the fact that the robot stops to perform a turn. At the same time, considerable time is spent, which reduces the average speed. As the length of the linear segments of the path increases, the average speed also increases. With an increase in the speed of movement up to 20 cm/s, an increase in the error of movement is observed. The instability of the robot’s step performance increases with increasing speed, so the movement error also increases. The average speed is also increasing but remains quite low compared to the linear speed. When using VTR gate, the average speed of movement is significantly higher than when using linear approximation. This is because the robot does not waste time stopping to turn on the spot. The average speed is slightly lower than the linear travel speed. The speed of movement of the legs external to the center of the circle is equal to the linear speed. The speed of the other parts of the robot is lower because they are closer to the center of rotation (Fig. 4a). The accuracy of robot movement when using VTR gate is slightly lower than with piecewise linear approximation. During the movement, the trajectory correction was not carried out, so errors in the execution of steps accumulated. It should be noted that in order of magnitude of the displacement error for the considered cases of piecewise linear displacement and VTR gate approximately coincide. To accurately track the trajectory, it is necessary to adjust the direction of movement of the walking robot depending on its deviation from the target trajectory. The difficulty in this case lies in the discrete nature of the steps and the complex control of the robot’s walking process.

5 Conclusion Moving a walking robot along a trajectory is a more difficult task than for wheeled robots. The robot’s steps are discrete and must be synchronized with each other in time. Note that for walking robots, movement planning and step execution are closely related, since when walking, the robot moves relative to the objects surrounding it. Known ways of moving along the path are the approximation of the trajectory by simple curves along which the robot can move. The article considers two types of approximation are piecewise linear approximation and approximation by arcs of a circle. The proposed methods of trajectory approximation allow solving the problem of movement along the trajectory for home quadruped robot. Approximation of the trajectory using rectilinear sections and turns allows you to solve the problem with fairly simple robot movements. The disadvantage of this representation is the intermittent movement of the robot, which reduces the average speed of movement. Approximation of the trajectory using arcs of circles allows you to build a universal gait for a walking robot (VTR gate). Such a gait can replace the rectilinear movement of the robot and turns on the spot. The calculation of the trajectories of the

Movement Along the Trajectory of a Home Quadruped Robot

135

reference points of the robot’s legs for this gait is more complex than for rectilinear gaits. The average speed of movement is slightly lower than for rectilinear movement and significantly higher than for piecewise linear approximation. Experiments have shown that when performing the movement of a walking robot using VPN gate, errors accumulate that occur when performing individual steps. To increase the accuracy of movement, it is necessary to adjust the position of the center of rotation and the radius of rotation. Note that to move a home robot in a home environment, high accuracy of movement is not required. It is enough that the robot does not touch obstacles and confidently passes difficult sections of the path. It should be noted that improving the accuracy of moving a walking robot is a separate difficult task and deserves separate consideration. The proposed methods of trajectory approximation for a home robot have a small computational complexity. This reduces the requirements for the necessary computing power for the control system of a home walking robot.

References 1. Karur, K., Sharma, N., Dharmatti, C., Siegel, J.E.: A survey of path planning algorithms for mobile robots. Vehicles 3(3), 448–468 (2021) 2. Jogeshwar, B.K., Lochan, K.: Algorithms for path planning on mobile robots. IFACPapersOnLine 55(1), 94–100 (2022) 3. Zhou, C., Huang, B., Fränti, P.: A review of motion planning algorithms for intelligent robots. J. Intell. Manuf. 33(2), 387–424 (2022) 4. De Santos, P.G., Garcia, E., Estremera, J.: Quadrupedal locomotion: an introduction to the control of four-legged robots, vol. 1. Springer, London (2006) 5. Matsuoka, K.: Mechanisms of frequency and pattern control in the neural rhythm generators. Biol. Cybern. 56(5–6), 345–353 (1987) 6. Shao, J., Ren, D., Gao, B.: Recent advances on gait control strategies for hydraulic quadruped robot. Recent Patents Mech. Eng. 11(1), 15–23 (2018) 7. Ijspeert, A.J.: Central pattern generators for locomotion control in animals and robots: a review. Neural Netw. 21(4), 642–653 (2008) 8. Barron-Zambrano, J.H., Torres-Huitzil, C., Girau, B.: Hardware implementation of a CPGbased locomotion control for quadruped robots. In: Diamantaras, K., Duch, W., Iliadis, L.S. (eds.) ICANN 2010. LNCS, vol. 6353, pp. 276–285. Springer, Heidelberg (2010). https://doi. org/10.1007/978-3-642-15822-3_35 9. Kim, H., Won, D., Kwon, O., Kim, T.J., Kim, S.S., Park, S.: Foot trajectory generation of hydraulic quadruped robots on uneven terrain. IFAC Proc. Vol. 41(2), 3021–3026 (2008) 10. Adak, O.K., Erbatur, K.: Bound gait reference generation of a quadruped robot via contact force planning. Int. J. Mech. Eng. Robot. Res. 11(3), 129–137 (2022) 11. Xu, X., Yang, Y., Pan, S.: Motion planning for mobile robots. In: Róka, R. (ed.) Advanced Path Planning for Mobile Entities. InTech (2018). https://doi.org/10.5772/intechopen.76895 12. McGhee, R.B., Frank, A.A.: On the stability properties of quadruped creeping gaits. Math. Biosci. 3, 331–351 (1968) 13. Estremera, J., de Santos, P.G.: Free gaits for quadruped robots over irregular terrain. Int. J. Robot. Res. 21(2), 115–130 (2002) 14. Yamamoto, H., Kim, S., Ishii, Y., Ikemoto, Y.: Generalization of movements in quadruped robot locomotion by learning specialized motion data. Robomech. J. 7, 1–14 (2020) 15. He, J., Shao, J., Sun, G., Shao, X.: Survey of quadruped robots coping strategies in complex situations. Electronics 8(12), 1414 (2019)

136

D. Dobrynin

16. Shao, Y., Jin, Y., Liu, X., He, W., Wang, H., Yang, W.: Learning free gait transition for quadruped robots via phase-guided controller. IEEE Robot. Autom. Lett. 7(2), 1230–1237 (2021) 17. Gehring, C., Coros, S., Hutter, M., Bloesch, M., Hoepflinger, M.A., Siegwart, R.: Control of dynamic gaits for a quadrupedal robot. In: 2013 IEEE International Conference on Robotics and automation, pp. 3287–3292. IEEE (2013) 18. Dobrynin, D.: Gait synthesis of a home quadruped robot. In: International Conference on Interactive Collaborative Robotics, vol. 13719, pp. 179–188. Springer International Publishing, Cham (2022). https://doi.org/10.1007/978-3-031-23609-9_16

Study of Path Planning Methods in Two-Dimensional Mapped Environments Nizar Hamdan, Viacheslav Pshikhopov , Mikhail Medvedev, Dimitry Brosalin, Maria Vasileva, and Boris Gurenko(B) Joint Stock Company “Scientific-Design Bureau of Robotics and Control Systems”, 154, Socialist St, Taganrog, Rostov Region 347900, Russia [email protected]

Abstract. The article studies the problem of path planning in two-dimensional environments. The review and analysis of known planning algorithms are carried out. This article is devoted to the development of a modified rapidly growing random trees algorithm (RRT) and the study of its effectiveness in comparison with known methods. The presented modified RRT algorithm checks the path to some area near the specified node while planning the path to a new potential node of the tree, which reduces the constructed nodes of the tree. The developed algorithm is compared with the original RRT algorithm. The comparison criteria are the path calculation time, the amount of memory required, the path length, and the percentage of situations in which the trajectory to the target point was successfully found. Next, the developed algorithm is compared with the planning algorithms of other classes. The study uses representative samples of numerical experiments and various environments that differ in the density of obstacles and the presence of mazes. A study of planning algorithms using the results of experiments on a ground-based wheeled robot has also been conducted. Keywords: Path Planning · Two-dimensional Environment · Random Tree Method · Optimization of Planning Algorithms · Comparative Analysis

1 Introduction Today, mobile robots have reached a level of development that allows them to be used autonomously. In this regard, the problem of optimal path planning in real-time is of importance. Path planning is a complex task that does not have a universal solution. Hence, the development and modification of various algorithms used in path planning, applicable and optimal for various conditions, is needed. In general, the task of path planning consists in calculating the sequence of states of a moving object that ensures its transfer from the initial state to the target state. Note that the transition from the initial state to the target can usually be done in several ways. In this regard, additional requirements are imposed on how the transition to the target state will be carried out. As a rule, the requirements for the optimal existential feasibility of the movement path are imposed. The requirements of existential feasibility include, for example, that the © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 137–150, 2023. https://doi.org/10.1007/978-3-031-43111-1_13

138

N. Hamdan et al.

requirement to go to the target point at a specified time and with limited control actions, including the absence of any collisions in the optimal trajectory and this means reducing or maximizing the quality criterion, for example, travel time, route length, energy expended, etc. Thus, the planning algorithm is understood as a set of rules which calculates the state of a moving object for each planning moment. As a result of the action of the planning algorithm, a set of states is formed at the output, which can be converted into a set of agent’s actions. According to [1], the trajectory planning of a moving object can be represented in general terms as the sequential execution of actions, including calculating the trajectory without collisions, smoothing the trajectory considering the characteristics of the object, and calculating control actions. Most often, the path is planned in the configuration space [1]. The main advantage of solving the path planning problem in configuration space is that this method reduces the calculations of a solid body to the calculations of a material point, which greatly simplifies the process [2]. For a moving object in two-dimensional Euclidean space, all configurations can be described by a vector (x, y, j), where x, y are the linear coordinates of the object, and j is the orientation angle. Space configurations are divided into subgroups {“Free”}, {“Obstacles”} and {“Unknown”}.

2 Overview of Motion Planning Methods Today, methods based on cellular decomposition are often used, which use the division of the map into cells. In the case of representing the configuration space in the lattice form, popular planning methods are Dijkstra’s algorithm and A* and D* based on it [3, 4]. Algorithm A* is a development of Dijkstra’s algorithm. In A*, the computational complexity is reduced due to the heuristic function of estimating the cost of the path. At the same time, this algorithm requires a lot amount of memory usage, and its computational complexity increases significantly as the number of cells on the map increases. In this regard, there are quite many modifications of the A* algorithm that are aimed at reducing these disadvantages [5–8]. The D* algorithm [4] is optimized for path planning in a dynamic environment, when information is updated during the robot’s movement. Note that the considered methods of path planning, as a rule, do not give smooth trajectories. In this regard, they are supplemented by various algorithms for smoothing the path [9]. Smooth movement paths are provided when using the artificial potential fields method for planning [10, 11]. Due to its simplicity, high computational efficiency and the indicated possibility of obtaining smooth trajectories of motion, the method of potential fields has become widely used [12, 13]. However, there are a few problems, which include local minimalism, the complexity of considering the dynamics of a moving body, and the reasonable selection of repulsive and attractive forces [14]. In this regard, potential fields are often used as a part of complex planning algorithms [15]. Planning methods using intelligent technologies, especially learning systems, have recently become popular [16, 17]. The advantage of neural network planning methods is their high adaptability to uncertain dynamic environments. The main disadvantage of planning systems is the high requirements for onboard computing systems [12].

Study of Path Planning Methods

139

Algorithms based on representing states in graph form are widely used. These include a planning algorithm for building a vision graph [18], Voronoi diagrams [19], rapidlygrowing random trees [20], and probabilistic road maps [21]. The main limitation of the visibility graph construction algorithm is its high computational complexity, which is estimated as O(n2 log(n)), where n is the number of nodes in [22], the problem of USV path planning in the waters of the coast of South Korea was considered. Due to many islands, the graph contains 54625 vertices. In this regard, the direct construction of the visibility graph is impractical. Therefore, in the article, adaptive meshing is first applied using the quadrant tree (quadtree) technology. Next, Dijkstra’s algorithm finds a path on the quadtree and builds a visibility graph using the obtained path points. The Voronoi diagram is a section of a plane with n centers in a set of convex polyhedrons such that any point within the polyhedron is closer to its center than the other centers [19]. This property allows you to plan the path farthest from obstacles if the centers of the latter act as points for the Voronoi diagram. The computational complexity of constructing a Voronoi diagram is estimated as O(n · log(n))[12]. The main disadvantages of the Voronoi diagram method are its low computational efficiency, which requires the use of additional algorithms and map processing. The probabilistic road map method [23] is used to solve problems of local and global planning. Due to high computational costs, the main efforts of researchers are aimed at improving the efficiency of the roadmap method, especially for narrow environments [24, 25]. The main role here is played by the method of generating new potential nodes of the graph. For example, in [24], a path planning method was proposed that uses a random roadmap algorithm. A new procedure for generating nodes in a limited area has been developed, and its effectiveness is shown in comparison with the original method of a randomized road map in a confined space. The paper [25] also studies the problem of enhancing the roadmap method in narrow spaces. For this, virtual fields are used that define the spaces in which potential new nodes of the roadmap are generated. Numerical experiments have been carried out, confirming the reduction of the time required to build a trajectory in a roadmap manner in bottlenecks. Thus, the roadmap method is common in path planning, but its computational efficiency is reduced in labyrinths, corridors, passages, and other bottlenecks. The method of rapidly growing random trees (RRT - rapidly random trees) is more computationally efficient than the probabilistic road maps method. However, the main limitations of this method for use in dynamic environments are low convergence with an optimal solution and high memory usage requirements. For this reason, an increase in computational efficiency will expand the scope of the application of the method of RRT. The problem of applying the RRT algorithm under conditions of partial uncertainty is considered in [26], where it is proposed to divide the area of operation into separate regions. The RRT algorithm is used to find paths between regions. Path planning within local areas is performed by a simple algorithm using straight line segments. The uncertainty of the map is considered by introducing the probability that the map cell is occupied by an obstacle. The main result presented in this paper is a method that ensures the achievement of the target point with a probability of at least a certain value

140

N. Hamdan et al.

pmin . The main disadvantage of this method is a significant increase in computational complexity with an increase in the value of pmin . The problem of reducing computational complexity and required memory was also considered in [26]. The work [26] combines the RRT algorithm and artificial potential fields (the P-RRT* method). In this method, a random tree is generated in the direction of reducing the potential repulsion field, which makes it possible to increase the rate of convergence to the optimal path. The computational complexity of the method remains very high; moreover, the methods of [26] give non-smooth paths. In [27], an improved version of the RRT* algorithm [28] was used. The bidirectional growth of a random tree is used in the RRT algorithm to increase the rate of convergence to the optimal solution. The main results of [27] indicate that bidirectional random search in combination with potential fields reduces the number of iterations to find the shortest path. In addition, these algorithms allow moving in narrow corridors without falling into local minima. To reduce the amount of memory required trees are computed sequentially by time, which increases the time complexity. In [29], is used a two-stage path planning procedure in a dynamic environment. The global planner computes the initial path of the first approximation using the RRT algorithm. Next, a dynamic planner is applied, which changes the position of the path nodes using repulsive and inertial forces. The prediction procedure is used to calculate moving obstacles. In this work, is used a heuristic optimization function, the choice of which depends on various requirements and limitations. The main problem of the method is the rational choice of this function. In [10], a two-stage planning algorithm is proposed, in which virtual obstacles are added to the map to consider the dynamic limitations of the mobile robot, blocking impossible sections of motion [30]. A modified version of the RRT algorithm, the MPN-RRT algorithm, is used as a global planning algorithm. This algorithm, unlike the original version, uses several parent nodes, which reduces the length of the planned path compared to the original version of RRT with a single node. To solve the problem of slow convergence of algorithms based on random search and path smoothing, the RRT algorithm is supplemented with a local optimization algorithm. The RRT algorithm searches for a global path, which is smoothed and optimized by the iterative local algorithm proposed in [31]. In this article, the task is to develop a modified RRT algorithm that is more efficient in terms of computation time than the original RRT algorithm. The task is also to conduct a study of the proposed algorithm in comparison with the original RRT method and other methods of path planning.

3 Mathematical Description of the Robot Movement and the Environment In the general case, the robot model can be represented by kinematics and dynamics equations of a rigid body and actuators [32]. For a ground-based wheeled robot mathematical model can be represented as [9, 12, 15]: x˙ = (ωl + ωr )r cos φ, y˙ = (ωl + ωr )r sin φ,

(1)

Study of Path Planning Methods

141

r φ˙ = (ωl − ω2 ) , l

(2)

ω˙ l = k1 ul , ω˙ r = k2 ur ,

(3)

where (x, y) are linear coordinates of the robot; ϕ is the robot orientation angle; ωl,r are the angular velocity of the left and right wheels; ul , ur are control actions; r is the radius of the wheels; l is the distance between the wheels; k 1 , k 2 are the constant coefficients. The operating environment of the robot is shown in Fig. 1.

Fig. 1. The operating environment of the robot.

The robot with coordinates (x, y) is indicated in Fig. 1 with an unshaded circle. The target with coordinates (xt , yt ) is represented as a double hatched circle. Obstacles are shown as shaded boxes and circles. Individual obstacles can form complex configurations, dead ends roads, walls, labyrinths, etc. Since obstacles can form random configurations, then it does not make much difference in the shape of individual obstacles. For this reason, both shapes will be used below – rectangles and circles. The path appears as a dotted line. The task is to plan the collision-free path L in such a way that: • for a given computation time t p , the path length L is minimized; • when the condition L ≤ L allowed is met, the computation time does not exceed the specified value t allowed . The RRT algorithm is adopted as the basic algorithm, which allows the bypass the problem of local minima, can be used in uncertain conditions, but as mentioned earlier, requires a long computation time.

4 Modified Rapidly-Growing Random Trees Algorithm To increase the efficiency of the RRT algorithm, the following heuristics methods are used to compensate for the existing disadvantages. This optimization is achieved through the following principles:

142

N. Hamdan et al.

• instead of checking the free path from the current node to the target node, it checks whether there is a path from the current node to an obstacle-free region around the target node, as shown in Fig. 2; • instead of checking the free path from the potential node to the parent node, the presence of a path from the current potential node to the free node is checked from the obstacles in the area around this parent node.

Fig. 2. Checking the path in the area near the target point.

Figure 3 shows a block diagram of the modified algorithm. The main part of the RRT algorithm has remained unchanged. The difference between RRT algorithm and the proposed algorithm is the additional verification. Before forming a new node, it is checked that there is a free direct path from the current node 4 in Fig. 2 to target node 1. If the straight path is crossed by obstacles, the free area around the target node is calculated. A diameter is built through the center of the free zone, straight lines are generated on it with a given step expand from the current node. Nodes are initialized at the intersection of lines with the boundaries of the free region. It checks for a free direct path from the current node to the node on the boundary of the free zone. If a free direct path from the current node was found, then a direct path is laid. Further, if the path is found then algorithm completes the execution, otherwise, the algorithm continues to build the graph according to the original method. Similarly, the free path from the newly generated node to the parent node is checked the same way. Thus, the application of the proposed heuristic method expands the list of adjustable parameters of the planning algorithm. In the implemented classical algorithm, the following are configured: the size of the rand_area region, in which a random point is selected in the configuration space; the maximum distance expand_dist that a tree branch can grow from the parent node towards an arbitrary point in space; The target_sample_rate probability that the path target node is chosen as the point towards which the tree branch grows; the maximum number of max_iter attempts to build a new tree node. If this number is exceeded, the algorithm terminates. Since checking the path to a given area introduces additional computational costs, it is not performed every step. In this regard, an additional configurable parameter is a frequency of checking the presence of a path to the free zone boundary around the target

Study of Path Planning Methods

143

Start

Initialization of the start and end node of the tree

Getting environment data

Target point reached?

Yes

No Free path to the target point?

Yes

No Definition of R-free area around the target point Finding points on R-free boundary with D-expand step

Free direct path to a point on R-free?

Yes

No Selection of coordinates of a random point P-rand

Adding a new tree node, coinciding with a free point on R-free

Choosing the N-nearest tree node closest to the point P-rand Path computation from new node and end point Compute a new path of length Dexploration from N-nearest in the direction of P-rand

New-path intersects with obstacles? Yes

No

Finding a path from the start to the end node

Adding a new tree node The end

Fig. 3. Modified RRT algorithm.

point straight_path_to_area_check. Determination of the presence of a direct path from the current node to the boundary of the free zone around the target point or the generated new node is performed according to the following algorithm: Step 1. Calculation of the radius of the free zone around the target point or the new generated node: Rfree – equal to the distance to the boundary of the nearest obstacle.

144

N. Hamdan et al.

Step 2. Segmentation of the diameter of the free zone d free into segments with the step d expand specified in the algorithm parameters. Step 3. Finding the intersection points of the straight lines from the current node to the points on the diameter of the free zone with the boundary of the free zone. Step 4. Determination of the presence of a collision-free path along one of the straight lines going from the current node to a point on the border of the free zone around the target point.

5 Comparison of the Original and the Modified RRT Algorithm A comparative modeling of two methods was carried out: the original RRT algorithm and the modified algorithm shown in Fig. 3. The algorithms were modeled using the Python 3.9 programming language. The simulation was performed using software simulation in the PyCharm integrated development environment. The simulated environment is represented as a two-dimensional coordinate grid, at each point of which a robot can be located, described by a positioning point (x, y) and an orientation angle ϕ (Fig. 1). Obstacles are described by a set of circles, which include a safety buffer zone. Single action means a change in the state of the agent, namely one or more of its coordinates (x, y). The modeling complex is implemented in the PyCharm environment. The moving object is represented by Eqs. (1), (2) and (3). The modeling of the navigation system and vision system blocks are implemented by the description presented in [32]. The calculation of the global path is implemented in the path planning system, where both the classical RRT and the modified version proposed in this article are implemented. The actuator sensors are considered as ideal, and the actuators themselves are proportional inertia–free links. The controller was synthesized by the method of positional trajectory control [12, 32]. Each obstacle in the form of a circle is described by an ordered list of the form (xobst , yobst , robst ), where x obst , yobst are the coordinates of the center of the obstacle, r obst is the radius of the danger zone around the center of the obstacle. A rectangular obstacle is described by a tuple of the form (xobst , yobst , aobst , bobst ), where x obst , yobst – the coordinates of the lower left corner of the obstacle, aobst , bobst are the width and length of the obstacle. In the present study, 270,000 iterations of the simulation were carried out. Each iteration includes the following steps: 1. Generation of a random situation on the map: coordinates of the initial and target points of the path, a given number N obst of obstacles with random parameters. 2. Simulation of the robot’s movement on the map using the original RRT algorithm. 3. Simulation of the robot’s movement on the map using the modified RRT algorithm. 4. Fixing the performance indicators of the original and modified RRT algorithms. The indicators include algorithm running time (t RRT , t new_RRT ); the amount of memory required (mRRT , mnew_RRT ); the length of the resulting path (lRRT , l new_RRT ); the frequency of successful goal achievement without collisions ( f RRT , f new_RRT ). A series of launches was performed on maps of different dimensions mapdim = [50:50, 500:500, 5000:5000], with a different number of obstacles mapobst = [10, 30, 50], with different positions of the starting and target points.

Study of Path Planning Methods

145

The simulation results are represented in Tables 1, 2. Examples of the operation of the classical and modified RRT algorithms are shown in Figs. 4 and 5. The following designations are used in the tables: t av – the average running time, mav – the average amount of used memory, lav – the average path length, in – the percentage of unsuccessful iterations of modeling, imax – the maximum number of iterations, mapdim – the map dimension. Modified path calculated in 3.4591 sec

-20

-20

0

0

20

20

40

40 6

60

80

80 100

Original path calculated in 19.7626 sec

0

20

40

60

A) modified RRT

80

100

-40 -20

0

20

40

60 80 100 120

B) Original RRT

Fig. 4. An example of modeling classical and modified RRT methods for constructing a path.

According to the data presented in Table 1, using of the modified RRT algorithm with an increased number of obstacles allowed to reduce the computation time iby17.4– 46.8%, reduce the amount of memory used by 45.6–86.0%, and reduce the path length by 3.0–13.5%. In addition, the original RRT method failed to find the path, depending on the size of the map and the maximum number of iterations, in 0.13–11.7% of cases. At the same time, the modified RRT algorithm did not find the path in 0–1.7% of cases. A total of 90000 experiments were performed using the parameters described above. Analyzing to the data presented in Table 2, we determine that using of the modified method, with the number of obstacles 50, allowed to reduce the computation time by 14.3–47.3%, reduce the amount of memory used by 29.8–80.2%, and reduce the path length by 1.9–8.9%. In addition, the original RRT method failed to find the path, depending on the size of the map and the maximum number of iterations, in 6.9–88.7% of cases. At the same time, the modified RRT algorithm did not find the path in 0.2–4.6% of cases. Note that as the number of obstacles and the dimension of the map increase, the effect of the heuristics used in this article is decreases a little. So, from Figs. 6 and 7 it can be seen that on a map with a dimension of 5000 × 5000 with the number of obstacles equal to 50, the modified RRT algorithm loses in time to the classical algorithm if the maximum number of iterations is more than 1000. Due to the fact that on a large map with many obstacles, it becomes increasingly difficult to determine the direct path from the current point to a given area Table 3 shows the success percentage of the modified algorithm according to all criteria for all 270,000 experiments conducted.

146

N. Hamdan et al.

original path calculated in 83.8 sec

0

0

10

10

20

20

30

30

40

40

50

50

60

60

Modified path calculated in 24 sec

-10 0

10

20

30

40

50

60 70

-10 0

10

A) modified RRT

20

30

40

50

60 70

B) Original RRT

Fig. 5. An example of modeling the classical and modified RRT methods for constructing a path in the presence of two walls. Table 1. Simulation results for N obst = 30. imax 700 1500 3000 imax 700 1500 3000

Method

t av , second

mapdim

50

500

5000

50

l av , unit 500

5000

RRT

0.0097

0.0266

0.0591

338108

1709574

2760827

M-RRT

0.0051

0.0141

0.0400

326533

1565291

2387731

RRT

0.015

0.0474

0.1374

399790

2930239

11103805

M-RRT

0.008

0.0309

0.1135

385079

2745523

10001515

RRT

0.0190

0.1124

0.2827

413072

3305314

29437582

M-RRT

0.0104

0.0619

0.2288

400049

3106780

27433970

Method

mav , byte

mapdim

50

500

5000

50

500

5000

RRT

1224.01

2502.32

3475.78

11.7%

43.22%

84.54%

M-RRT

636.28

713.05

488.05

1.69%

1.52%

0.13%

RRT

1541.64

3455.61

6107.81

3.05%

18.16%

57.96%

M-RRT

816.14

1229.59

1238.44

0.4%

1.41%

0.6%

RRT

1679.25

5821.56

6905.18

0.74%

11.76%

17.52%

M-RRT

914.10

2105.76

2254.93

0.08%

0.93%

1.25%

in , %

Study of Path Planning Methods

147

Table 2. Simulation results for N obst = 50. imax

Method mapdim

50

500

5000

50

500

5000

700

RRT

0.0133

0.0326

0.0772

222692

570338

1826683

M-RRT

0.0088

0.0172

0.0652

216445

519433

1608241

RRT

0.0268

0.0666

0.1600

345404

1370839

6354044

M-RRT

0.0188

0.0486

0.1866

338594

1288650

5814408

RRT

0.0463

0.1328

0.2980

427098

2317289

14420732

M-RRT

0.0336

0.1062

0.4708

419115

2198826

13550898

Method

mav , byte

mapdim

50

500

5000

50

500

5000

RRT

1073.89

2456.67

3095.97

35.96%

74.55%

88.73%

M-RRT

679.88

691.59

612.80

4.63%

1.14%

0.2%

1500

RRT

1625.69

3733.38

5158.43

16.82%

53.0%

72.33%

M-RRT

1089.77

1463.61

1390.18

3.14%

3.0%

1.29%

3000

RRT

2174.86

5289.81

7529.37

6.94%

33.07%

51.21%

M-RRT

1526.72

2466.74

2863.20

1.11%

3.53%

2.67%

in , %

0.03

0.04

RRT-500-10 Modified RRT-10 RRT-500-10 Modified RRT-10 RRT-500-10 Modified RRT-10

0.02

700

Calculation time, sec

imax

0.01

3000

l av , unit

0

1500

t av , second

0

500

1000 1500 2000 Maximum iterations/unit

2500

3000

Fig. 6. Graph of the change in the path search time on the map 50 × 50.

N. Hamdan et al. RRT-500-10 Modified RRT-10 RRT-500-10 Modified RRT-10 RRT-500-10 Modified RRT-10

0.3 0.2 0

0.1

Calculation time, sec

0.4

148

0

500

1000 1500 2000 Maximum iterations/unit

2500

3000

Fig. 7. Graph of the change in the path search time on the map 5000 × 5000. Table 3. Simulation results grouped by the size of the environment. Map dimensions

t av , second

mav , byte

l av , unit

50 × 50

55,48%

54,67%

8,03%

500 × 500

53,06%

70,56%

9,56%

5000 × 5000

21,16%

81,46%

11,27%

6 Conclusion An analysis of the results of a series of experiments allows us to draw a few conclusions in order to determine in which situations the use of the modified RRT algorithm is effective, and in which situations the original algorithm is more effective. With the increase in the complexity of the environment and the increase in the number of obstacles, the advantage of the modified algorithm decreases (but remains) in terms of t average , maverage , l average . This is due to an increase in computational costs for determining the presence of a direct path - the duration of this procedure is directly proportional to the number of obstacles on the map. This disadvantage can be partially compensated by optimizing of computations. The decrease in the advantage value in terms of mav and l av is associated with an increase in the number of maneuvers that the agent must perform when going around many obstacles. Many obstacles on the map reduces the number of sections in which the modified algorithm outstands the original one. As the dimensions of the environment increase, the advantage of the modified algorithm in terms of t av decreases but remains. The advantage of mav and l av increases. The change in the advantage is associated with an increase in the number of iterations of tree

Study of Path Planning Methods

149

growth performed by the algorithm; respectively, the modified algorithm performs more computations that are heuristic and spends more time on them. Using the modified algorithm increases the percentage of the probability of finding a solution within the specified requirements for the allowed number of imax (Tree growth iterations). Among all the experiments, the modified algorithm did not find a solution in 1.07% of the cases; meanwhile the original algorithm solution did not find a solution in 30.13% of the cases. Thus, the modified algorithm showed an advantage: in cases of environment without obstacles; in cases where there is a limit on the maximum number of iterations; in cases where the memory space allowed for storing the computation tree is limited. The disadvantages of the algorithm include the dependence of temporal computational efficiency on the environment dimension for medium and high obstacle density maps. Acknowledgements. The work was supported by the grant of the Russian Science Foundation No. 22-29-00533, carried out in JSC “Scientific Design Bureau of Robotics and Control Systems”.

References 1. LaValle, S.: Planning Algorithms, pp.1–842. Cambridge University Press (2006) 2. Lozano-Perez, T.: Spatial planning: a configuration space approach. IEEE Trans. Comput. 32(2), 108–120 (1983) 3. Hart, P.E., Nilsson, N.J., Raphael, B.A.: Formal basis for the heuristic determination of minimum cost paths. IEEE Trans. Syst. Sci. Cybern. 2, 100–107 (1968) 4. Stentz, A.: Optimal and efficient path planning for partially known environments. In: Intelligent Unmanned Ground Vehicles, pp. 203–220. Springer, Boston, MA, USA (1997) 5. Wang, Q., Hao, Y., Chen, F.: Deepening the IDA* algorithm for knowledge graph reasoning through neural network architecture. Neurocomputing 429, 101–109 (2021) 6. Zhou, R., Hansen, E.A.: Memory-bounded {A*} graph search. In: The Florida AI Research Society Conference – FLAIRS, pp. 203–209 (2002) 7. Holte, R., Perez, M., Zimmer, R., MacDonald, A.: Hierarchical A*: searching abstraction hierarchies efficiently. Proc. Thirteenth Natl. Conf. Artif. Intell. 1, 530–535 (1996) 8. Liu, B., Xiao, X., Stone, P.: A lifelong learning approach to mobile robot navigation. IEEE Robot. Autom. Lett. 6(2), 1090–1096 (2021) 9. Pshikhopov, V., Medvedev, M., Kostukov, V., Houssein, F., Kadim, A.: Algorithms for trajectory planning in a two-dimensional environment with obstacles. Inf. Autom. 21(3), 459–492 (2022) 10. Khatib, O.: Real-time obstacles avoidance for manipulators and mobile robots. Int. J. Robot. Res. 5(2), 90–98 (1986) 11. Platonov, A., Karpov, I., Kirilchenko, A.: The method of potentials in the laying of the route. M.: Preprint of the Institute of Applied Mathematics of the USSR Academy of Sciences 27 (1974) 12. Pshikhopov, V.: In: Beloglazov, D., et al.: (eds.) Path Planning for Vehicles Operating in Uncertain 2D Environments, Elsevier, Butterworth-Heinemann, ISBN: 9780128123058, 312 (2017)

150

N. Hamdan et al.

13. Filimonov, A., Filimonov, N.: Issues of motion control of mobile robots by the method of potential guidance. Mekhatronika, Avtomatizatsiya, Upravlenie 20(11), 677–685 (2019) 14. Pshikhopov, V., Medvedev, M.: Group control of the movement of mobile robots in an uncertain environment using unstable regimes. Proc. SPIIRAN 60, 39–63 (2018) 15. Malone, N., Chiang, H.-T., Lesser, K., Oishi, M., Tapia, L.: Hybrid dynamic moving obstacle avoidance using a stochastic reachable set-based potential field. IEEE Trans. Rob. 33(5), 1124–1138 (2017) 16. Gaiduk, A., Martyanov, O., Medvedev, M., Pshikhopov, V., Hamdan, N., Farkhud, A.: Neural network control system for a group of robots in an unidentified two-dimensional environment. Mechatron., Autom., Control 21(8), 470–479 (2020) 17. Wang, Y., Cheng, L., Hou, Z.-G., Yu, J., Tan, M.: Optimal formation of multirobot systems based on a recurrent neural network. IEEE Trans. Neural Netw. Learn. Syst. 27(2), 322–333 (2016) 18. De Berg, M., Cheong, O., Van Kreveld, M., Overmars, M.: Computational Geometry: Algorithms and Applications, 3rd edn. Springer-Verlag (2008) 19. Guibas, L.J., Knuth, D.E., Sharir, M.: Randomized incremental construction of delaunay and voronoi diagrams. Algorithmica 7(1), 381–413 (1992) 20. LaValle, S.M., Kuffner, J.J.: Rapidly exploring random trees: progress and prospects. In: Workshop on the Algorithmic Foundations of Robotics, pp. 293–308 (2000) 21. Kedem, K., Sharir, M.: An efficient motion planning algorithm for a convex rigid polygonal object in 2-dimensional polygonal space. Discrete Comput. Geom. 5(1), 43–75 (1990) 22. Lee, W., Choi, G.-H., Kim, T.-W.: Visibility graph-based path-planning algorithm with quadtree representation. Appl. Ocean Res. 117, 102887 (2021) 23. Kavraki, L.E., Svestka, P., Latombe, J.C., Overmars, M.H.: Probabilistic roadmaps for path planning in high-dimensional configuration spaces. IEEE Trans. Robot. Autom. 12(4), 566– 580 (1996) 24. Kamil, A.R.M., Shithil, S.M., Ismail, Z.H., Mahmud, M.S.A., Faudzi, A.A.M.: Path planning based on inflated medial axis and probabilistic roadmap for duct environment. Lecture Notes in Electrical Engineering 834 (2022) 25. Chen, G., Luo, N., Liu, D., Zhao, Z., Liang, Ch.: Path planning for manipulators based on an improved probabilistic roadmap method. Rob. Computer-Integrated Manuf. 72, 102196 (2021) 26. Qureshi, A., Ayaz, Y.: Potential functions-based sampling heuristic for optimal path planning. Auton. Robot 40, 1079–1093 (2016) 27. Karaman, S., Frazzoli, E.: Sampling-based algorithms for optimal motion planning. Int. J. Robot. Res. 30(7), 846–894 (2011) 28. Chen, L., Shan, Y., Tian, W., Li, B., Cao, D.: A fast and efficient double-tree RRT∗-like sampling-based planner applying on mobile robotic systems. IEEE/ASME Trans. Mechatron. 23(6), 2568–2578 (2018) 29. Wang, J., Meng, M.Q.-H., Khatib, O.: EB-RRT: optimal motion planning for mobile robots. IEEE Trans. Autom. Sci. Eng. 17(4), 2063–2073 (2020) 30. Medvedev, M., Pshikhopov, V., Gurenko, B., Hamdan, N.: Path planning method for a mobile robot with maneuver restrictions. In: Proceedings of the International Conference on Electrical, Computer, Communications and Mechatronics Engineering (ICECCME) (2021) 31. Kostjukov, V., Medvedev, M., Pshikhopov, V.: Method for optimizing of mobile robot trajectory in repeller sources field. Inf. Autom. 20(3), 690–726 (2021) 32. Pshikhopov, V., Medvedev, M.: Multi-loop adaptive control of mobile objects in solving trajectory tracking tasks. Autom. Remote Control 81(11), 2078–2093 (2020)

DHC-R: Evaluating “Distributed Heuristic Communication” and Improving Robustness for Learnable Decentralized PO-MAPF Vladislav Savinov1(B)

and Konstantin Yakovlev2

1 St Petersburg State University, 7-9, Universitetskaya Emb., St. Petersburg 199034, Russia

[email protected] 2 Federal Research Center “Computer Science and Control” of the Russian Academy of

Sciences, 44/2, Vavilova street, Moscow 119333, Russia

Abstract. Multi-agent pathfinding (MAPF) is a problem of coordinating the movements of multiple agents operating a shared environment that has numerous industrial and research applications. In many practical cases the agents (robots) have limited visibility of the environment and must rely on local observations to make decisions. This scenario, known as partially observable MAPF (POMAPF), can be solved through decentralized approaches. In recent years, several learnable algorithms have been proposed for solving PO-MAPF. However, their performance is oftentimes not validated out-of-distribution (OOD), and the code is often not properly open-sourced. In this study, we conduct a comprehensive empirical evaluation of one of the state-of-the-art decentralized PO-MAPF algorithms, Distributed Heuristic Communication (DHC), Ma, Z., Luo, Y., Ma, H.: Distributed heuristic multi-agent path finding with communication. In: 2021 International Conference on Robotics and Automation (ICRA), pp. 8699–8705. IEEE, Xi’an, China (2021), which incorporates communication between agents. Our experiments reveal that the performance of DHC deteriorates when agents encounter complete packet loss during communication. To address this issue, we propose a novel algorithm called DHC-R that employs a similar architecture to the original DHC but introduces randomness into the graph neural network-based communication block, preventing the passage of some data packets during training. Empirical evaluation confirms that DHC-R outperforms DHC in scenarios with packet loss. Open-sourced model weights and the codebase are provided: https://github.com/acforvs/dhc-robust-mapf. Keywords: PO-MAPF · Reinforcement Learning · Generalization · Out-of-distribution · AI safety

1 Introduction Multi-agent pathfinding (MAPF) is a problem of coordinating the movements of multiple agents to achieve a common goal in a shared environment. One of the most challenging scenarios in MAPF arises when agents cannot observe the entire environment in © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 151–163, 2023. https://doi.org/10.1007/978-3-031-43111-1_14

152

V. Savinov and K. Yakovlev

which they operate and must make decisions based on partial observations. In such cases, communication with a central controller might also be restricted, and each agent might plan its movements in a decentralized manner. This scenario, known as partially observable multi-agent pathfinding (PO-MAPF, see Fig. 1), poses a complex challenge as agents must cooperate without explicit coordination. Addressing this challenging task has traditionally relied on classical computer science or heuristic algorithms. However, with the increasing availability of data and advancements in computational capabilities, researchers have begun exploring learnable algorithms for these problems. Reinforcement learning (RL), in particular, has emerged as a promising tool for addressing these challenges.

Fig. 1. Example of a single PO-MAPF scenario; image from [1].

Previous learnable decentralized algorithms did not provide thorough empirical evaluation of algorithms and failed to consider important real-world scenarios, such as a complete packet loss in communication-based algorithms. This work is aimed at closing this gap. Our main contributions are: 1. We extend the empirical evaluation for one of the state-of-the-art learnable decentralized PO-MAPF algorithms, DHC [2], and benchmark it out-of-distribution. We provide empirical evaluation for this algorithm in scenarios where a complete packet loss occurs. 2. We introduce a novel algorithm, DHC-R, built upon the DHC architecture, that handles the scenario of a complete packet loss significantly better than the original algorithm. 3. We fully open source the code base and set up programming instruments for easy development and experiments.

2 Background The problem of multi-agent pathfinding (MAPF) consists of the computation of collisionfree paths for a group of agents from their location to an assigned target. The elements of a classical MAPF problem are as follows: 1. An undirected graph G = (V , E). 2. A set of N agents; each agent i is associated with its starting point si ∈ V and the unique target point gi ∈ V . It is assumed that the time is discrete, and that each agent can perform one action at each timestep. Path, or plan, for the i-th agent is a sequence of vertexes v0 , v1 , . . . , vN : vj ∈ V ; v0 = si , vN = gi . The vertex vj is a result of a taken action (usually wait or move)

DHC-R: Evaluating Distributed Heuristic Communication

153

  from the previous vertex vj−1 , e.g., for every j ≥ 1 either vj = vj−1 or vj−1 , vj ∈ E, meaning that two consecutive vertexes in a path are either adjacent or identical. The collision between two agents occurs when one of the following transpires: 1. At timestamp t two agents end up being in the same vertex s ∈ V . This can happen if they either came from different vertexes to s or one of the agents stayed in place, and the other moved to s at timestamp t. 2. Two agents, being located at s1 ∈ V and s2 ∈ V at timestamp t − 1 changed their positions to s2 and s1 correspondingly at t. The solution for the MAPF problem is a set of N collision-free paths, one for each agent. In the partially observable setting, agents do not receive the full graph as an input. Instead, at each timestamp, they can observe only the part of the environment, called field of view or FOV. In the example shown in the Fig. 1, FOV is represented by a small square window around the agent. The PO-MAPF problem can be formulated as a Markov decision process (MDP) and described by a 7-tuple (N , S, {Ai }, {Oi }, {Ri }, P, γ ), where N is a number of agents, S is a set of states, Ai is a set of actions allowed for agent i, Oi is a set of observations for agent i, P is the dynamics of the MDP, Ri : S × A → R is a reward function for  the i-th agents and γ is a discount rate, 0 < γ < 1. Dynamics of the  MDP P s , o|s, a outputs the probability of moving to the state s from s and receiving an observation o for it by selecting an action a . A reward function for the i-th agent outputs a numerical signal for the action that was taken from a given state. Discount rate determines how much the future rewards are important to the current state. In a finite MDP, Ai , Oi and S are finite sets. A policy maps each state s in the state space to a probability π (a|s) for each action a in the action space. The solution to the PO-MAPF problem for each agent is a policy that maximizes the sum of discounted with the rate γ rewards. In this manuscript, we consider PO-MAPF problem to be located on a 2D map, called a grid, which consists of empty cells where agents can be located and blocked cells occupied by obstacles. In our setting, each agent can undertake 1 out of 5 available actions: move left, right, up or down, or stay still. Collisions which were described above are also not allowed. The goal of each agent is to reach the target. In this work, a policy is a neural network that receives an encoded state/observation from the environment and outputs π(a | s). To evaluate a policy once it is found, several metrics are usually utilized: 1. Collective success rate (CSR): The percentage of tasks that were fully solved, meaning that each agent successfully reached its corresponding goal. The higher, the better. 2. Individual success rate (ISR): The percentage of agents that arrived at their goals. The higher, the better. During test invocation, these metrics are usually averaged across different runs. Reinforcement learning [3] is learning from interactions with an environment to maximize a numerical signal, called reward. In general, there are two types of tasks in RL: episodic and continuing tasks. In episodic ones, the agent’s interactions with the environments break into subsequences called episodes. Each episode ends in a terminal state, which is a special state that resets the agent’s state to one of the starting states. In

154

V. Savinov and K. Yakovlev

continuing tasks, there are no terminal states, and the agents interacts with the environment forever. The goal of a reinforcement learning agent is to maximize the expected return after timestep t which can be defined as follows: Gt =

∞ 

γ k rt+k+1 .

k=1

In the equation above, ri is the reward received at timestep i. If the task is episodic with the terminal state at timestamp T , all rT +1 , rT +2 , . . . are zeroed. A policy π : S × A → [0, 1] is a mapping from states to actions; π(a | s) is the probability of choosing the action a in s. The value function of a state s under policy π , Vπ (s), is the expected return that the agent receives starting at state s and following the given policy π . Vπ (s) can be formally defined as follows: ∞   k Vπ (s) = E[Gt |St = s] = E γ rt+k+1 |St = s . k=0

The state-action value function Qπ (s, a) outputs the expected return when following the policy π starting at s and taking the action a as the initial action. ∞   k Qπ (s, a) = E[Gt |St = s, At = a] = E γ rt+k+1 |St = s, At = a . k=0

In addition, an advantage function is usually defined as the different between the Q-function and a V -function: Aπ (s, a) = Qπ (s, a) − Vπ (s). In the past, numerous methods have been suggested to find the optimal policy that maximizes the sum of discounted rewards. While these methods are beyond the scope of this study, one commonly used approach is to approximate or precisely calculate the Qfunction. In this work, we adopt this approach and train a neural network to approximate the Q-function that can be then written as Q(s, a, θ ). Once the network is trained, agents can utilize it to select the action with the highest expected return.

3 Related Work Heuristic Algorithms. In a single agent setting, the goal is to find the optimal path from a starting point to a finishing point. The standard approach is to use the A* [4] algorithm. This algorithm uses a heuristic function to estimate the distance from the current location of the agent to its goal and guide the algorithm towards the optimal solution. Although A* can find the optimal solution, it may not be suitable for real-world usage due to its time complexity. There are many other single-agent algorithms, but since they are not our main focus in this work, we will not be investigating them further. For multiagent scenarios, there exist several methods that extend the A* algorithm. As a basic modification, one could run A* in a joint space of actions. However, this would result in an exponential growth of the number of vertexes in a graph, and the algorithm would

DHC-R: Evaluating Distributed Heuristic Communication

155

take a long time to output a solution for a large number of agents. Further modifications include LRA* [5], M* [6] or OdRM* [7] that plan each agent independently but transfer the planning to a higher dimensional space once the collision occurs. Other techniques, including Conflict Based Search (CBS)-based [8–11] algorithms were also introduced for this problem. In addition, reduction-based algorithms [12, 13] that reduce the MAPF problem to a SAT problem were suggested. Although these algorithms may provide optimal solutions, they are not polynomial and can take a long time to converge. In situations where optimality is not guaranteed, conflicts can arise during plan execution, and agents may need to communicate with a central controller for replanning. This may result in frequent calls to the global expert, causing agents to wait for a response and remain idle. Learning-based Algorithms. In response to the challenges outlined above, researchers have proposed several non-heuristic methods for MAPF. Among these, RL approaches have been particularly appealing [2, 14, 15]. While some RL-based methods may still rely on a central planner, they use it not as a sole point of coordination, but as an expert to guide learning during training. The expert generates trajectories, and agents try to replicate them. These algorithms typically use imitation learning (IL) and can be executed in a decentralized manner. They can also accommodate partially observable maps. For instance, in [16], a switch determines whether the next stage will be RL- or ILbased. During the IL phase, the central planner generates new trajectories that the agent can follow using behavioral cloning. During the RL stage, agents regress on generated trajectories to learn. However, IL builds upon the idea that experts generate trajectories of good quality. This is not the only problem with the described approach. First, the experts themselves may be too complex and time-consuming to run, making it impractical to use them for training. Additionally, while regressing on expert demonstrations, the model may not learn how to efficiently deviate from them even if doing so would be beneficial [2]. Recently, there has been growing interest in communication-based algorithms that leverage the exchange of packets between agents to improve coordination. Early works in this area used a collective reward to guide agent behavior, but later approaches employed individual rewards, enabling decentralized execution. All these methods have been trained in a centralized manner. More recent works have explored decentralized training methods as well. For example, in [17], agents form clusters based on learned priorities, and communication occurs only within clusters. Central agents receive and aggregate data packets from others, then transmit the results back to their clusters. The development of graph neural networks (GNNs) has led to recent works leveraging this powerful architecture for the MAPF problem. One early work is [18], which combines imitation learning with communication using CBS as a central planner. However, it is worth noting that the number of works that utilize communication and decentralized training together remains limited. For instance, in [2], agents receive information from all neighbors and consider packets only from the two closest ones. Another study [19] adopts a similar architecture but employs a request-reply mechanism instead of broadcasting, allowing only the most crucial agents to exchange information.

156

V. Savinov and K. Yakovlev

4 Method The objective of this work is to address a scenario where communication-based algorithms are unable to utilize inter-agent communication during inference, which can occur in real-world situations due to technical issues. To accomplish this, we propose a novel algorithm called DHC-R, which builds upon DHC [2]. Additionally, our aim is to validate the original DHC algorithm [2] across various settings to expand upon previous empirical evaluations. To begin, we thoroughly reproduce the DHC algorithm and refer to this reproduced version as DHC-paper. Building upon the reproduced version, we then train our algorithm, DHC-R. In this section, we provide a detailed description of the architecture of the original DHC algorithm, along with implementation details that were not explicitly covered in the original paper.

Fig. 2. Up: DHC architecture (image from “Distributed Heuristic Multi-Agent Pathfinding with Communication”). Down: DHC-R architecture.

DHC-R: Evaluating Distributed Heuristic Communication

157

The code that was used for evaluations can be found here [20]. 4.1 Architecture The original DHC algorithm has an architecture which consists of three main components (Fig. 2, above): 1. Observation encoder receives stacked matrixes that represent a single observation of an agent and outputs an embedding. 2. Communication block is an attention-based neural network that outputs an embedding that incorporates information from the two nearest neighbors, if any. Each agent sends the result of its observation encoder to every other agent inside its FOV. When messages are received by an agent, only those from the two nearest neighbors are selected for further processing. Then, the selected observations are passed through two communication blocks consisting of a multi-head attention and a GRU. 3. Q-network consists of two heads, the advantage and state-value, and combines their outputs to approximate the value of the Q-function. The final encoding is passed through the Q-network, which calculates the value V (s) and a vector of advantages {A(s, ai )}i for every action ai . These values are then summed to obtain a vector {Q(s, ai )}i . DHC-R. The robust version of DHC follows the same architecture as the original DHC but introduces randomness in the communication block. During training, there is a 30% chance of disabling the communication block, simulating communication failure scenarios. We propose this algorithm due to the anticipated poor performance of classical DHC in adversarial scenarios, particularly in cases involving communication loss. The architecture of DHC-R is depicted in Fig. 2 at the bottom.

4.2 Observations Observations are represented by a 6 × N × N tensor, with N × N being the FOV of the agent. The tensor uses six channels to encode the following information: 1. Binary mask encoding agents inside the FOV: a value of 1 represents the position of an agent, while 0 denotes an empty position. 2. Binary mask encoding obstacles inside the FOV: a value of 1 represents an obstacle, while 0 marks an empty cell. 3. Four heuristic channels: a representation of the potential direction an agent should take to get closer to its target. Each heuristic channel indicates whether an agent gets closer to its goal by taking a certain action. Refer to Fig. 3 for an example. Heuristic channels are necessary as the agents do not rely on imitation learning and need to acquire navigation skills. These channels provide agents with information about the location of the goal which helps them to navigate towards it.

158

V. Savinov and K. Yakovlev

Fig. 3. Heuristic channel; image from “Distributed Heuristic Multi-Agent Path Finding with Communication”.

4.3 Training In the original paper, the DHC algorithm was trained on maps with randomly generated obstacles. To train DHC-R, we utilize the same type of maps, which we will refer to as random maps. We generate them and corresponding scenarios using the GenerateRandomScen procedure, presented in Fig. 4 (algorithm 1). In comparison to the original implementation, we speed up the GenerateRandomScen procedure by up to 400 times for large map instances. It is guaranteed that each agent can reach its end point when no other agents are present when the described algorithm is applied.

Fig. 4. GenerateRandomScen algorithm.

However, authors do not validate that each end goal can be reached from the starting point when other agents are present, and we employ the same strategy. If the target

DHC-R: Evaluating Distributed Heuristic Communication

159

becomes unreachable for one of the agents, the experimental results of different algorithms would be equally impacted since the test set is generated beforehand. The loss function is a multi-step TD error: L(θ ) = Huber(Gt − Q(st , at , θ ), κ),   where Gt is the total return of an agent; Gt = rt + rt+1 γ + . . . + γ n Q st+n , at+n , θ , θ represents the parameters of a target network and Huber is the Huber loss. The authors employ curriculum learning to handle tasks with complex configurations on large maps with many agents. They start by training a single agent on a 10 × 10 map to learn how to navigate and reach goals in small-sized environments using generated trajectories. The policy is trained from a single-agent perspective, treating other agents as part of the environment. During training, 16 actors are set up on the CPU, and a single learner is set up on the GPU. The actors generate episodes with a copy of the global network, which are collected into a shared prioritized replay buffer. An ε-greedy exploration strategy is used during training, where each actor chooses a random action with a probability ranging 7i from 0.0008 to 0.4, with actor i having a probability of 0.41+ num_actors−1 . During inference, the action with the maximum Q-value is selected. To implement curriculum learning, a list of settings in the form of (map size, number of agents) is initialized with (10, 1). As actors interact with the environment, they store useful information obtained from the environment in the local buffer. Every 400 steps, the collected transitions are added to the global prioritized buffer. Every 10 seconds, statistics are calculated for each setting based on the information in the global buffer. If the success rate for a specific configuration exceeds 90%, the list of settings is extended with two additional configurations: one with the same number of agents but a larger map (10 cells added to each dimension), and the other with the same map size but one additional agent added. The training stops either when the maximum possible configuration is reached and successfully solved, which is determined by a CSR exceeding 90% of the last 200 runs, or when 600.000 training steps have been completed. We train each algorithm on a single A100 GPU, and the training takes around 3–4 full days.

5 Empirical Evaluation Adaptability to unfamiliar environments is a critical factor for running the algorithm in the real world. This section aims to assess the algorithm’s generalizability on non-randomly structured maps. For this study, we select multi-agent scenarios from the MovingAI collection [21], which includes maps with different structures and characteristics. Specifically, we choose four maps: lak303d.map, den520d, Berlin_1_256.map and ht_chantry.map (displayed in Fig. 5). The MovingAI website offers various scenarios for each map, where each scenario consists of a series of rows representing the start and goal positions for a single agent. Typically, these scenarios are used to evaluate singleagent performance. To generate test sets for multi-agent pathfinding problems with N agents, we select consecutive batches of N rows from each scenario, as illustrated in Fig. 6. During the experiments, we set the maximum episode length to 1024 steps. As

160

V. Savinov and K. Yakovlev

Fig. 5. MovingAI maps used for evaluation.

Fig. 6. Creating a multi-agent scenario using the.scen file from MovingAI.

shown in Fig. 7, the value of ISR consistently exceeded 90%. However, while most agents individually achieved their goals, the collective metric on this set of experiments was significantly worse (Fig. 7, left) than on random maps (results for random maps were reported in the original DHC [2] paper and were replicated by us prior to the start of OOD testing). These results suggest that communication can sometimes mislead an agent and negatively affect cooperation. In real-world scenarios, however, perfect communication between agents cannot always be guaranteed. Weather conditions, deceptive agents in the environment, or the malfunction of an agent can result in the loss of data packets. As communication is the primary tool for cooperation in this algorithm, the loss of packets can potentially impact the performance of the system. To analyze the impact of packet loss on the performance of the algorithm, we evaluate two different algorithms, DHC-R and DHC-paper, on randomly generated maps with varying obstacle densities, where message passing was not allowed at all. We generate a set of 200 tests on an M × M random map with density p using the procedure GenerateRandomScen explained in detail in Fig. 4. In each test, agents act in parallel by choosing one of five allowed actions: move left, right, up, down, or stay. Upon reaching the goal, an agent remains stationary and does not receive a new target. Note that an agent can still communicate with other agents to coordinate movements, such as stepping aside to allow another agent to pass. Execution stops either when all agents have reached their respective targets or upon reaching a terminal episode, which is set to 256. We report an average metric across all 200 runs. Charts depicting the collective success

DHC-R: Evaluating Distributed Heuristic Communication

161

Fig. 7. DHC. Performance on MovingAI maps. Left: CSR; right: ISR.

rates for each agent in scenarios with obstacle densities of 0.1 are presented in Fig. 7 (right). As evident from the results, DHC-R demonstrates significant improvement over DHC on small random maps with a density of 0.1, as observed by the 20% (absolute) increase in performance (Fig. 8).

Fig. 8. DHC-R vs. DHC-paper, CSR, random square maps (10, 20, 30, 40). Density 0.1.

6 Conclusion In this paper, we have studied the problem of partially observable multi-agent pathfinding and extensively analyzed the existing DHC [2] algorithm in various scenarios. The paper’s original experimental setup was replicated and extended, and novel experiments on MovingAI maps and custom maps were conducted to test agent performance. Furthermore, this study introduced a new experiment testing the performance of the algorithm under complete packet loss, revealing a significant drop in performance. We suggested a modification to the algorithm, named DHC-R, to mitigate these issues and improve the algorithm’s performance under these conditions and open sourced our implementation (including model weights).

162

V. Savinov and K. Yakovlev

Acknowledgement. The authors gratefully acknowledge the Mathematics and Computer Science Department of St. Petersburg State University, with special thanks to Aleksandr Avdiushenko for providing access to computational resources.

References 1. Partially-Observable Grid Environment for Multiple Agents: https://github.com/AIRI-Instit ute/pogema. Last accessed 11 May 2023 2. Ma, Z., Luo, Y., Ma, H.: Distributed heuristic multi-agent path finding with communication. In: 2021 International Conference on Robotics and Automation (ICRA), pp. 8699–8705. IEEE, Xi’an, China (2021) 3. Sutton, R.S., Barto, A.G.: Reinforcement Learning: An Introduction. MIT Press, Cambridge, MA, USA (1998) 4. Zelinsky, A.: A mobile robot exploration algorithm. IEEE Trans. Robot. Autom. 8(6), 707–717 (1992) 5. Hart, P.E., Nilsson, N.J., Raphael, B.: A formal basis for the heuristic determination of minimum cost paths. IEEE Trans. Syst. Sci. Cybern. 4(2), 100–107 (1968) 6. Wagner, G., Choset, H.: M*: A complete multirobot path planning algorithm with performance bounds. In: 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 3260–3267. IEEE, San Francisco, CA, USA (2011) 7. Ferner, C., Wagner, G., Choset, H.: ODrM* optimal multirobot path planning in low dimensional search spaces. In: 2013 IEEE International Conference on Robotics and Automation, pp. 3854–3859. IEEE, Karlsruhe, Germany (2013) 8. Sharon, G., Stern, R., Felner, A., Sturtevant, N.R.: Conflict-based search for optimal multiagent pathfinding. Artif. Intell. 219, 40–66 (2015) 9. Boyarski. E.: Iterative-deepening conflict-based search. In: Proceedings of the TwentyNinth International Joint Conference on Artificial Intelligence (IJCAI-20), pp. 4084–4090. International Joint Conferences on Artificial Intelligence Organization, Yokohama (2020) 10. Barer, M., Sharon, G., Stern, R., Felner, A.: Suboptimal variants of the conflict-based search algorithm for the multi-agent pathfinding problem. Front. Artif. Intell. Appl. 263, 961–962 (2014) 11. Andreychuk, A., Yakovlev, K., Surynek, P., Atzmon, D., Stern, R.: Multi-agent pathfinding with continuous time. Artif. Intell. 305, 103 (2022) 12. Surynek, P., Felner, A., Stern, R., Boyarski, E.: Efficient SAT approach to multi-agent path finding under the sum of costs objective. In: Proceedings of the Twenty-Second European Conference on Artificial Intelligence, pp. 810–818. IOS Press, The Hague, The Netherlands (2016) 13. Lam, E., Le Bodic, P., Harabor, D.D., Stuckey, P.J.: Branch-and-cut-and-price for multi-agent pathfinding. In: Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence (IJCAI-19), pp. 1289–1296. International Joint Conferences on Artificial Intelligence Organization, Macao, China (2019) 14. Yakovlev, K.S., Andreychuk, A.A., Skrynnik, A.A., Panov, A.I.: Planning and learning in multi-agent path finding. Dokl. Math. 106(1), 79–84 (2022) 15. Sunehag, P.: Value-decomposition networks for cooperative multi-agent learning. In: Proceedings of the 17th International Conference on Autonomous Agents and Multi-Agent Systems, pp. 2085–2087. International Foundation for Autonomous Agents and Multiagent Systems, Stockholm, Sweden (2018) 16. Sartoretti, G.: PRIMAL: pathfinding via reinforcement and imitation multi-agent learning. IEEE Robot. Autom. Lett. 4(3), 2378–2385 (2019)

DHC-R: Evaluating Distributed Heuristic Communication

163

17. Li, W., Chen, H., Jin, B., Tan, W., Zha, H., Wang, X.: Multi-Agent Path Finding with Prioritized Communication Learning. In: 2022 International Conference on Robotics and Automation (ICRA), pp. 10695–10701. Philadelphia, PA, USA (2022) 18. Li, Q., Gama, F., Ribeiro, A., Prorok, A.: Graph Neural Networks for Decentralized Path Planning. In: Proceedings of the 19th International Conference on Autonomous Agents and Multi-Agent Systems, pp. 1901–1903. International Foundation for Autonomous Agents and Multiagent Systems, Auckland, New Zealand (2020) 19. Ma, Z., Luo, Y., Pan, J.: Learning selective communication for multi-agent path finding. IEEE Robot. Autom. Lett. 7(2), 1455–1462 (2022) 20. Learnable Decentralized MAPF using reinforcement learning with local communication. https://github.com/acforvs/dhc-robust-mapf. Last accessed 11 May 2023 21. Stern, R., et al., Multi-agent pathfinding: definitions, variants, and benchmarks. In: Symposium on Combinatorial Search (SoCS), pp. 151–158 (2019)

Ground Mobile Robot Localization Algorithm Based on Semantic Information from the Urban Environment Artur Podtikhov(B)

and Anton Saveliev

St. Petersburg Federal Research Center of the Russian Academy of Sciences (SPC RAS), 39, 14th Line, St. Petersburg 199178, Russia [email protected] Abstract. This paper presents the SLAM algorithm, which use the semantic information extracted from the urban environment to increase the accuracy of ego-vehicle localization in ORB-SLAM2 system. For this purpose, a semantic segmentation module is added to the standard algorithm to assign an object on each frame to one of a given set of classes. The CARLA Simulator was used as a simulation environment, which generates a photorealistic urban environment with the ability to run an arbitrary number of active elements in it, which usually make localization difficult, causing interference with the system. Based on the environment, a training dataset for semantic segmentation was collected. The training dataset consists of 3,696 pairs of city images and corresponding segmentation masks in which each pixel corresponds to one of 23 semantic labels. Using this dataset, the DeepLabV3+ segmentation model was trained with mean per-class IoU metric equals to 81.48%. By using semantic information to filter potentially dynamic objects and matching key points, we were able to increase the localization accuracy relative to the base algorithm by an average of 23% and build a semantic map of the environment. Keywords: SLAM · ORB-SLAM2 · DEEPLAB · CARLA · Robot

1 Introduction Simultaneous Localization and Mapping (SLAM) is a technique used in mobile autonomous vehicles to build a map of an unknown environment or to update a map of a known environment while simultaneously keeping track of agent’s location and the traveled path within it. In general terms, the control scheme of a modern mobile robot moving in a known environment can be represented in the following chain of actions: obtaining information about the world around; determining one’s own position on a predetermined map; traffic planning with regard to the environment; control over the implementation of planned actions and transmission of control signals to actuators (motors, wheels and other manipulators). However, if the environment is not known in advance, then first you need to build a map of the area. Traditional mapping algorithms require an estimate of the robot’s position, while accurate localization requires a previously known map. That is why SLAM methods can be called complex, because they are aimed at solving two mutually dependent tasks: localization and map construction. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 164–174, 2023. https://doi.org/10.1007/978-3-031-43111-1_15

Ground Mobile Robot Localization Algorithm

165

This approach was first proposed at the IEEE Conference on Robotics and Automation in San Francisco in 1985 [1]. Then, and over the next few years, it was solved using various active sensors, such as a laser range finder, lidar or sonar, to determine the position of landmarks in space. SLAM is a cornerstone for autonomous navigation tasks in unknown environment, its applications are found in unmanned vehicles [2, 3], aircraft [4], underwater vehicles [5], virtual reality [6], in space exploration, for example, the surface map of Mars was constructed using SLAM methods [7]. The relevance of solving the problem of simultaneous localization and mapping is due to the fact that maps commonly used for agent navigation mainly reflect the type of space fixed at the time of their construction, and it is not at all necessary that the type of space will be the same at the time the maps are used. At the same time, the complexity of the technical process of determining the current location with the simultaneous construction of an accurate map is due to the low accuracy of the instruments involved in the process of calculating the current location. Recently, visual SLAM methods, which are based on information from cameras, have become very popular, since cameras are cheaper to purchase and operate, while they can provide more information about the world around the robot. For instance, only cameras can transmit color, therefore, in unmanned vehicles they are used. Although the use of cameras increases the complexity and resource intensity of the algorithms, since they do not allow you to directly calculate the distance to the object of interest. Most visual SLAM methods rely solely on geometric information, building a map of the unknown terrain in dense/semi-dense (DTAM [8], LSD-SLAM [9]) or keypointbased (PTAM [10], ORB-SLAM [11]) point clouds. Such maps are homogeneous: the dots on them indicate only the presence or absence of an obstacle and do not carry any additional information. At the same time, visual SLAM works with camera images – a rich source of additional information. Often, when working with images, not the points themselves are used, but the objects they form, for example, various algorithms for analyzing biomedical images are based on this approach. Such enlarged objects are usually obtained using object detection methods, and if more accurate prediction of the boundaries of objects is necessary, using segmentation methods. Simultaneous localization and mapping methods that use this approach to working with images from cameras are grouped under the name Semantic SLAM. All semantic SLAM methods can be divided into 2 broad categories according to the type of problem being solved: improving map representation [12–17] and improving localization [18–24]. The purpose of using methods that improve the presentation of a map is to add an additional “semantic” layer to it, so that points on it are distinguishable from each other and belong to a certain class. Such map representations can be useful in navigation, often in articles the following example is given: semantic information provides an ideal level of abstraction for a robot to understand and execute human commands (e.g., “bring me a cup of coffee”, “leave the house through the red door”) and provide people with models of the environment that are easy to understand. In turn, methods aimed at improving localization consider segmentation not as a goal, but as a tool that helps to take into account additional non-geometric information during localization. Such methods, for instance, include filtering moving objects and localization or mapping

166

A. Podtikhov and A. Saveliev

solely on the basis of those objects that a priori cannot change their location in the world, thus helping the robot to localize in the so-called “dynamic” environments. This paper presents a new method of semantic SLAM, which uses one of the most stable and accurate algorithms ORB-SLAM2 [25] as a basic algorithm for localization and map building. But it also considers semantic information using the DeepLabV3+ [26] model for semantic segmentation in order to: (a) build meaningful maps, where each point is associated with the class of the object to which it belongs, and (b) use semantic information to increase localization accuracy (by excluding potentially dynamic scene objects and building associations between points from different frames).

2 Description of the Training Data Collection Methodology for Semantic Segmentation To collect a dataset for training the segmentation network, a high-quality map “Town10HD” from the CARLA Simulator [15] was used, which is an urban area with various infrastructure facilities. On this map, software developers pre-set a list of locations in which it is recommended to spawn cars in order for them to appear on the road directed towards traffic (Fig. 1).

Fig. 1. Schematic image of the Town10HD city map from the CARLA Simulator. The points where the training images from the camera were collected are marked in orange (Color figure online).

The recommended points for car spawn were used to collect a dataset of camera images and corresponding ground-truth segmentation masks according to the following algorithm: 1. The car was spawned at a given point (X i , Y i , Z i ), parallel to the ground surface with a rotation angle relative to the perpendicular to the surface equal to 0.

Ground Mobile Robot Localization Algorithm

167

2. On the hood of the car, 2 pseudo-cameras were spawned: standard RGB and segmentation, both with a resolution of 800 × 600 pixels. 3. The car (with the cameras) turned through an angle of 15°. 4. The images received from the cameras were recorded and saved to disk. 5. Steps 3–4 were repeated until the car made a complete turn. 6. The car and both cameras were destroyed. 7. The transition to the next spawn point and, respectively, to point 1 was performed. Thus, 3,696 pairs of images with segmentation masks were collected from the 154 recommended vehicle spawn points. Figure 2 shows an example of an image obtained with an RGB camera mounted on a car hood (left) and its corresponding ground-truth segmentation mask (right), in which each pixel belongs to one of the given classes. In total, CARLA provides a segmentation map for 23 classes, which are listed in Table 1. A random subset of images of 80% of the original data set was used for training, with the remaining 20% exclusively for validating the results.

Fig. 2. An example of an RGB camera image (left) and a ground-truth semantic segmentation mask (left) from the training dataset.

3 Segmentation Model DeepLabV3+ was used as the segmentation model, with the resnext50_32x4d encoder [27] pre-trained on the ImageNet dataset [28]. A small number of augmentations were used: random cropping the image to a size of 512 × 512 pixels, horizontal flipping (with probability 0.5), adding normally distributed noise (with probability 0.2), and performing a random four-point perspective (with probability 0.5). The loss function chosen was FocalLoss [29] since the class distribution in the dataset is highly irregular. Optimization was performed using AdamW optimizer [30]. The training batch size was set to 6 and the learning rate was set to 1e-4. The table shows that for large objects the segmentation accuracy is quite high, while for objects with a small area (such as traffic lights, road signs and poles) it is less. However, the obtained distribution of accuracy for different classes is consistent with the distribution of accuracy of the best segmentation models of the CityScapes benchmark, so this distribution can be associated with the limitations of modern semantic segmentation architectures.

168

A. Podtikhov and A. Saveliev

Table 1. Metrics reflecting the quality of segmentation on the validation dataset. The MISSING label marks classes that do not exist in the Town10HD map. ID

Class label

Per-class IoU

Per-class Accuracy

ID

Class label

Per-class IoU

Per-class Accuracy

0

Unlabeled

SKIP

SKIP

12

TrafficSign

69.65%

76.47%

1

Building

95.05%

97.66%

13

Sky

93.70%

96.20%

2

Fence

32.83%

42.83%

14

Ground

94.17%

97.62%

3

Other

84.67%

89.02%

15

Bridge

MISSING

MISSING

4

Pedestrian

MISSING

MISSING

16

RailTrack

98.10%

98.97%

5

Pole

58.49%

66.19%

17

GuardRail

MISSING

MISSING

6

RoadLine

84.98%

89.98%

18

TrafficLight

78.44%

87.59%

7

Road

98.65%

99.47%

19

Static

81.89%

90.01%

8

SideWalk

97.01%

98.43%

20

Dynamic

79.31%

89.68%

9

Vegetation

84.76%

93.74%

21

Water

63.54%

74.03%

10

Vehicles

90.62%

95.70%

22

Terrain

79.56%

85.09%

11

Wall

82.62%

87.96%

4 The Algorithm Developed As previously mentioned, the ORB-SLAM2 algorithm was chosen as the base algorithm for simultaneous localization and mapping. Interaction with the CARLA simulation environment was performed using the ros_bridge package which allows to receive sensor and odometry information from the simulator and publish them to ROS topics. Figure 3 shows a generalized architecture of the proposed algorithm. The architecture is almost the same as that of ORB-SLAM2, except for the new block responsible for semantic segmentation included in the Tracking thread. The rest of the changes are internal and adjust some functions, which will be described below. Semantic Segmentation Block. In order to integrate the image segmentation model into the system, it was converted from the PyTorch format to TorchScript, after which it became possible to use it in scripts written in C++. The resulting model is initialized by the GPU in the Tracking module and applied after each new frame is received, thus, at the start of the algorithm, there is not only the image itself, but also a segmentation mask that matches each pixel of the image with a semantic class. Extract ORB Block. Since storing a full segmentation mask for each frame requires a significant amount of RAM, the corresponding semantic information is stored only for selected key points. For this purpose, at the moment of extracting key points and ORB descriptors, semantic information is added to these key points, indicating that the point belongs to one of the 23 classes, after which the rest of the mask is removed. New Points Creation Block. When a map point is created, semantic information is also added to it, with each point storing a list of all predicted semantic classes, when it is

Ground Mobile Robot Localization Algorithm

169

seen from different angles, at the current moment its class is the class it takes most often. This approach reduces the segmentation error and eliminates outliers. Also, a map point is considered inactive and does not participate in further calculations if it belongs to one of the potentially dynamic (or low-informative) classes (Unlabeled, Other, Pedestrians, Vehicles, Sky, Dynamic).

Fig. 3. Generalized architecture of ORB-SLAM2, to which a semantic segmentation block has been added, which is triggered on receipt of each new frame.

Key Point Association. The keypoint (or map points and keypoints) association algorithm is one of the central algorithms of ORB-SLAM2, since it is used in almost all submodules. In the original algorithm, the association is performed solely based on the calculation of the distance between the two ORB descriptors. To account for semantic information, a penalty factor equal to 0.5 * current distance is added to this distance.

170

A. Podtikhov and A. Saveliev

5 Results Analysis To evaluate the quality of localization, 3 experiments were conducted in a simulation environment, lasting from 1 to 3 min, sensor information and ground-truth odometry were stored at a frequency of 20 frames per second. Simulations were run from various recommended vehicle spawn points, in addition, except ego-vehicle, 20 cars with a built-in autopilot and 10 pedestrians were also generated in the environment. After that, ORB-SLAM2 model and the developed modification were launched separately. Having ground-truth and predicted odometry, it is impossible to compare them directly, because when using a monocular camera, it is possible to restore the world coordinates of a map point with an accuracy to scale constant. The Horn algorithm [31] was used to estimate the scale constant and alignment of coordinate systems. For each simulation, it shows the ground-truth trajectory of the car (blue), the trajectory obtained using the ORB-SLAM2 algorithm (orange) and the trajectory obtained using the developed algorithm (green). From the motion trajectories it is difficult to draw conclusions about the increase in localization accuracy, therefore, plots of localization errors are also attached (Fig. 4). The reconstructed trajectories for all simulations are shown in Fig. 5. The horizontal axis denotes the frame number, the vertical axis denotes the distance between the groundtruth position of the vehicle at a given time and the predicted position. Comparison of localization accuracy over the entire route was performed using the metric of the mean percentage absolute error in the Cartesian coordinate system, calculated by the formula:     100%  xi − xˆ i , [yi − yˆ i ] . (1) MAPE([x, y], [ˆx + yˆ ]) = MEAN nsamples [|xi |, |xi |] The results of comparing the quality of localization are in Table 2. It can be seen that the proposed algorithm performs slightly better than the basic algorithm in determining the location of the vehicle, while from the reconstructed trajectories (Fig. 5) it can be concluded that the predictions change slightly, mainly due to reducing the probability of wrong key points matching. At the same time, if it is strictly forbidden to assign points corresponding to different semantic classes, it fails to initialize the map (due to the fact that the number of matches falls below the threshold value), therefore, to further improve the approach, it is necessary to improve the quality of semantic segmentation. A side effect of our work is the construction of a semantic map of the environment; after a complete route around of the city, the map of the area looks like Fig. 6. It can be concluded that the algorithm for determining key points basically extracts points from buildings (white color on the map), road markings (purple) and trees (green).

Ground Mobile Robot Localization Algorithm

171

Fig. 4. Plots for estimating position errors for each frame (in meters). Green colors indicate the errors of the implemented algorithm, orange is ORB-SLAM2 (Color figure online).

Table 2. Localization accuracy in different simulations. The table shows the duration of the route, it’s length (in simulator units) and the mean absolute percentage error of estimating the ego vehicle location by the base and developed implementation. The last column displays the % change in the localization error of the developed algorithm relative to the base one. No. Duration (s) Length (m) ORB-SLAM2 Developed Algorithm Relative change in localization error 1

172.8

610

25.85%

21.7%

2

83.2

241.7

1.10%

1.02%

3

105.6

456.36

7.83%

4.22%

–16.05% –7.27% –46.1%

172

A. Podtikhov and A. Saveliev

Fig. 5. The trajectory of the car. The ground-truth trajectory is shown in blue, the trajectory predicted by the ORB-SLAM2 algorithm in orange, and the trajectory predicted by developed algorithm in green. The order of the images corresponds to the sequence numbers of the simulations in Table 2.

Fig. 6. Semantic map of the city Town10HD in the form of a point cloud, obtained after simulating the movement of a car throughout the city. Different colors indicate urban infrastructure objects belonging to different classes (Color figure online).

6 Conclusion In this paper, we proposed an algorithm for simultaneous localization and mapping, taking into account semantic information about the objects of the urban environment. The proposed approach excludes potentially dynamic objects from the consideration of the algorithm and improves the matching of key points. The developed algorithm has demonstrated a 23% increase in localization accuracy on mean absolute percentage error relative to the basic algorithm, at the same time, it requires more computing resources to apply the segmentation model. The quality of the current segmentation model does not allow to completely eliminate the comparison of key points assigned to different semantic classes, so further development of the algorithm should be aimed at increasing the segmentation accuracy and increasing the number of semantic classes under consideration.

Ground Mobile Robot Localization Algorithm

173

References 1. Chatila, R., Laumond, J.: Position referencing and consistent world modeling formobile robots. In: Proceedings. International Conference on Robotics and Automation, vol. 2, pp. 138–145. IEEE (1985) 2. Henning, L., Andreas, G., Bernd, K.: Visual slam for autonomous ground vehicles. In: IEEE International Conference on Robotics and Automation, Shanghai, pp. 1732–1737, China (2011) 3. Qin, T., Chen, T., Chen, Y., Su, Q.: Avp-slam: semantic visual mapping and localization for autonomous vehicles in the parking lot. In: 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 5939– 5945. IEEE (2020) 4. Milford, M.J., Schill, F., Corke, P., Mahony, R., Wyeth, G.: Aerial slam with a single camera using visual expectation. In: 2011 IEEE international conference on robotics and automation, pp. 2506–2512. IEEE (2011) 5. Ribas, D., Ridao, P., Tardo´s, J.D., Neira, J.: Underwater slam in man-made structured environments. J. Field Robot. 25(11–12), 898–921 (2008) 6. Jinyu, L., Bangbang, Y., Danpeng, C., Nan, W., Guofeng, Z., Hujun, B.: Survey and evaluation of monocular visual-inertial slam algorithms for augmented reality. Virtual Reality Intell. Hardware 1(4), 386–410 (2019) 7. Zheng, B., Zhang, Z.: An improved EKF-SLAM for mars surface exploration. Int. J. Aerosp. Eng. 2019, 1–9 (2019) 8. Newcombe, R.A., Lovegrove, S.J., Davison, A.J.: Dtam: Dense tracking and mapping in real-time. In: 2011 International Conference on Computer Vision, pp. 2320–2327. IEEE (2011) 9. Engel, J., Scho¨ps, T., Cremers, D.: Lsd-slam: Large-scale direct monocular slam. In: Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 Sep 2014, Proceedings, Part II 13, pp. 834–849. Springer (2014) 10. Klein, G., Murray, D.: Parallel tracking and mapping for small AR workspaces. In: 2007 6th IEEE and ACM International Symposium on Mixed and Augmented Reality, pp. 225–234. IEEE (2007) 11. Mur-Artal, R., Montiel, J.M.M., Tardos, J.D.: Orb-slam: a versatile and accurate monocular slam system. IEEE Trans. Rob. 31(5), 1147–1163 (2015) 12. Dub´e, R., Cramariuc, A., Dugas, D., Nieto, J., Siegwart, R., Cadena, C.: Segmap: 3d segment mapping using data-driven descriptors, arXiv preprint arXiv:1804.09557 (2018) 13. Hermans, A., Floros, G., Leibe, B.: Dense 3d semantic mapping of indoor scenes from rgbd images. In: 2014 IEEE International Conference on Robotics and Automation (ICRA), pp. 2631–2638. IEEE (2014) 14. McCormac, J., Handa, A., Davison, A., Leutenegger, S.: Semantic fusion: dense 3dsemantic mapping with convolutional neural networks. In: 2017 IEEE International Conference on Robotics and automation (ICRA), pp. 4628–4635. IEEE (2017) 15. Rosinol, A., Abate, M., Chang, Y., Carlone, L.: Kimera: an open-source library for realtime metric-semantic localization and mapping. In: 2020 IEEE International Conference on Robotics and Automation (ICRA), pp. 1689–1696. IEEE (2020) 16. Salas-Moreno, R.F., Newcombe, R.A., Strasdat, H., Kelly, P.H., Davison, A.J.: Slam++: simultaneous localisation and mapping at the level of objects. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1352–1359 (2013) 17. Stückler, J., Behnke, S.: Multi-resolution surfel maps for efficient dense 3d modeling and tracking. J. Vis. Commun. Image Represent. 25(1), 137–147 (2014) 18. Bowman, S.L., Atanasov, N., Daniilidis, K., Pappas, G.J.: Probabilistic data association for semantic slam. In: International Conference on Robotics and Automation (ICRA), pp. 1722– 1729. IEEE (2017)

174

A. Podtikhov and A. Saveliev

19. Fuentes, O., Savage, J., Contreras, L.: A SLAM system based on Hidden Markov Models. Inform. Autom. 21(1), 181–212 (2022). https://doi.org/10.15622/ia.2022.21.7 20. Mahamudul Hashan, A., Md Rakib Ul Islam, R., Avinash, K.: Apple leaf disease classification using image dataset: a multilayer convolutional neural network approach. Inform. Autom. 21(4), 710–728 (2022). https://doi.org/10.15622/ia.21.4.3 21. Ganti, P., Waslander, S.L.: Network uncertainty informed semantic feature selection for visual slam. In: 2019 16th Conference on Computer and Robot Vision (CRV), pp. 121–128. IEEE (2019) 22. Gawel, A., Del Don, C., Siegwart, R., Nieto, J., Cadena, C.: X-view: graph-basedsemantic multi-view localization. IEEE Robot. Autom. Lett. 3(3), 1687–1694 (2018) 23. Lianos, K.N., Schonberger, J.L., Pollefeys, M., Sattler, T.: Vso: Visual semantic odometry. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 234–250 (2018) 24. Stenborg, E., Toft, C., Hammarstrand, L.: Long-term visual localization using semantically segmented images. In: 2018 IEEE International Conference on Robotics and Automation (ICRA), pp. 6484–6490. IEEE (2018) 25. Mur-Artal, R., Tardo´s, J.D.: ORB-SLAM2: an open-source SLAM system for monocular, stereo, and RGB-D cameras. IEEE Trans. Robot. 33(5), 1255–1262 (2017) 26. Chen, L.C., Zhu, Y., Papandreou, G., Schroff, F., Adam, H.: Encoder-decoder with atrous separable convolution for semantic image segmentation. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 801–818 (2018) 27. Xie, S., Girshick, R., Doll´ar, P., Tu, Z., He, K.: Aggregated residual transformations for deep neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1492–1500 (2017) 28. Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: Imagenet: a large-scale hierarchical image database. In: Conference on Computer Vision and Pattern Recognition, pp. 248–255. IEEE (2009) 29. Lin, T.Y., Goyal, P., Girshick, R., He, K., Dolla´r, P.: Focal loss for dense object detection. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2980–2988 (2017) 30. Reddi, S.J., Kale, S., Kumar, S.: On the convergence of a dam and beyond, arXivpreprint arXiv:1904.09237 (2019) 31. Horn, B.K.: Closed-form solution of absolute orientation using unit quaternions. Josa a 4(4), 629–642 (1987)

Remote Control Robotic System for the Perimeter Security Azad Bayramov1(B)

and Samir Suleymanov2

1 Institute of Control Systems, 68, B. Vahabzadeh street, Baku AZ1141, Azerbaijan

[email protected] 2 Institute of Geology and Geophysics, 119, H. Javid avenue, Baku AZ1073, Azerbaijan

Abstract. The the results of the development, preparation and preliminary laboratory and field tests of the robotic security system along the perimeter of any object remotely controlled by radio waves have been presented in paper. Radio signals are coded using a cryptographic method. The robotic system consists of a remote control and controlled units (blocks) that can perform various functions to protect the perimeter of the selected zone. The principle of operation and the structure of the robotic security system are presented. To ensure safety, the blocks are controlled by coded signals. The working principle, management, operation rules, management and telemetry of the security system were explained. Cryptographic coding based on the Neyman algorithm was used to encrypt the telemetry of the security system. The robotic security system can be used for the protection of territory, borders or objects, including for military purposes. Such control can be carried out around the clock, in any weather, at any area and distance. The working principle and management of the protection system are given in the paper. Keywords: Security System · Robotic · Remotely Controlled · Coded Signals · Control Block · Sensor Block

1 Introduction Effective control of unauthorized access to any territory, or crossing the borders, or the perimeter of any object is one of the urgent problems in industry, in the military field, in the border service and many others. The most effective method is an automated control, which is carried out remotely. Such control can be carried out around the clock, in any weather, at any area and distance. Here the human factor plays a minimal role, so the probability of error is minimized here. A number of the scientific works (for example, [1–4]) is devoted this problem. In particular, in the work [3] it was analyzed a perimeter defense problem in which a single turret, having a finite range and service time, is tasked to defend a perimeter against at most N intruders that arrive in the environment. An offline as well as an online version of this setup was considered. In the offline setup in which N intruders have already arrived in the environment, it was established that the problem is equivalent to solving a Travelling Repairperson Problem with Time Windows. This paper presents the developed unified automated robotic system for remote control of a given territory and access protection. Seismic sensors, acoustic sensors, fuses, © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 175–184, 2023. https://doi.org/10.1007/978-3-031-43111-1_16

176

A. Bayramov and S. Suleymanov

video observation blocks, etc. can serve as remote sensor units that are activated by coded radio signals. The advantage of this system is that it is controlled by coded signals, the frequency and control protocol of which are set by the developers, so this system is safe and not available to a third party.

2 Principle of Operation of the Security System The common functional block diagram of the system is shown in Fig. 1. Here: RC – Remote Control, SB – sensors blocks, 1 – device programmer (operating mode selection block), 2 – signal encryption block, 3 – antennas and feeders, 4 – signal decoding block, 5 – start relay, 6 – functional sensors. The operating modes of the secutity system are selected through the programmer located in the control panel. These modes are activation of one or all sensor blocks and their sequence. The sequence of: fc (S, t) = S1 (t1 ) + S2 (t2 ) + S3 (t3 ) + S4 (t4 ) + ... + Sn (tn ),

(1)

signals are generated in this block and is passed to the encryption block and encrypted there.

Fig. 1. The common functional block diagram of a remotely controlled radio system.

Encryption occurs due to the change of the time sequence of signals. This sequence can be represented as a function as follows:  fk (Sk , tk ). fc (S, t) = S4 (t4 ) + S1 (t1 ) + S3 (t3 ) + S2 (t2 ) + . . . = The signals through the feeder antenna, are spread into space and received by the antennas of the sensor blocks located at a certain distance. Electromagnetic signals in sensor blocks are decoded through selection:  fk (Sk , tk ) = S1 (t1 ) + S2 (t2 ) + S3 (t3 ) + S4 (t4 ) + . . . + Sn (tn ). (2) fdc (S, t) =

Remote Control Robotic System for the Perimeter Security

177

If the f dc (S,t) function does not coincide with the f c (S,t) function, fdc (S, t) = fc (S, t), that is, the time sequence is violated, then the sensor block is activated. If the f dc (S,t) function coincide with the f c (S,t) function, fdc (S, t) = fc (S, t), then, then the received signals affect the electric relay, the sensor block is activated through the electric relay. After that, the sensor block records the variouse change in the environment at a certain distance. The encryption code is based on Neyman’s algorithm [5, 6].

3 Choosing the Encryption Algorithm of the Security System In order to choose the encryption algorithm to be used in the created cryptosystem, first of all, it is necessary to pay attention to the following features of the algorithms [5, 6]: • Crypto durability. The algorithm must be carefully analyzed by the world cryptographic community for a long time (at least five years) and be considered cryptographically robust against various types of attacks; • The length of the key. The key used in the encryption algorithm must be at least 256 bits for symmetric encryption algorithms and at least 2048 bits for open key algorithms. This is intended to make it impossible to open the code directly by brute force (strong force) in the 21st century; • Encryption speed. Therefore, the data encryption speed for the selected algorithm should be very high to avoid interruptions in data transfer. • Resourcefulness. The algorithm must be optimized for implementation in hardware implementation. The amount of RAM and the performance of the microprocessor needed should be within the limits of general purpose microcontrollers. • The system encryption algorithm was selected from the following group of algorithms: Des, Aes, Cast, Rc4, Verman. All of the above algorithms meet the requirements of different levels for them. For the implementation of the algorithm, operations on large numbers (increase in degree) are needed, for the fast execution of which it is necessary to use special microcircuits in the project. Public key algorithms cannot be used due to their low speed. It is best not to use stream ciphers. They are useful for encrypting the flow of information, for example, information packages in computer networks, or conversations on a telephone line. Also, the “disposable notepad scheme” is popular in Vernam’s cipher cryptography. The solution is the symmetric encryption system, which was invented in 1917 by AT&T employees Major Joseph Moborn and Vernam Gilbert. Vernam’s cipher is the most secure available cryptosystem. Taking these into account, it was accepted to use the Vernam encryption algorithm in radio control of security system.

178

A. Bayramov and S. Suleymanov

Now let’s give detailed information about the encryption process and the password itself. Since this cipher was invented for computer systems, it should be noted that it is based on binary arithmetic. The radio security robotic system consists of 10 radio block detectors and 1 control panel. After the radio block detectors are activated, the keys for each radio block detectors are selected from the random number machine block of the control panel, sent to the block detectors and written to the control panel memory accordingly. Random keys are 32 bytes long. During operation, the command sent by the remote control to the selected radio block detectors is encrypted with the key stored in the memory for that radio block detectors. Vernam encryption is based on the absolute summation logical operation of addition by module 2. In Bull’s algebra, the absolute addition operation 2 is a function of two variables (the operands of the operations or the arguments of the function). The summation of the result is calculated according to the truth Table (see Table 1). Table 1. Truth table {\displaystyle true > false}. a

b

ab

0

0

0

0

1

1

1

0

1

1

1

0

10010111 11101001 xxxxxxxx 10010111 (the start-command for the selected radio block detectors). 01100110 10111100 xxxxxxxx 00011010 (the key for selected radio block detectors). 11110001 01010101 xxxxxxxx 10001101 (the result of a logical operation between an address and a key). 11110001 01010101 xxxxxxxx 10001101 (adopted command). 01100110 10111100 xxxxxxxx 00011010 (the key for the selected radio block detector). 10010111 11101001 xxxxxxxx 10010111 (the result of a logical operation between the received command and the key, or the command to start for the selected radio block detectors). Thus, the encryption of signals prevents the security system from being activated by the outside on purpose or by accident.

4 Structure of the Security System The main tasks of the security system:

Remote Control Robotic System for the Perimeter Security

179

– activation of sensor blocks through an electric relay in the front line and areas close to it, in the radius of visual distances; – control of 10 sensor blocks with one controller. The security system is managed by one operator. The security system is made according to the modular principle and consists of a control unit (block) (remote control), 10 sensor units (blocks), antennas and special narrow iron plates for burying in the ground. The tactical and technical indicators of the security system are given in Table 2. Working conditions, ambient temperature: –20° ÷ +50o . The security system complies with the IP67 protection class against dust and moisture. This means that the equipment is completely protected from dust and solid bodies with dimensions of at least 1.0 mm; has protection against water ingress into the shell when immersed at a certain depth and time, and is regulated by IEC 60529 (DIN 40050, GOST 14254) standards [7]. The sensor blocks are controlled by coded one-way radio signals during operation mode and two-way radio signals during control mode. The coding scheme used completely excludes arbitrary (without operator permission) activation. Table 2. Tactical and technical indicators of the security system. No.

Indicators

Quantification of the indicators

1

Control radius (with direct view)

2 km

2

Standby time

> 6 months

3

Duration of operation in activation mode

72 h

4

Activation period

3s

5

Weight of device (with case)

5 kg

6

Mass of 1 sensor block

280 gr

7

The mass of the control panel

500 gr

8

Overall dimensions of the device

45 × 50 × 20 cm

9

Power supply

3,7V Li – battery

10

Power of the radio transmitter

1W

11

Retention period: in stock field conditions -

10 years 2 years

Assignment of commands coming from the control unit (block): – putting the security system into working condition; – disable the security system; – activate the sensor blocks or neutralize (deactivate) the sensor blocks in the active state;

180

A. Bayramov and S. Suleymanov

– check the current state of each sensor block individually. Remote control of the sensor blocks with a remote control should take place either in individual mode, or in group mode, all of them are activated by command. In operation mode, the control unit (remote control) sends a coded radio signal that is unique to each sensor block or group of sensor blocks. First of all, the security system is designed for the use of special-purpose army engineers. The security system is programmed to work in the following modes: Commands In command mode, the controlling remote control unit transmits a specific command to the selected sensor unit. It is planned to send the commans to several sensor blocks. Checking This mode allows checking the current status of the sensor block selected through the “TEST status” command. This command responds to putting the sensor unit in the ready working state, removing it from the ready working state, checking the state or neutralizing the sensor unit. Bring to Ready Working Condition The ready working state is created after the “Connect” command is issued. This command is intended to load the firing capacitor. Remove from Ready Worker State Unloading from the ready worker state occurs after issuing the “Open” command. This command discharges the firing capacitor. Neutralization By pressing the “Neutralization” button, the sensor block is neutralized. At this time, the relay is disconnected from the sensor block. Start Operating Mode The “Run” command must be transmitted after the “Connect” command. The “Start” command can be sent to each sensor block or to all sensor blocks.

5 Description of the Security Robotic System The control unit of the security system is powered by a 7.4V Li-battery. The battery is located in a waterproof compartment. The display of the control unit is permanently connected. The main components of the security system are located in a hard case. Between the components is filled with foam. Cases for automobile, aviation, etc. transport is possible. The control unit of the security system concentrates all the elements of the control for processing. The command of the security system is carried out by radio communication. There are two “Encoding” buttons in the control block, through which various commands are selected from the menu. A LED type indicator is placed on the cover of the security system, which shows the connection of the food. There is also a pin on the cap, which, in “Neutralization” mode, protrudes 10 mm above the cap for better control of the situation. The power button and remote protection button are located at the bottom of the security system. In two-way radio communication mode, when a request is received

Remote Control Robotic System for the Perimeter Security

181

Fig. 2. The security robotic system in the suitcase.

from the control unit, the protection can send information about the current state of the system. General view of security robotic system in the suitcase shown in Fig. 2. Figure 3 shows two versions of the remote control for securiry robotic system: mini (pocket version) and stationary.

Fig. 3. Two versions of the remote control for securiry robotic system: mini (pocket version) and stationary.

Figure 4 shows the developed topology diagram of the security robotic system. A two-layer board was used to create the topological scheme.

6 Security Robotic System Operation 6.1 Preparation for Operation 1. Remove the control block unit and the power source (battery) from the suitcase. Placing the battery in the lower part of the control block unit. 2. Remove the protective cap and insert the antenna into the antenna junction located on the front side of the control block unit.

182

A. Bayramov and S. Suleymanov

Fig. 4. Topology diagram of the security robotic system on two-layer board.

6.2 Operation Algorithm Connect. The connection is made within 1 s by pressing the “ON” button. The connection is observed by the reflection of information on the display. Each time the control block unit is connected, the battery voltage is measured, and if the voltage is lower than normal, the display shows “LOW BAT” (in percent). At this time, it is necessary to feed the battery. Command Selected. First, the remote control is connected with the “On” button. By “ +”, “–” (up, down) buttons using the number of block detector is hanged (from 1 to 10). The display shows the selected command. The selected block detector launcher can be activated or deactivated using the “A/D” (Active/Disable) button. If the “A/D” button is pressed on with the selected number, then the selected block detector number is shown on the display, and if it is necessary to select other block detectors, the numbers of the block detectors are changed with the “ +” and “–” keys. The number of block detectors to be used can be confirmed by pressing the “T/P” (Confirm/Start) button after activating (or deactivating) them with the “A/D” button. Confirmed start of block detectors need to press the “T/P” button again to transmit the start command to the block detector. By selecting the “ALL” button on the display, the command applies to all block detectors. Figure 5 shows the developed electrical scheme of the security system. Developed topological scheme of security robotic system telemetry, that is, remote control with electromagnetic waves shown in the Fig. 6. Figure 7 shows the schematic diagram of security sistem telemetry.

7 Results of Laboratory and Field Tests of Security System The tests were carried out in laboratory and field conditions with 2 different remote controls from the initial test. 1 control panel is designed for every 10 seismic sensor blocks. At first, the MRP-10M type control panel was used. After preliminary tests in laboratory conditions, the field tests were conducted in open air conditions and the results of the tests were as follows. The field tests were conducted at a distance of 300–1200 m. At a distance of 300– 1000 m, both on the ground and at a height of 1 m, the system worked perfectly normally.

Remote Control Robotic System for the Perimeter Security

183

Fig. 5. Security robotic sistem electrical circuit.

Fig. 6. Topological scheme of security robotic system telemetry.

At 1200 m, 1 m above the ground surface, the data sent was not received. Later, the tests were conducted using the MRP-10 m control panel. These tests were carried out at distances of 200–400 m. The tests were successful both on the ground surface and at a height of 1 m above the ground surface. Tests were conducted in different weather conditions. Weather conditions did not adversely affect the tests. A total of 100 field tests were conducted. The tests showed positive results. During the tests, there was no violation of the working mode of the protection system and it showed 100% reliability.

184

A. Bayramov and S. Suleymanov

Fig. 7. Security sistem telemetry electrical diagram.

8 Conclusion Thus, the paper shows the results of the development, preparation and preliminary laboratory and field tests of the security system working in the telemetry mode. The working principle, management, operation rules, management and telemetry of the security system were explained. Cryptographic coding based on the Neyman algorithm was used to encrypt the telemetry of the security system. During 100 laboratory and field tests of the security robotic system, no violations of the system’s operation mode were observed.

References 1. Vorona, V., Tixonov, V.: Control Systems and Access Control. Goryachiya Liniya-Telekom, Moscow (2010). (In Russ) 2. Hashimov, E., Bayramov, A., Abdullayev, F., Suleymanov, S., Nazarov, M.: Analysis of systems that detonate mines from a distance. In: Proceeding of “Defence and Srcurity” Scientificpractical Conference, p. 24. Bilik Fondu, Baku (2017) 3. Bajaj, S., Bopardikar, S.D., Von Moll, A., Torng, E., Casbeer, D.W.: Perimeter Defense using a Turret with Finite Range and Service Times. arXiv preprint arXiv:2302.02186 (2023) 4. Guerrero-Bonilla, L., Nieto-Granda, C., Egerstedt, M.: Robust perimeter defense using control barrier functions. In: 2021 International Symposium on Multi-Robot and Multi-Agent Systems (MRS), pp. 164–172 (2021) 5. Standard, D.E.: Federal information processing standards publication 46. National Bureau of Standards, US Department of Commerce 23, 1–18 (1977) 6. GOST 28147-89: Information development systems. Cryptographic protection. Cryptographic transformation algorithm (1989) (in Russian) 7. International Electrotechnical Commission. IEC 60529: degrees of protection provided by enclosures (IP code). National Electrical Manufacturers Association, Rosslyn, VA, USA (2004)

Development of a Robot for Agricultural Field Scouting Olga Mitrofanova1,2(B)

, Ivan Blekanov1 , Danila Sevostyanov1 and Evgenii Mitrofanov1,2

, Jia Zhang1

,

1 St. Petersburg State University, 7-9 Universitetskaya Emb., St. Petersburg 199034, Russia

[email protected] 2 Agrophysical Research Institute, 14, Grazhdansky Pr., St. Petersburg 195220, Russia

Abstract. Since agricultural environments are mostly in unstructured feature information, in order to facilitate agricultural robots to better adapt to environmental change problems in agricultural environments, this paper proposes a robot system architecture for agricultural fields scouting. The article proposes: 1) the analysis of some existing field agricultural robots; 2) a novel approach that first constructs a map using the Rtabmap SLAM technique and second employs a particle filter-based Monte Carlo method to estimate the robot’s position post-hoc with a loopback detection process; 3) the pipeline of robot works (which is developing), including its architecture, taking into account selected hardware and software components. Firstly, the hardware architecture of the robot and its required sensors are considered, then the AMCL route planning algorithm is applied. The depth camera+lidar+RTK approach is used for the robot’s map construction, the simulation model of robot route planning by ROS, the robot design by SolidWorks, and finally the required sensors and hardware structure parts are analyzed and summarized. Based on the considered project, it is planned to create a field agricultural scouting robot, which will become part of the implementation of a digital twin in crop production. Keywords: Agricultural Scouting Robots · Route Planning · Sensors · Smart Agriculture · Digital Twin · Crop Production

1 Introduction Smart agriculture will critically combine technologies such as computer vision and multisensor fusion to provide effective and efficient agricultural services. Smart agriculture can utilize a wide range of advanced technologies such as artificial intelligence, the Internet of Things and robotics. In smart agriculture, the “eyes” of agricultural robots (i.e., depth cameras) are used to identify and understand the soil or vegetation in a farm field in order to obtain valuable information to accomplish manually predetermined tasks and purposes. The aggravation of the food crisis in the world determines the urgent need for the development of this industry, including in the field of modern technologies and artificial intelligence. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 185–196, 2023. https://doi.org/10.1007/978-3-031-43111-1_17

186

O. Mitrofanova et al.

Against the backdrop of the departure of leading international companies in the field of computerization and electronation of agriculture, the issue of replacing developments with domestic ones is in the first place. Indeed, without the constant improvement of agricultural technologies, it is impossible to ensure the stable development of the industry. One of the main challenges of our country is the digitalization of the crop industry, the transition to resource-saving and environmentally friendly agricultural technologies. In this regard, a team based on two organizations (AFI and St. Petersburg State University) as the main direction of research work has chosen a set of tasks in the field of so-called smart-farming, which covers various scale levels (small farm, large agricultural fields, region, etc.), concentrating on automatic driving equipment, collecting and analyzing data using robotic systems, predicting yields and improving the efficiency of existing field processing systems. Robotic unmanned technologies have proven themselves well in solving the problems of smart agriculture, many processes can be automated in greenhouses [1], as well as in orchards [2]. In recent years, the development of field robots for agroscouting [3], which allow us to automate problems related to monitoring the state of soil and crops, has also become promising. So, for example, the problem of detecting and recognizing pests in an agricultural field [4] or weeds [5] is solved. However, most of these developments are aimed at application for crops such as maize [6], vineyard [7], strawberry [8], etc. There is a need to create a field agro-robot for scouting on grain crops, which would take into account the physiological characteristics of the culture. Currently, in the agricultural environment, most agricultural sites are in an unstructured and uncertain working environment. The growing environment of crops not only depends on the terrain conditions, but is also directly affected by seasonal and weather conditions. In open agricultural environments, there are various disturbing factors, such as light, overlap, and occlusion, which make target identification difficult; in agricultural environments, there are variations in scale and perspective as well as various noises and errors that lead to inaccurate 3D reconstruction; in agricultural environments, there are semantic and functional differences with blurred and uncertain boundaries that lead to incomplete semantic separation. These are the difficulties and challenges faced by the adaptation of agricultural robot environments. Agricultural robots need to make appropriate and timely adaptations to different working conditions to ensure successful work. Agricultural robots work with crops. The spatial shape of crops varies as they grow. Agricultural robots must be able to adapt to the different spatial shapes of the crops while ensuring that the critical physiological parts of the crops are not damaged. Since the SLAM problem in smart agriculture has high academic value and broad application prospects, it has become a very active research area in academia. For navigation and crop scouting of maize fields in agricultural fields, Schmitz A [4] et al. developed an RGV vehicle that uses low-cost ultrasonic sensors to detect the location of maize rows. [9] developed a small autonomous field inspection vehicle that uses cameras and GPS for localization to identify field contours and crop types. The end of the crop rows were detected by threshold segmentation of photos of the crop and a straight line from the center of the crop rows was used for localization. Pak J [10] analyzed and

Development of a Robot for Agricultural Field Scouting

187

compared Dijkstra’s algorithm, A* algorithm and RRT algorithm and finally A* algorithm was used for greenhouse localization. Piersthal [11] proposed the Graph-SLAM method for accurate localization of forest vehicles in a forest environment using 3D LiDAR and stereo cameras. Therefore, graph search-based trajectory planning methods, such as Dijkstra’s algorithm and A* algorithm, use analytical methods for discretization to guarantee the existence of feasible solutions. However, it is difficult to apply them to trajectory planning in multidimensional space due to the large computational effort. The trajectory planning method based on random sampling probability, which only requires the positions of computer robots to collide, can effectively solve the trajectory planning problem in multidimensional space and complex constraints. Moreover, its effectiveness is related to the collision detection module, and its performance in the low-dimensional case needs to be improved. In addition, most researchers’ SLAM-based studies are mostly used indoors, while there is less literature on SLAM studies in outdoor agricultural scenarios, which are highly influenced by ambient light. In this study, we propose a novel approach that first constructs a map using the Rtabmap SLAM technique and second employs a particle filter-based Monte Carlo method to estimate the robot’s position post-hoc with a loopback detection process. This approach allows for fast and accurate robot localization in outdoor agricultural scenarios. The innovation of this paper is that using LIDAR + RTK for SLAM map construction and navigation is difficult to guarantee high accuracy in large agricultural environments, so based on the above navigation and positioning algorithm, the feature data collected by vision is added to the SLAM algorithm as crop feature correction data to correct the map construction error and improve the real-time accuracy of navigation and positioning. The objective of our research is to bridge the gap between the robotic technologies applying and optimization of precision farming agrotechnological processes for grain crops.

2 Movement Organization 2.1 Precise Positioning with RTK To ensure accurate positioning, the GPS RTK (Real-Time Kinematic) system, model RTK Emlid REACH RS2+, was chosen. This model allows you to receive the following signals: GPS/QZSS L1C/A, L2C – GLONASS L1OF, L2OF – BeiDou B1I, B2I – Galileo E1-B/C, E5b – SBAS L1C/A and others. The essence of this complex is to install a rover (AVG) on a robot, which, in turn, communicates with a base installed at a distance of up to 8 km. Hours of operation, depending from the regime, starts from 16 h. Thus, this kit is excellent for positioning the robot in the agricultural field. Processing of the signals received by the base takes place in the open-source program RTKLib, which is free and offers the following set of functions: • • • •

support for various GNSS positioning modes, in real time and for post-processing; support for standard and precise positioning algorithms; support for various RTK complexes; support of communication protocols Serial, TCP/IP, NTRIP, FTP/HTTP.

188

O. Mitrofanova et al.

2.2 Building a Map in an Unknown Area When the robot first passes through the field, the map of the area and the route are not predetermined. To build a map and determine routes, it is proposed to use SLAM methods in conjunction with the Percipio RGBD camera. The essence of the method is to calculate the location estimate x t of the agent and the environment map mt from a series of observations zt over discrete time with a sampling step t. The goal of the problem is to compute (mt ,t |zt ). To build a map, the SLAM Rtabmap method is used – this is a method based on finding and matching the visual data of the sensors (In this case, the RGBD vision sensor FS830-HD has two main infrared cameras, a color camera and a fill light system. (An active binocular vision technique is used)). The benefits of using this camera, due to the fact that in strong natural light, the laser scatter in the structured light camera is susceptible to great interference, therefore, in outdoor conditions and can only be used indoors; and none of the ordinary binocular stereo vision cameras are suitable for use in dim environments or where features are not obvious. It can also play the role of image optimization by adjusting the brightness of the optical enhancement system or the exposure parameters of the image sensor according to the needs of the specific working environment light. The environment sensing component contains lidar, imu inertial measurement unit. LIDAR, as an important environment sensing sensor, can collect the outline of the surrounding obstacles in the current environment. Imu can quickly obtain the angle, angular velocity and acceleration of the angle change of the current environment and other information. LIDAR scans the surrounding 360-degree environment by rotation, and the acquired environmental data can be used by slam algorithm to quickly construct a map of the mobile robot’s environment and reflect the robot’s current position change by this map. The method of constructing a map using depth cameras is initially capable of scanning volumetric space. In this case, the algorithm will map the projections of all objects that it scans during the movement of the robot. As a result, the resulting map will be less visual, but more complete and secure for the robot to move around. Moreover, the robot’s path planner will build its route at once, taking into account all the objects; while when using a map built using lidar, the robot will first drive up to an obstacle not fixed by this same lidar, then recognize it with the help of a camera, and only after that, rebuild its route [12, 13]. Rtabmap SLAM builds not only a 2D map, but also creates a database of visual images. Thus, Rtabmap searches for global matches in visual images, while the compiled room map can be used by lidar to locally adjust the location of the robot in space. In an agricultural environment, the robot needs to achieve all inspection tasks through autonomous navigation. The key to navigation is robot positioning and path planning, which are based on ROS move_base and amcl. Among them, move_base provides the main operation and interaction interface for robot navigation. Meanwhile, in order to ensure the accuracy of the robot’s navigation path, the robot also needs to estimate its own location through amcl implementation. In the ROS-based robot navigation, the robot only needs to publish the necessary sensor information and the target location for navigation. Real-time obstacle avoidance of the robot is achieved by collecting information from

Development of a Robot for Agricultural Field Scouting

189

LiDAR/Laser_scan and publishing odometer information nav_msgs/Odometry and the corresponding TF transformations. AMCL Monte Carlo Positioning. In the process of outdoor farming operations, it is necessary to ensure that the robot accurately reaches each instrument monitoring station, which requires the robot to realize the autonomous positioning function. Autonomous positioning of mobile robots means that the robot can project its own position in the map in any situation, which is mainly divided into two parts: global positioning and local positioning [14]. The robot determines its position relative to the environment based on the given environment map, based on the perception of the environment and its own motion, and its positioning model is schematically shown in Fig. 1.

Fig. 1. Autonomous positioning model of mobile robot.

KLD Sampling. The size N of the sample set used to represent the confidence level is an important parameter for the efficiency of particle filtering. In order to avoid scattering due to Monte Carlo localization sampling consumption, a sufficiently large sample set must be chosen to achieve accurate global localization and position tracking of the robot. In the early stages of localization, a sufficiently large number of particle sets is needed to accurately represent the state confidence of the robot. Using KLD sampling in amcl localization, for each particle filtering iteration, the number of samples is determined with probability 1 − δ. That is, the number of particles required is determined based on statistical bounds on the quality of the sampled approximation such that the error between the true posterior and the sampling-based approximation is less than ε. The number of particles required for sampling is given in the algorithm. The more dispersed the particles are, the larger the value of k and the larger the Mχ , the larger the number of particles sampled. In robot navigation, the initial poses of the robot are unknown, and the initial confidence bel(x0 ) is used to reflect the initial positional information of the robot, initialized by a uniform distribution over the space of all free poses on the map as follows: bel(x0 ) =

1 , |X |

(1)

where: |X | - the volume of all positional spaces in the map. Simulation tests are performed in ros, as shown in the Fig. 2.

190

O. Mitrofanova et al.

In the initial state of the robot, the robot’s position in the map is unknown, and the particles are uniformly distributed in the map in terms of their poses. As the robot moves and the robot’s state is estimated based on the odometer input and the LIDAR observations, the confidence in the robot’s position increases and the particles tend to converge. It should be noted that it is additionally planned to provide for the option of loading a ready-made polygon map of an agricultural field in the kml format. In this case, the construction of the route is simplified.

Fig. 2. Amcl positioning.

2.3 Obstacle Avoidance Avoiding obstacles encountered in an agricultural field is proposed to be carried out using 2D LiDAR – a type of sensor that uses laser beams to measure distances and create accurate maps of the environment. The principle of operation of LiDAR is to emit laser beams and measure the time required for the light to reflect back to the sensor, which allows you to calculate the distance to the object, as well as its geometric characteristics. LiDAR “LSLIDAR N10P” is used for this purpose. The difference between 2D and 3D sensors is the level of detail that can be captured and the types of information that can be extracted from the data. Two-dimensional LiDAR, as the name implies, collects information in two dimensions, on a plane in front of it, the general principle of operation of such a sensor is shown in Fig. 3 [15]. To implement obstacle avoidance, the sensor must be installed at the bottom of the robotic complex in order to collect information about obstacles as fully as possible [16].

Development of a Robot for Agricultural Field Scouting

191

Fig. 3. How 2D LiDAR Works to Avoid Obstacles.

The algorithm used in our work is designed to be used on computers with limited computing capabilities, in other words, we can process data directly on the processor that controls the robot. The pseudocode of the algorithm used is shown in Table 1. Table 1. Algorithm of cluster contouring. Input: Set of clusters B Output: Set of obstacles O 1

O ← ∅;

2

for i = 1, 2, . . . , c do :

3

Oi ← ∅;

4

for j = 1, 2, . . . , (|Bi | − a)do :     (j−1) (j−1) (j) (j) s← θ , dmin , θ , dmin ;

5 6 7

add s to the set Oi ;     (j−1) (j−1) (j) (j) s← θ , dmax , θ , dmax ;

8

Add s to the set Oi ;

9

Add Oi to the set O;

10

end

11

end

192

O. Mitrofanova et al.

The data received from the sensor is described in the polar coordinate system: θ = (θ1 , θ2 , . . . , θn ),

(2)

d = (d1 , d2 , . . . , dn ),

(3)

where θ is the vector of angles, d is the vector of distances, n is the number of measurements. This data can be converted to a Cartesian coordinate system as follows:  xi = di cos θi , i ∈ [1, n]. (4) yi = di sin θi Thus, points xi , yi are obtained by scanning relative to the position of the robot. Also, the necessary data for the developed algorithm is the robot displacement vector S = (Sx , Sy , Sϕ ) from the previous scan by the LiDAR sensor to the current one. The value of this displacement vector may have an error (as a rule, it is impossible to accurately calculate the distance traveled by the robot) [17].

3 Pipeline of Robot Works 3.1 Robot Concept While doing research, we focused, among other things, on reducing the price of a robotic complex. Many parts are made in-house using machine tools or a 3D printer. Figure 4 shows a preliminary diagram of a robot with caterpillars and movement sensors.

Fig. 4. Preliminary scheme of the robot.

As we can see from the figure, the robot uses a caterpillar’s base, the sensors are mounted on the outer parts of the robot, the motors and the control board are hidden inside the body.

Development of a Robot for Agricultural Field Scouting

193

3.2 Choice of Components The choice of components for the robotic complex was carried out with a focus on price and availability. Next, the advantages and disadvantages of control panels and chassis will be considered. Table 2 shows the set of components for the robot. Table 2. Comparison of the cost of RTK complexes. Name

Quantity

Price per piece, rubles

Wheel with motor OR

4

7703,57

Rubber caterpillar

1

13744,84

Raspberry Pi 4

1

16115,46

Nvidia Jetson Nano

1

20490,98

Soil sensor

1

15067,50

Wiring

5m

1000

OR

Table 2 shows only those components that we do not have in stock. The rest of the components are available. These are components such as: • • • • •

an electric motor for moving the robot with the option of using caterpillars; rechargeable batteries for powering the robotic complex; 2D lidar for avoiding obstacles; RGBD camera to implement computer vision; RTK module for precise positioning.

The body and fasteners of the robot will be made using a 3D printer. Our team is more inclined to use a robot with a caterpillar, since movement on such base is easier than on wheels, it is cheaper and it is easier to implement turning mechanisms. Choosing a control panel is an important step in robot design. We agreed on the choice of Nvidia Jetson Nano, as it outperforms its competitors in key indicators (for example, Raspberry Pi4). The Nvidia Jetson Nano has a faster modern CPU, expandable RAM up to 8 GB, and more ports for connecting and testing possible sensors. In conclusion, we tend to use the Nvidia Jetson Nano, since there are a sufficient number of inputs/outputs, the board is completely complete, the power allows us to do the processes we need. 3.3 Pipeline of Our Model Figure 5 shows the pipeline of actions of our robot. It begins with the release of the robot on the field. After that, you can conditionally divide the actions of the robot into

194

O. Mitrofanova et al.

collecting data on the field itself and data on movement. The collection and processing of movement data is carried out in real time on the control panel of the robot, and is also collected on the server for subsequent construction of a field map.

Fig. 5. Pipeline of robot works.

The architecture pipeline consists of the following main modules: • After launching the robot in the field, we separate 2 aspects of its work: movement and data collection. • While moving, the robot moves across the field and collects data (marked as Movement Data). It collects information about obstacles using 2D Lidar and RGBD cameras, tracking GPS coordinates, collecting data for terrain mapping using the RTK complex. • On the Movement Processing on CPU block, all the data collected about the movement is processed in real time to optimize the movement. • In the Movement correction block, the processed data is used to change the robot’s movement, if necessary. • Data Collection block. This block deals with the data that the robot collects directly for future analysis, such as plant health data, soil data, etc.

Development of a Robot for Agricultural Field Scouting

195

• The Field data collection block describes directly the collection of field data, on the basis of which yield analysis, analysis of plant diseases, etc. will subsequently be carried out. • The collected data on plants are transferred to a common database (Data Base block), in which information will be collected and analyzed in the future. • The Database Processing block is directly responsible for processing the collected data: analyzing the results, comparing with past results, and predicting yields. The collected field data on plants and soil are transferred to a single service for further processing, including by artificial intelligence methods, and the output of analysis results for decision-making in various problems: yield forecasting, determining the required doses of fertilizers for application, biomass forecasting, etc. • In the end, all collected data is used to further combine with data from other sources (drone, underground sensor), which are used to forecasting the yield of the field.

4 Conclusion An increasingly actual and perspective direction is the creation of field agricultural scouting robots that allow real-time collection and processing of heterogeneous big data to solve a large number ‘of management tasks: from yields forecasting to building a map of fertilizer doses for applying specialized precision machines. The paper proposes a project of such robot. It is planned to use it for research problems on grain crops and potatoes. It will allow us to quickly collect various information about crops and soil (optical, agrochemical, agrophysical indicators, etc.) for further analysis. In fact, the considered task is one of the first stages of creating a digital twin for crop production. A digital twin is, in fact, a digital analogue of a real object, reflecting its state and behavior in virtual space over time. The use of digital twins in the future will allow agricultural producers to manage agro-technological operations remotely based on digital information in real time, to model the consequences of management actions based on real data [18, 19]. It is assumed that the data collected by the robot and the unmanned aerial system will be automatically processed in a web service, after which the system will issue a result with management decisions, for example, a map with fertilizer doses, which will be loaded into the tractor’s on-board computer for automatic differential application to the field [20].

References 1. Gao, J., et al.: Development and evaluation of a pneumatic finger-like end-effector for cherry tomato harvesting robot in greenhouse. Comput. Electron. Agric. 197, 106879 (2022) 2. Wakchaure, M., Patle, B.K., Mahindrakar, A.K.: Application of AI techniques and robotics in agriculture: a review. Artif. Intell. Sci. 3, 100057 (2023) 3. Yamasaki, Y., Morie, M., Noguchi, N.: Development of a high-accuracy autonomous sensing system for a field scouting robot. Comput. Electron. Agric. 193, 106630 (2022) 4. Schmitz, A., Badgujar, C., Mansur, H., Flippo, D., McCornack, B., Sharda, A.: Design of a reconfigurable crop scouting vehicle for row crop navigation: a proof-of-concept study. Sensors 22, 6203 (2022)

196

O. Mitrofanova et al.

5. Pradel, M., de Fays, M., Seguineau, C.: Comparative life cycle assessment of intra-row and inter-row weeding practices using autonomous robot systems in French vineyards. Sci. Total Environ. 838(3), 156441 (2022) 6. Manish, R., Lin, Y.C., Ravi, R., Hasheminasab, S.M., Zhou, T., Habib, A.: Development of a miniaturized mobile mapping system for in-row, under-canopy phenotyping. Remot. Sens. 13(2), 276 (2021) 7. Longo, D., Pennisi, A., Bonsignore, R., Muscato, G., Schillaci, G.: A multifunctional tracked vehicle able to operate in vineyards using GPS and laser range-finder technology. In: International Conference Ragusa SHWA2010-September 16–18 2010 Ragusa Ibla Campus-Italy “Work safety and risk prevention in agro-food and forest systems (2010) 8. Ge, Y., Xiong, Y., Tenorio, G.L., From, P.J.: Fruit localization and environment perception for strawberry harvesting robots. IEEE Access 7, 147642–147652 (2019) 9. Bengochea-Guevara, J.M., Conesa-Muñoz, J., Andújar, D., et al.: Merge fuzzy visual servoing and GPS-based planning to obtain a proper navigation behavior for a small crop-inspection robot. Sensors 16(3), 276 (2016) 10. Pak, J., Kim, J., Park, Y., et al.: Field evaluation of path-planning algorithms for autonomous mobile robot in smart farms. IEEE Access 10, 60253–60266 (2022) 11. Pierzchała, M., Giguère, P., Astrup, R.: Mapping forests using an unmanned ground vehicle with 3D LiDAR and graph-SLAM. Comput. Electron. Agric. 145, 217–225 (2018) 12. Labbé, M., Michaud, F.: RTAB-Map as an open-source lidar and visual simultaneous localization and mapping library for large-scale and long-term online operation. J. Field Robot. 35, 416–446 (2019) 13. Silva, B.M.F.D., Xavier, R.S., Gonçalves, L.M.G.: Mapping and Navigation for Indoor Robots under ROS: An Experimental Analysis. Preprints.org. 2019070035 (2019) 14. Al-Turjman, F.: A novel approach for drones positioning in mission critical applications. Transactions on Emerging Telecommunications Technologies 33(1), e3603 (2022) 15. Bouazizi, M., Mora, A.L., Ohtsuki, T.: A 2D-Lidar-equipped unmanned robot-based approach for indoor human activity detection. Sensors 23(5), 2534 (2023) 16. Baek, S., Lee, T.-K., Se-Young, O., Ju, K.: Integrated on-line localization, mapping and coverage algorithm of unknown environments for robotic vacuum cleaners based on minimal sensing. Adv. Robot. 25, 16511673 (2011) 17. Mochurad, L., Hladun, Y., Tkachenko, R.: An obstacle-finding approach for autonomous mobile robots using 2D LiDAR data. Big Data Cogn. Comput. 7, 43 (2023) 18. Dyck, G., Hawley, E., Hildebrand, K., Paliwal, J.: Digital Twins: a novel traceability concept for post-harvest handling. Smart Agr. Technol. 3, 100079 (2023) 19. Purcell, W., Neubauer, T.: Digital twins in agriculture: a state-of-the-art review. Smart Agr. Technol. 3, 100094 (2023) 20. Mitrofanova, O., Yakushev, V., Zakharova, E., Terleev, V.: An alternative approach to managing the nitrogen content of cereal crops. Smart Innov. Syst. Technol. 247, 481–491 (2022)

Optimization of the Placement of Measurement Points and Control of the Power of Moving Sources in Rod Heating Kamil Aida-Zade(B)

and Vugar Hashimov

Institute of Control Systems of the Ministry of Science and Education of Republic of Azerbaijan, 68, B. Vahabzadeh street, Baku AZ1141, Azerbaijan [email protected]

Abstract. The problem of synthesis of power control of point-wise heat sources of heating of a rod moving along the rod by given trajectories and optimization of the placements of temperature measurement points is considered. To form the current power values of each of the heat sources, it is proposed to use the formula of their linear dependence on the temperature of the rod at the measured points. In general, the original optimal control problem is reduced to finding a finitedimensional vector of feedback parameters and coordinates of measurement points that optimize the given objective functional. Regarding the feedback parameters and the coordinates of the measurement points, the necessary conditions for the optimality of the functional of the problem are formulated, containing formulas for the components of the gradient of the objective functional. The obtained formulas make it possible to use effective numerical first-order optimization methods for solving the problem. The results of numerical experiments obtained on initial test data are presented, and the analysis of the results is carried out. In particular, the influence of the temperature measurement errors at the measurement points on the quality of process control, namely on the value of the objective functional is analyzed. Keywords: Rod Heating · Feedback Control · Moving Sources · Feedback Parameters · Temperature Measurement Points

1 Introduction The problem of optimal synthesis of control of the process of heating a rod by lumped sources moving along given trajectories and rules is considered. The current values of the power of the sources are assigned depending on the measured values of the temperature of the rod at the points where the measuring devices are installed. The problem under consideration belongs to the class of problems of optimal control of objects with distributed parameters [1–6]. Various formulations of problems of control of lumped sources in distributed systems were studied by Butkovsky A.G. and his students [7, 8]. In contrast to the control problems for systems with lumped parameters [9, 10], the control problems for systems with distributed parameters, especially © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 197–207, 2023. https://doi.org/10.1007/978-3-031-43111-1_18

198

K. Aida-Zade and V. Hashimov

with feedback, have been studied much less [4, 11–14]. Numerical methods for solving problems of optimal control of objects with distributed parameters with feedback are even less developed. In this paper, main goal is firstly, an approach is proposed for building a control system with feedback, objects with distributed parameters, and secondly, the problem of optimizing the locations of control points used for feedback is posed and solved. Formulas for the components of the gradient of the objective functional with respect to the feedback parameters are obtained, which made it possible to use well-known effective numerical first-order optimization methods to solve the problem of control synthesis. The results of computer experiments obtained by solving the test problem are presented. The results of the work can be used in systems for automatic control and regulation of lumped sources for various objects described by some other types of initial-boundaryvalue problems.

2 Formulation of the Problem The problem of control of the heating of the rod, described by the equation: ut (x, t) = a2 uxx (x, t) − λ0 [u(x, t) − θ] +

Ns i=1

qi (t)δ(x − zi (t)),

x ∈ (0, l), t ∈ (0, T ],

(1)

with initial and boundary conditions: u(x, 0) = b(x) = b = const ∈ B,

(2)

ux (0, t) = λ1 (u(0, t) − θ ), ux (l, t) = −λ2 (u(l, t) − θ ), t ∈ (0, T ].

(3)

Here: u(x, t) is temperature of the rod at the point x ∈ [0, l] at the time t; a, λ0 , λ1 and λ2 are given parameters of the heating process; θ is ambient temperature; δ(·) is the Dirac delta function. The heating of the rod is carried out by Ns lumped sources moving along given trajectories with controlled powers determined by piecewise continuous functions qi (t), i = 1, 2, . . . , Ns such that:   (4) Q = qi ≤ qi (t) ≤ qi , i = 1, 2, . . . , Ns , t ∈ [0, T ] , where qi , qi , i = 1, 2, . . . , Ns are given. The given continuous functions zi (t) ∈ [0, l] determine the coordinates of the location on the rod of the i th source at the time t, i = 1, 2, . . . , Ns . The solution of the initial-boundary-value problem (1)–(3) is understood in the generalized sense [1, 6, 12], i.e. for an arbitrary function such that: ψ(x, T ) = 0, x ∈ [0, l], ψx (0, t) = λ1 ψ(0, t), ψx (l, t) = −λ2 ψ(l, t), t ∈ [0, T ),

Optimization of the Placement of Measurement Points

199

we have:   T l ∫ ∫ u(x, t) −ψt (x, t) − a2 ψxx (x, t) + λ0 ψ(x, t) dxdt 0 0

Ns  l T l T ∫ qi (t)ψ(zi (t), t)dt + ∫ ψ(0, t)b(x)dx + λ0 ∫ ∫ ψ(x, t)θ dxdt = i=1 0

0

T

T

0

0

0 0

+λ1 θ ∫ ψ(0, t)dt + λ2 θ ∫ ψ(l, t)dt, x ∈ (0, l), t ∈ [0, T ). In [1, 12], the existence and uniqueness of a generalized solution of the considered initial-boundary-value problem (1)–(3) were proved for arbitrary admissible functions qi (t), i = 1, 2, . . . , Ns . It is assumed that the initial temperature in (2) is the same at all points of the rod, but it is not exactly specified, and belongs to a given set B ⊂ R with a known density function ρB (b):  ρB (b) ≥ 0, b ∈ B, ρB (b)db = 1. B

The ambient temperature involved in (1), (3) does not change during heating, and its possible values are distributed over a set  ⊂ R with a known density function ρ (θ ) ρ (θ ) such that:  ρ (θ ) ≥ 0, θ ∈ , ρ (θ )d θ = 1. 

The above problem of  optimal control is consists of finding admissible values of source powers q = q(t) = q1 (t), q2 (t), . . . , qNs (t) ∈ Q, delivering to minimum on average over all possible values b ∈ B of the initial states of the rod and the ambient temperature values θ ∈  of the following objective functional: I(q) = ∫ ∫ I (ϑ; b, θ )ρB (b)ρ (θ )d θ db,

(5)

B

I (q; b, θ ) = ∫l0 μ(x)[u(x, T ) − U (x)]2 dx + εq(t) − qˆ 2 Ns

L2 [0,T ]

.

(6)

Here: u(x, t) = u(x, t; q, b, θ ) is the solution of the initial-boundary-value problem (1)–(3) with the initial condition b ∈ B, the ambiant temperature of the θ ∈  and at admissible values of the source powers q(t) ∈ Q; U (x) is the given temperature distribution of the rod, which is desirable to achieve at the end of the heating process; μ(x) ≥ 0, x ∈ [0, l] is a given weight function; ε > 0, qˆ are the given regularization parameters of the problem functional [6]. Now suppose that at given No points of the rod ξj ∈ [0, l], j = 1, 2, . . . , No , the temperature values are measured at these points continuously during its heating:  (7) uˇ j (t) = u ξj , t , t ∈ [0, T ], ξj ∈ [0, l], j = 1, 2, . . . , No .

200

K. Aida-Zade and V. Hashimov

The measured values are used to assign the current values of the power sources qi (t), i = 1, 2, . . . , Ns according to the following linear dependence on the measured temperature values: No j j qi (t) = αi [u(ξj , t) − βi ], (8) j=1

j

j

where αi , βi , ξj are feedback parameters, i = 1, 2, . . . , Ns , j = 1, 2, . . . , No [13, 14]. In (8), the value in square brackets determines the deviation of the temperature j measurement at the jth measurement point from the nominal value βi relative to the ith j source at the jth measurement point. αi is the corresponding gain factor. The nominal values are largely determined by the values of the given function U (x) at the measurement points x = ξj , j = 1, 2, . . . , No . Substituting expressions for powers with continuous feedback (8) into Eq. (1), we obtain: ut (x, t) = a2 uxx (x, t) − λ0 [u(x, t) − θ ] +

Ns  No 

j

j

αi [u(ξj , t) − βi ]δ(x − zi (t)), x ∈ (0, l), t ∈ (0, T ].

(9)

i=1 j=1

Equations (9) belong to the class of loaded equations due to the participation of the values of the searched function u(x, t) in them at the measurement points at ξj [15]. For the corresponding initial-boundary-value problems, in [16–18] the conditions for the existence and uniqueness of their solution were studied, and numerical methods for their solution were proposed and studied. In general, in the problem below, it is required to determine the considered   feed j back parameters α = (αi ) = α11 , α12 , . . . , α1No , . . . , αN1 s , αN2 s , . . . , αNNso , , β =       j (βi ) = β11 , β12 , . . . , β1No , . . . , βN1 s , βN2 s , . . . , βNNso , , ξ = ξj = ξ1 , ξ2 , . . . , ξNo , i = 1, 2, . . . , Ns , j = 1, 2, . . . , No , taking into account restrictions (4) on the power of the sources, under which the objective  will take the minimum possible value.  functional j j Denote by the parameter vector y = αi , βi , ξj ∈ Rn , n = No (2Ns +1) optimized in the j

j

problem, which consists of: No Ns parameters αi , No Ns parameters βi and No parameters ξj . The objective functional (5), (6) of the problem under consideration is finally written as follows: I(y) = ∫B ∫ I (y; b, θ )ρB (b)ρ (θ )d θ db,

(10)

I (y; b, θ ) = ∫l0 μ(x)[u(x, T ) − U (x)]2 dx + εy − yˆ 2Rn .

(11)

Here: u(x, t) = u(x, t; y, b, θ ) is the solution of the initial-boundary-value  value problem  j j with respect to Eq. (9) with optimized parameters y = αi , βi , ξj , initial condition u(x, 0) = b and ambient temperature θ .

Optimization of the Placement of Measurement Points

201

Restrictions (4) on the power of sources when using dependence (7) for feedback will turn into the following joint restrictions on the optimized parameters y and temperature at the measurement points ξj ∈ [0, l], j = 1, 2, . . . , No . qi ≤

No j=1

j

j

αi [u(ξj , t) − βi ] ≤ qi , i = 1, 2, . . . , Ns , t ∈ [0, T ].

(12)

which we denote and write in the following equivalent form: gi (t; y) = |gi0 (t; y)| − gi0 (t; y) =

qi − qi

2 qi + qi 2



≤ 0, i = 1, 2, . . . , Ns , t ∈ [0, T ]. No j=1

(13) j

j

αi [u(ξj , t) − βi ].

Problem (9), (2)–(4), (10), (11) considered below is a problem of parametric optimal control of objects with  distributed parameters with feedback. The finite-dimensional  j j vector y = αi , βi , ξj ∈ Rn is optimized. The features of the problem are: the loaded differential equation; participation in the equation Dirac δ-functions; the value of the objective functional is determined not by one solution of the initial-boundary-value problem, but by a bunch of solutions, provided that the initial condition and the ambient temperature of the take not one value, but a set of values, respectively, from the sets B and ; comparatively not very large for problems of synthesis of control systems with distributed parameters, the dimension of the resulting finite-dimensional optimization problem, determined by the double product of the number of sources and measurement points.

3 Approach and Formulas for the Numerical Solution of the Problem By direct verification according to the definition of the convexity of the functional, it is easy to show that the objective functional of the original optimal control problem (1)–(6) without feedback is convex in q(t). It is easy to show that the functional of the considered control problem (10), (11) with continuous feedback (8) is not convex in terms of the optimized feedback parameters y. The admissible range of parameters y defined by inequalities (12) is also not convex, which follows from the obvious, in the general case, non-linearity of the dependence of the solution of the initial-boundaryvalue u(x, t; y, b, θ ) on the parameters y = (α, β, ξ ), which implies the non-convexity of the considered problem of optimizing feedback parameters (8) as a whole. Nevertheless, the formulas obtained below for the components of the functional gradient can be used to numerically determine the locally optimal values of the feedback parameters or to locally refine any of their values given by an expert. It is clear that to find the optimal values of the parameters, known methods of global optimization can be used together with methods of local improvement of the values of parameters by first-order optimization methods using the approach to the numerical solution proposed below and the obtained formulas for calculating the gradient of the objective functional of the control problem by the feedback parameters.

202

K. Aida-Zade and V. Hashimov

For the numerical solution of problem (1)–(6), namely, finding the local minimum of the objective functional (10), (11), it is proposed to use the external penalty method to take into account constraints (13) [6]. Considering the possible multi-extremality of the problem due to its non-convexity, the above approach can be used to solve the problem using, for example, the “multistart” method, which uses various randomly selected starting points for the optimized parameter vector using algorithms for parallelizing the computational process. We choose the penalty functional with respect to functional (10), (11) in the following form: IR (y) = ∫B ∫ I (y; b, θ )ρB (b)ρ (θ )d θ db, IR (y; b, θ ) = ∫l0 μ(x)[u(x, T ; y, b, θ ) − U (x)]2 dx + εy − yˆ 2Rn + RG(y), Ns

2 ∫T0 gi+ (t; y) dt. G(y) =

(14)

(15)

i=1

Here R is the penalty coefficient tending to +∞, the function gi+ (t; y) = 0, if gi+ (t; y) ≤ 0, and gi+ (t; y) = gi (t; y), if g i (t; y) > 0. We have the following theorem. Theorem. The objective functional IR (y) of the problem (9), (2)–(4), (14), (15) for each given penalty coefficient R and continuous feedback (8) is differentiable with respect to the synthesized parameters y = (α, β, ξ ) and the components of its gradient are determined by the formulas:  

  ∂IR (y) j = ∫B ∫ − ∫T0 ψ(zi (t), t) + 2Rgi+ (t; y)sgn gi0 (t; y) u(ξj , t) − βi dt j ∂αi   j j +2 αi − αˆ i ρB (b)ρ (θ )d θ db, (16)      ∂IR (y) j j j = ∫B ∫ ∫T0 αi ψ(zi (t), t) − 2Rgi+ (t; y)sgn(gi0 (t; y)) dt + 2 βi − βˆi j ∂βi ×ρB (b)ρ (θ )d θ db,    Ns T  j ∂ IR (y) ∂ IR (y) ∫ αi ψ(zi (t), t) + 2Rgi+ (t; y)sgn(gi0 (t; y)) ux (ξj , t)dt+ = ∫B ∫ − i=1 0 ∂ξj ∂ξj    Ns T  j + 0 ∫0 αi ψ(zi (t), t) + 2Rgi (t; y)sgn(gi (t; y)) ux (ξj , t)dt = ∫B ∫ − i=1

  +2 ξj − ξˆj ρB (b)ρ (θ)d θdb,

(17)

(18)

where i = 1, 2, . . . , Ns , j = 1, 2, . . . , No , ψ(x, t) = ψ(x, t; y, b, θ, R) with the current vector of parametersy, admissible initial condition b ∈ B, ambient temperature θ ∈  and penalty coefficient R is the solution of the following conjugate-boundary-value problem: No  δ x − ξj ψt (x, t) = −a2 ψxx (x, t) + λ0 ψ(x, t) − j=1 (19)  Nc  j + 0 × αi ψ(zi (t), t) + 2Rgi (t; y)sgn(gi (t; y)) , x ∈ (0, l), t ∈ [0, T ), i=1

Optimization of the Placement of Measurement Points

203

ψ(x, T ) = −2μ(x)(u(x, T ) − U (x)), x ∈ [0, l],

(20)

ψx (0, t) = λ1 ψ(0, t), ψx (l, t) = −λ2 ψ(l, t), t ∈ [0, T ).

(21)

4 Numerical Experiments The solution of the problem under consideration was carried out with the following values of the data involved in its formulation: a = 1, l = 1, T = 1, λ0 = 0.01, λ1 = 0.001, Ns = 2, No = 4, μ(x) ≡ 1, U (x) = 10, x ∈ [0; 1], qi = −1, qi = 14, i = 1, 2, B = [0.1; 0.2; 0.3], ρB (b) = 1/3,  = [0.2; 0.3; 0.4], ρ (θ ) = 1/3,     z1 (t) = 0.05 + 0.9 sin2 (π t) , z2 (t) = 0.75 − 0.6 sin2 (π t) , t ∈ (0; 1]. The penalty coefficient successively, starting from R = 5, increased by 5 times for three times. The dimension of the vector of synthesized parameters y in this case is equal to No (2Ns + 1) = 20. The setsB,  and the corresponding integrals in the objective functional were discretized uniformly by three points. To approximate the Dirac δ-function, we used, as is easy to see, an everywhere smooth (differentiable) function:    x − zˇ  > σ, 0      δσ x; zˇ = 1 x−ˇz   2σ 1 + cos σ π , x − zˇ ≤ σ. In this case, for an arbitrary value σ ,  +σ ∫zzˇˇ −σ δσ x; zˇ dx = 1. The value of σ was set equal to 3hx , where hx is the step of the grid approximation of the segment x ∈ [0; 1][0; 1], which was varied during numerical experiments. Such an approximation of the Dirac δ-function smoothes the dependence ψ(x, t) of the solution of the conjugate-boundary-value problem (19)–(21), and hence the gradient of the objective functional (16)–(18), on the optimized coordinates of the location of the measurement points ξj , j = 1, 2, . . . , No , which are the arguments δ-functions [7, 19]. An iterative procedure for minimizing the penalty functional by the gradient descent method for a given penalty coefficient R was built using the formula:   yk+1 = yk − αk grady IR yk , k = 0, 1, . . . ,    (22) αk = arg minIR yk − αgrady IR yk , α≥0

with a stop criterion on argument increment yk+1 − yk R20 ≤ 0.0001.

204

K. Aida-Zade and V. Hashimov Table 1. Problem solutions obtained for two initial values y10 and y20 .   ∗ IR y 0 y

y0 1 −0.1058 −0.2652 0.0194

5.0693

0.0147

 IR y∗ M it −0.4523 0.0210

0.0002

45

−0.2608 0.0003

82

−0.4408 −0.0272 −0.312

−0.8643 0.4891

−0.2752 −0.2143 12.000

−0.5668 −0.4421 12.243

12.000

12.000

12.000

12.582

12.167

12.928

12.000

12.000

12.000

11.973

13.072

12.939

12.000

0.2537

0.5671

12.811

0.1564

0.4584

0.9244

0.7963

0.6272

0.8956

2 −0.3542 −0.3123 −0.3460 2.7898

−0.2220 0.1870

−0.7126

−0.6856 −0.3571 −0.2619

−0.9727 −0.6362 0.3212

−0.2635 −0.4339 12.044

−0.2897 −0.6535 12.499

12.026

12.041

12.185

12.202

12.530

13.167

12.047

12.005

12.006

13.073

12.243

12.737

12.079

0.9283

0.3178

13.151

0.9067

0.2764

0.7433

0.3249

0.8592

0.0500

 Fig. 1. Graphs of the function IR t; y∗ : the values of the objective functional calculated for y1∗ for the process time t.

One-dimensional minimization with respect to α in (22) was carried out using the golden section method. The implicit scheme of the grid method [20] was used to solve the direct and conjugate boundary-value problems, and the methods proposed in [17, 18] was used to take into account the loading of differential equations. The regularization parameters ε and yˆ at each value of the penalty coefficient R changed according to the algorithm proposed in [6], namely, ε decreased by a factor of 5, starting from ε = 0.05, and yˆ at the beginning for each R were chosen to be zero, and then were equated y∗ found in the previous step.

Optimization of the Placement of Measurement Points

205

Table 1 shows the results of solving the problem obtained for two initial values y0 ∈ R20 of the iterative process (22). It can be seen from the table that the corresponding values of the parameters α, β and ξ obtained in this case differ significantly, but the values of the functional IR (y) are quite small, and the number of iterations M it differs little. Here it is possible that the objective functional is either multi-extremal or has a strong ravine structure. On Fig. 1a, b shows the graphs of the function IR (t; y∗ ) of the value of the objective functional, calculated with y∗ = y1∗ for the end time of the process t. The time t = T = 1 corresponds to the time interval [0;1] for which the source synthesis problem was solved. Next, for t > 1, the process of heating the rod was observed using the synthesized vector of parameters y1∗ . Due to the relatively large values of IR (t; y∗ ) at t > 0.8, its graph is divided into two using different scales along the ordinate axis. Figure 1.b shows that the heating control process at t > 1.8 has stabilized. This means that the value of the functional at t > 1.8, which characterizes the deviation of the current temperature state on the entire rod from the desired desired temperature U (x) = 10, x ∈ [0; 1], is small, and therefore the heating control process has switched to the regularization process with feedback. Figure 2 shows the results of calculating the temperature of the rod at t = T = 1 using the optimal feedback parameters given in the first row of Table 1 in the presence of noise in measurements (7). Noisy measurements uˆ (ξj , t), j = 1, 2, . . . , No , were determined by the formula:  

uˆ ξj , t = u ηj , t 1 + ξj (t)χ , j = 1, 2, , . . . , No , t ∈ [0, T ], where ξj (t) are uniformly distributed over a function whose random values for each t ∈ [0; T ] were generated using the RAND program; χ noise level: χ = 0 corresponds to no noise, χ = 0.01 corresponds to 1% noise, χ = 0.03 to 3% noise.

Fig. 2. Calculated rod temperature for T = 1 for the optimal feedback parameters shown in the first line of Table 1 under measurement noise χ = 0 (1), χ = 0.01 (2), χ = 0.03 (3).

As can be seen from Fig. 2, the order of magnitude of the deviation from the desired state of the rod at T = 1 is quite consistent with the noise level, and the process of heating with feedback can be considered quite stable.

206

K. Aida-Zade and V. Hashimov

5 Conclusion The paper proposes an approach to solving the problem of feedback control of the power of point sources of heating of a rod moving according to given rules and trajectories. In the problem to be optimized is also the location of the points of the current state of the process. To form the current values of the powers of each of the sources, it is proposed to use for them the formula of a linear dependence on the measured values of the temperature of the rod at the measured points. Thus, the original optimal control problem is reduced to finding a finite-dimensional vector of feedback parameters and placement of measuring points that optimize the given objective functional. Formulas are obtained for the gradient components with respect to the parameters to be optimized, which allow efficient first-order optimization methods to be used for the numerical solution of the problem. The results of numerical experiments obtained on test initial data and their analysis are presented. In particular, an analysis was made of the influence of the temperature measurement error at measurement points on the quality of process control, namely, on the value of the objective functional. The approach presented in the paper and the scheme for obtaining formulas can be extended to many other problems of optimal control of lumped sources in systems with distributed parameters, which are described by other types of initial-boundary-value problems, including those with a higher dimension of the spatial variable. The proposed approach can be easily generalized to the problems of controlling two-dimensional and three-dimensional technological processes arising in metallurgy, economics and other areas. The analogues obtained above those given in the article have no fundamental differences and are associated only with technical difficulties.

References 1. Lions, J.L., Lelong, P.: Contrôle optimal de systèmes gouvernés par des équations aux dérivées partielles (1968) 2. Deineka, V.S., Sergienko, I.V.: Optimal Control of Inhomogeneous Distributed Systems. Naukova Dumka, Kiev (2003) 3. Ray, W.H.: Advanced Process Control. McGraw-Hill Book Company (1981) 4. Egorov, A.I.: Principles of the Control Theory, p. 504. Fizmatlit (2004). (In Russ.) 5. Sirazetdinov, T.K.: Optimization of systems with distributed parameters. (No title) (1977). (In Russ.) 6. Vasil’ev, F.P.: Optimization Methods. Faktorial Press, Moscow (2002). (In Russ.) 7. Butkovskiy, A.G.: Methods of Control of Distributed Parameter Systems. Nauka, Moscow (1984). (In Russ.) 8. Butkovskiy, A.G., Pustylnikov, L.M.: The Theory of Mobile Control of Systems With Distributed Parameters. M. Nauka (1980). (In Russ.) 9. Utkin, V.I.: Sliding Conditions in Optimization and Control Problems. Nauka, Moscow (1981). (In Russ.) 10. Polyak, B.T., Khlebnikov, M.V., Rapoport, L.B.: Mathematical Theory of Automatic Control, p. 500. M. LENAND (2019). (In Russ.) 11. Sergienko, I.V., Deineka, V.S.: Optimal Control of Distributed Systems with Conjugation Conditions, vol. 75. Springer Science & Business Media (2005)

Optimization of the Placement of Measurement Points

207

12. Lions, J.-L., Magenes E.: Problemes Aux Limites Non Homogenes at Application, vol. 1. Paris (1968) 13. Guliyev, S.Z.: Synthesis of zonal controls for a problem of heating with delay under nonseparated boundary conditions. Cybern. Syst. Anal. 54(1), 110–121 (2018) 14. Aida-zade, K.R., Abdullaev, V.M.: Optimizing placement of the control points at synthesis of the heating process control. Autom. Remote. Control. 78(9), 1585–1599 (2017) 15. Nakhushev, A.M.: Loaded Equations and their Application. Nauka, Moscow (2012). (In Russ.) 16. Alikhanov, A.A., Berezgov, A.M., Shkhanukov-Lafishev, M.X.: Boundary value problems for certain classes of loaded differential equations and solving them by finite difference methods. Comput. Math. Math. Phys. 48(9), 1581–1590 (2008) 17. Abdullaev, V.M., Aida-Zade, K.R.: Optimization of loading places and load response functions for stationary systems. Comput. Math. Math. Phys. 57(4), 634–644 (2017) 18. Aida-zade, K.R.: An approach for solving nonlinearly loaded problems for linear ordinary differential equations. Proc. Instit. Math. Mech. Natl. Acad. Sci. Azerbaijan 44(2), 338–350 (2018) 19. Aida-zade, K.R., Bagirov, A.G.: On the problem of spacing of oil wells and control of their production rates. Autom. Remote. Control. 67(1), 44–53 (2006) 20. Samarsky, A. A.: Theory of Difference Schemes. PML (1989). (In Russ.)

Design of Hybrid Control System for Nonaffine Plants Anatoly Gaiduk , Viacheslav Pshikhopov , Mikhail Medvedev(B) Vladislav Gissov, Ali Kabalan, and Evgeny Kosenko

,

Southern Federal University, 105/42, Bolshaya Sadovaya St, Rostov-On-Don 344006, Russia [email protected]

Abstract. The purpose of this article is to develop a design method that ensures the stability of the zero-equilibrium position of a closed nonaffine control system in a certain area. The objects described by a nonlinear system of differential equations with one control and one output are considered. A constraint is introduced, which consists in the differentiability of the right-hand sides of differential equations with respect to all state variables. The task of designing control in the form of a function of the setting action, state variables and control values at previous points in time is set. This problem is solved using a quasilinear model of the control object. In the quasilinear model, matrices and vectors are functions of the variables of the state of the control object. The control is found using an algebraic polynomial matrix method. This article presents the expressions for calculating the control in accordance with the polynomial-matrix method. Based on the given coefficients of the desired polynomial, as a result of solving an algebraic system of equations, coefficients are found that are a function of control and state variables. The fulfillment of the controllability condition guarantees the existence of a solution of the specified algebraic system. Keywords: Nonlinear System · Nonaffine Control Plant · Quasilinear Model · Polynomial Matrix Method · Controllability Condition · Static Error

1 Introduction Nonlinear control plants are characterized by a wide variety of nonlinear characteristics. In particular, this leads to one of the difficult tasks of control theory – designing the control systems for nonaffine control objects [1–4]. In the equations of such objects, control enters nonlinearly, and the influence of the control action on the derivatives of the state variables of the plant is not additive. Such plants include underwater vehicles, surface vessels, aircraft, mobile robots, etc. [2, 4, 5]. Various approaches are used to design the control systems for nonaffine objects. Most often, these include a block approach with discontinuous controls [2], systems with sliding modes [3], the maximum principle, leading to a program piecewise constant control or optimal control in terms of speed [6, 7]. The article [2] considers the problem of designing a control for nonaffine plants under the uncertainty of mathematical models describing them. It is noted that the solution of © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 208–220, 2023. https://doi.org/10.1007/978-3-031-43111-1_19

Design of Hybrid Control System for Nonaffine Plants

209

this problem by the methods of structural synthesis [8] or backstepping [9] leads to complex expressions for the obtained controls. In this regard, the paper [2] proposes a method based on sliding modes [10–12] that makes it possible to obtain a control with a simpler structure. In the article [12], the calculation expressions for the control calculated from the output are obtained, and the boundedness of the error and state variables is shown. A significant disadvantage of systems with a variable structure is the need for high-frequency switching of drives in saturation modes. Such modes require the use of special drives and can be difficult to implement, because the switching frequency in such systems is limited by the inertial properties of the drives. In [6, 7], a method for controlling plants with a block structure [11] is proposed, for which, based on the maximum principle [13, 14], a piecewise constant control is obtained, which is then approximated by a continuous function close to it. This allows maintaining the robustness of the system and avoiding the use of sliding modes. In [6, 7], the controllability and stability conditions were also obtained in the form of inequalities corresponding to [15]. However, in [6, 7] the question of the influence of the approximation of the “signum” function by a continuous function on the degree of optimality of the resulting system is not studied. The control is obtained in a continuous form and its implementation in digital form is not considered. In this article, to solve the problem of designing control systems for nonaffine plants, it is proposed to apply an algebraic polynomial-matrix design method based on quasilinear models [16]. These models are easily constructed by an analytical or numerical method [17], if the nonlinearities of the plant are differentiable, and make it possible to find a control that ensures stability, the required speed, and other indicators of the quality of the control process. It will be shown below that this control is easily implemented using digital automation tools, without loss of accuracy and quality of the nonlinear control system. As a result, a hybrid control system is formed, which includes a continuous nonaffine control plant and a discrete control device.

2 Formulation of the Problem Let a nonlinear nonaffine plant with one control and one controlled variable be described by a differential equation of the form: x˙ = ζ (x, u, f ).

(1)

where x = [x 1 x 2 … x n ] is the n-vector of state variables; ζ(x,u,f ) is the nonlinear n-vector function, differentiable with respect to all variables x i , i = 1, 2, . . . , n, as well as with respect to u and f , and ζ(0,0,0) = 0; u is the control; f is the external disturbance; y is the controlled variable; ψ(x) is the scalar function, also differentiable in all variables x i ; the state vector x is assumed to be measurable; and 0 is the null n-vector. The task is to determine the control as a function of the setting action g = g(t), the state vector x and, possibly, the control. This control should ensure the stability of the system; zero value of static error ε(t) = g(t)-y(t) at t → ∞. The duration of the transient process t pp according to the setting effect g(t) = g0 1(t) under zero initial conditions ∗ . Here g is the value at which the control should not exceed the specified value of tpp 0 module u = u(x) does not exceed the allowable values.

210

A. Gaiduk et al.

Note that since the control plant is nonlinear, it is impossible to specify in advance the possibility of ensuring the stability of the position x = 0 of the control system as a whole or in some area of its state space, since this significantly depends on the properties of nonlinearities from Eq. (1) of each specific object [16, 17].

3 Solution of the Problem Since the vector function ζ(x,u,f) is differentiable, moreover, ζ(0,0,0) = 0 Eqs. (1) can be represented by the following quasilinear model: x˙ = A(x)x + b(x, u)u + bf (x, u, f )f ,

y = cT (x)x,

(2)

where: ⎤ ⎡ ⎤ b1 (x, u) a11 (x) a12 (x) . . . a1n (x) ⎢ ⎥ ⎢ a21 (x) a22 (x) . . . a2n (x) ⎥ ⎥, , b(x, u) = ⎢ b2 (x, u) ⎥ A(x) = ⎢ ⎦ ⎣ ⎣ ... ... ... ... ... ⎦ an1 (x) an2 (x) . . . ann (x) bn (x, u) ⎤ ⎡ b1f (x, u, f ) ⎢ b2f (x, u, f ) ⎥ T ⎥, c (x) = [c1 (x), c2 (x), . . . , cn (x)], bf (x, u, f ) = ⎢ ⎦ ⎣ ... bnf (x, u, f ) ⎡

(3)

The functional coefficients in (2) and (3) when obtained by the analytical method [17, 18] are determined by the expressions:  1  1  aij (x) = ζij (x1 , . . . , xj−1 , θ xj , 0n−j , 0, 0)d , θ, ·bi (x, u) = ζiu (x, θ u, 0)d θ, 0



1

b1f (x, u, f ) = 0

ζif (x, u, θ f )d θ, ·cj (x) =

 0

0

1

(ψj x1 , . . . , xj−1 , θ xj , 0n−j )d θ, (4)

where ζij (x, u, f ) = ∂ζi (x, u, f )/∂x; ζiu (x, u, f ) = ∂ζi (x, u, f )/∂u; ζif (x, u, f ) = ∂ζi (x, u, f )/∂f ; ψj (x) = ∂ψi (x)/∂x · ψj (x) = ∂ψj (x)/∂x; and 0n-j is the sequence 0, …, 0, containing n – j zeros. We emphasize that the quasilinear model (2) – (4) describes the nonlinear plant (1) exactly, i.e. with the preservation of all its nonlinear features. Expressions for determining the coefficients from the expressions (2) and (3) by the numerical method can be found in [17]; however, in this case, the model (2) – (3) becomes discrete-continuous and describes the nonlinear plant (1) approximately. In the expressions (2) – (3), the nonaffinity in control of the considered object (1) is manifested in the fact that the input vector b(x,u) depends on the control u. As can be seen, the Eqs. (2) are similar in form to the equations of one-dimensional linear control plants [19], except that the matrix and vectors in these equations are nonlinear functions of plant variables, which is why they were called a quasilinear model [20, 21].

Design of Hybrid Control System for Nonaffine Plants

211

Note that for f = f (t) = 0 and ∂ 2 ζiu (x, u)/∂u2 = 0 vector b(x), as well as the matrix A(x), depend only on the state vector. In the Western literature, such models (2) are called state-dependent coefficient (SDC) [22, 23], while the method of constructing the matrix A(x) and vector b(x) is usually not covered in such works. As in the linear case, the design problem here has a solution if the quasilinear model (2) satisfies the controllability condition, which has the form: |detU (x, u)| ≥ ζ1 > 0, x ∈U , u ∈ Iu ,

(5)

where ζ1 > 0 is some not too small number; U is some region of the state space Rn , which includes the point x = 0, and in which the condition (5) is satisfied; I u is the interval of allowable control values u; and the functional matrix of controllability U(x,u) U (x, u) is determined by the expression

U (x, u) = b(x, u)A(x)b(x, u) . . . An−1 (x)b(x, u) . (6) The design of a control system for a nonaffine plant (1) based on a quasilinear model (2) – (4) under condition (5) is conveniently carried out using the algebraic polynomial-matrix method [16], in which the control law is sought in the form: u = u(x) = h0 g − hT (x)x = h0 g − [h1 (x)x1 + h2 (x)x2 + . . . + hn (x)xn ].

(7)

i.e. in the form of influence and state control [24]. To derive the calculated relations in this case, we substitute the control (7) into the first Eq. (2), taking into account the second Eq. (2), and present similar terms; as a result we get: x˙ = D(x)x + b(x, u)h0 g + bf (x, u, f )f , y = cT (x)x,

(8)

D(x) = A(x) − b(x, u)hT (x).

(9)

where:

The characteristic polynomial D(p,x) = det[pE – D(x)] of the matrix D(x) (9) by formula (A.25) from [25] can be represented as D(p, x) = A(p, x) +

n



hi (x)Vi (p, x, u), A(p, x) = det pE − A(x) .

(10)

i=1

where: n−1 

 Vi (p, x, u) = ei adj pE − A(x) b(x, u) = vij (x, u)pj , i = 1, n, i=0

(11)

ei is the i-th row of the identity n × n-matrix E. The dependence of the coefficients vij (x,u) of polynomials (11) on the control u is also a consequence of the nonaffinity in control of the plant (1). Note. The matrix D(x) (9) is included in the equation of the quasilinear model (8) of the closed system (1), (7), but for brevity, we will call it the system matrix.

212

A. Gaiduk et al.

In accordance with the algebraic polynomial-matrix method, the functional polynomial D(p,x) in the first expression (10) is replaced by the Hurwitz polynomial D* (p), which has constant, real and different roots, i.e. λ∗j , i.e. D∗ (p) =

n j=1

  ∗ p − λ∗j = pn + δn−1 pn−1 + . . . + δ1∗ p + δ0∗ ,

(12)

    where: λ∗j < −ζ2 < 0; λ∗j − λ∗l  > ζ3 , j = l, j,l = 1,2,…,n, ζ2 > 0, ζ3 > 0 [18, 20]. Having performed the indicated replacement in the equality (10) and moving the polynomial A(p,x) to its left side, we obtain a polynomial equation with respect to the coefficients hi (x) from the expression (7): n

hi (x)Vi (p, x, u) = R(p, x),

(13)

i=1

R(p, x) = D∗ (p) − A(p, x) =

n−1 j=0

ρj (x)pj .

(14)

It is advisable to obtain the solution of the polynomial Eq. (13) by passing to the equivalent system of algebraic equations [16, 20], which in this case has the form: ⎡

⎤ ⎡ ⎤ ⎡ ν1,0 ν2,0 h1 ρ0 . . . νn,0 ⎢ ν1,1 ν2,1 ⎥ ⎢ ⎥ ⎢ . . . νn,1 ⎥ ⎢ h2 ⎥ ⎢ ρ1 ⎢ = = ⎣ ... ... ... ... ⎦ ⎣ ... ⎦ ⎣ ... ν1,n−1 ν2,n−1 . . . νn,n−1 hn ρn−1

⎤ ⎥ ⎥. ⎦

(15)

Note that many coefficients in system (15) in this case are functions of the vector x and control u, but for brevity, the arguments of these coefficients in (15) are omitted. We also emphasize that the existence of a solution to system (15) is guaranteed by the fulfillment of the controllability condition (5). The solution of system (15) is determined by the vector of coefficients from (7): h(x, u) = [h1 (x, u) h2 (x, u) . . . hn (x, u)],

(16)

which in the case of a nonaffine plant can be functions not only of the vector x, but also of the control u. Therefore, the system matrix can also depend on the control, i.e. in the general case, it has the form D(p,x,u). Nevertheless, its characteristic polynomial D(p,x,u) is equal to the polynomial D* (p) (12), i.e. the roots of the polynomial D(p,x,u) are constant, distinct, real, and negative. Therefore, it follows from the theorem proved in [21] that there is some neighborhood 0 ∈ u of the equilibrium position of system (8) n, which the following condition is satisfied: lim x(t, x0 ) = 0, ∀x0 ⊂ 0 , x ∈ x .

t→∞

(17)

Here x(t, x 0 ) is the solution of the free system (8), i.e. at g(t) = f (t) = 0; x0 is the vector of initial conditions for this solution.

Design of Hybrid Control System for Nonaffine Plants

213

It follows from condition (17) and Eq. (8) that in the steady state at f (t) = 0 and g(t) = g0 1(t) the following equality takes place:     ◦ ◦ (18) 0 = D x◦ , ug x◦ + b x◦ , ug h0 g0 . From here, taking into account the second equality (8) and determining the deviation of the system ε = g − y, we derive the following expressions:         ◦ ◦ ◦ ◦ x◦ = −D−1 x◦ , ug b x◦ , ug h0 g0 , y◦ = −cT D−1 x◦ , ug b x◦ , ug h0 g0 ,    

◦ ◦ ε◦ = 1 + cT (x◦ )D−1 x◦ , ug b x◦ , ug h0 g0 .

(19)

In equalities (18), (19) x 0 , y0 , ε0 are the steady-state values of the corresponding variables. From (19) it follows that the steady value of the controlled variable of the nonlinear system (8) may have a non-zero value, and its error ε0 can be provided equal to zero, if only the following condition is met: 1 + cT (x)D(−1) (x)b(x, u) = 0.

(20)

When condition (20) is met, to ensure the zero value of the static error for the setting action g(t) = g0 1(t), the coefficient h0 from the expression (7) should be equal to: h0 (x, u) = −1/cT (x)D−1 (x, u)b(x, u).

(21)

The fulfillment of condition ε0 = 0 n determining the coefficient h0 (x,u) by equality (21) is due, first of all, to the fact that for all x and u the matrix D(x,u) (9) has an inverse, and secondly, due to condition (16) in the steady state x(t) → x 0 , and u(t) → u0 , respectively. The condition (20) and equality (21) can be simplified if we take into account that D−1 (x,u) = [adjD(x,u)]/detD(x,u). When designing nonlinear Hurwitz systems by the algebraic polynomial-matrix method, for all values of the vector x = x(t) and control u(t) the condition D(p,x,u) = det[pE – D(x,u)] = D* (p) is satisfied. Therefore, considering equality (12) and the properties of the determinant, we have the equality detD(x,u) = D(p,x,u)p=0 = (-1)n δ0∗ . Therefore, the denominator in equality (21) can be represented as: (-1)n cT (x)[adjD(x,u)]b(x,u)/δ0∗ . . On the other hand, according to (9), D(x,u) = A(x)b(x,u) hT (x,u), therefore, in accordance the with expression (A.26) from [25], 



adjD(x, u) b(x, u) = adj A(x) − b(x, u)hT (x, u) b(x, u) =

= adjA(x) b(x, u).

(22)

Here, adj denotes the adjoint matrix [26]. From the foregoing and relations (21), (22) it follows that the equality ε0 = 0 at g(t) = g0 1(t) can be ensured if only: h0 (x, u) = (−1)n+1 δ0∗ /γ (x, u), γ (x, u) = 0, x ∈ x , u ∈ Iu ,

(23)

214

A. Gaiduk et al.



γ (x, u) = cT (x) adjA(x) b(x, u).

(24)

˙ y = cT x, In the case of linear plants described by a system of equations x = Ax + bu; (n−1) Kob detA, where Kob is the transmission coefficient of the the value γ = (−1) channel u → y. Obviously in the nonlinear case γ(x,u) has the same meaning, but in this case, it is a nonlinear function. Assuming condition γ(x,u) = 0 to be satisfied and substituting the obtained coefficients (16) and (23) into expression (7), we obtain a continuous control: u=

(−1)n+1 δ0∗ g − hT (x, u)x, γ (x, u) = 0, x ∈ x , u ∈ Iu , γ (x, u)

(25)

As a continuous control, (25), of course, is physically unrealizable. However, if on the left side of this expression the control u = u(x(t)) is replaced by the control uk , k = 1, 2, 3, …, then with a sufficiently small T, with virtually no loss of quality, the vector x(t) and control u(x) on the right sides of the above expressions can be replaced by x k and uk-1 , respectively, assuming u0 = u(0) = 0. In particular, the expression (25) in this case will have the form: uk =

(−1)n+1 δ0∗ gk−1 − hT (xk , uk−1 )xk , γ (xk , uk−1 ) = 0, xk ∈ x , uk ∈ Iu , γ (xk , uk−1 )

(26)

where, with accordance to (24),



γ (xk , uk−1 ) = cT (xk ) adjA(xk ) b(xk , uk−1 ), k = 1, 2, 3, . . .

(27)

The discrete control, defined by expressions (26), (27), is obviously physically realizable by some digital control device. In this case, the resulting nonlinear control system (1), (26), (27) is practically continuous, since rather small values of the sampling period T are usually required, which are typical for solving non-rigid differential equations on a computer using standard programs. It follows from the above expressions that the main feature of the he nonaffine systems (8) synthesized by the algebraic polynomial-matrix method based on the approach proposed here is the constancy of the coefficients of the Hurwitz characteristic polynomial D(p,x,u) of the system matrix D(x,u) (9). As noted above, for the stability of the equilibrium position of such nonlinear systems at x(t,x 0 ) ∈u , u ∈ Iu , x 0 ⊂ 0 ∈ u ∈ Rn it is sufficient that all the roots of the indicated Hurwitz polynomial be real and different. Let us show the effectiveness of the proposed approach using a numerical example.

4 Example of Control Design and Simulation Results Let’s design a control system for the direction of movement (course) of a surface vessel moving at a constant longitudinal speed. To change the course of the vessel, a hydrodynamic steering element is used [2]. The process of controlling the direction of movement of the vessel is described by the equations: x˙ 1 = x1 , x˙ 2 = u + u3 , y = x1 + 0.5x23 ,

(28)

Design of Hybrid Control System for Nonaffine Plants

215

where x 1 , x 2 are the state variables, u is the control, and y is the controlled variable (vessel’s course). In this case, if g = g0 1(t) then the condition ε0 = g0 -y0 = 0 must be satisfied, and the duration of transient processes when changing course should not exceed 2 s. Solution. This plant is nonaffine, since the control enters its equations in a nonlinear way, so we will use the proposed method. First, the quasilinear model (2), (3) of the plant (28) is found:    

01 0 , y = 1 0.5x x, (29) x˙ = x+ 2 00 1+u where x = [x 1 x 2 ]T and the controllability of the system (29) is checked under the conditions (5), (6):   0 1 + u2 = −(1 + u2 )2 , det U = det 1 + u2 0 It follows from this that the controllability condition (5) is obviously satisfied in the entire space Rn , i.e. the area ∈ u = R2 . It follows that the designed nonlinear system will have one equilibrium position x = 0, i.e. instead of the words “the equilibrium position of the system is stable or unstable” one can simply say “the system is stable or unstable”. By (10), (11) and (29) we determine the polynomials:   p −1 A(p) = det = p2 , 0 p     

1 + u2 p −1 0  = , adj(pE − A) b(u) = 1 + u2 p 0 p 1 + u2  

1 + u2  = 1 + u2 , V1 (u) = 1 0  1 + u2 p    

1 + u2   (30) = 1 + u2 p, V2 (u) = 0 1 2 1+u p and the value     

0 −1 0 2 γ (x, u) = 1 0.5x22 = − 1 + u . 0 0 1 + u2

(31)

As can be seen, the value (x, u) = 0 for all x and u. Therefore, the design problem has a solution. In this case n = 2, therefore, in accordance with (7), the continuous control has the form: u(x) = h0 g − [h1 x1 + h2 x2 ].

(32)

To determine its coefficients, a second-order Hurwitz polynomial D∗ (p) = p2 + is formed, the roots of which satisfy conditions (12), and the difference R(p) =

δ1∗ p + δ0∗

216

A. Gaiduk et al.

p2 +δ1 p+δ0 −p2 = δ1 p+δ0 is found from (14). Based on the coefficients of polynomials (30) and R(p), the system (15) is formed:    ∗   h1 δ 1 + u2 0 = 0∗ . δ1 0 1 + u2 h2 The solution of this system leads to the expressions: h1 (u) = δ0∗ /(1 + u2 ), h2 (u) = + u2 ). . The expression (23) determines the coefficient h0 (x, u) = δ0∗ /(1 + u2 ), which allows using (32) to write down the continuous control u(x) = (δ0∗ g − δ0∗ x1 − δ1∗ x2 )/(1 + u2 ). This expression, taking into account the definition of ε = g-y and the second equality (28) y = x1 + 0.5x23 , can be represented as follows:     u(x) = δ0∗ + 0.5δ0∗ x22 − δ1∗ x2 / 1 + u2 . δ1∗ /(1

In this case, the implemented discrete control is described by the expression: 

 2 − δ∗ x δ0∗ εk + 0.5δ0∗ x2,k 1 2,k   uk = , k = 1, 2, 3, . . . . (33) 2 1 + uk−1 where εk = gk – yk . Let’s pay attention to the fact that the system was designed with the symbolic coefficients δj∗ . This makes it possible to study the dependence of the properties of control systems designed using the algebraic polynomial-matrix method, including non-affine plants, both on the sampling period T and on the values of the roots of the characteristic polynomial D(p,x,u) of the system matrix D(x,u) by simulation. In this case, to determine the coefficients δj∗ we use a correlation that connects the moduli of the real roots δj∗ of the Hurwitz characteristic polynomial D(p,x,u) of the ∗ of the synthesized system matrix with the desired duration of the transient processes tpp system:     minλ∗j  ≥ (5 ÷ 7)/ttr∗ , (34) ∗ = 2 s from which often used in the synthesis of linear control systems. For a given tpp    ∗ (33) it follows: minλj  ≥ (5 ÷ 7)/2 = 2.5 ÷ 3.5. The values of the roots adopted for modeling and the coefficients corresponding to them are given in Table 1.

Table 1. The values of the roots adopted for modeling Variant

λ∗1

λ∗2

δ0∗

δ1∗

1

–0.6

–2

1.2

2.6

2

–1.5

–3

4.5

4.5

3

–3

–6

18

9

The simulation was carried out in MATLAB under various initial conditions x 0 = [0 0]T ; x 0 = [2 -1]T ; x 0 = [-2 3]T , setting step action g = g0 1(t) at g0 = 0,1,5 and sampling period T ∈ [0.001; 0.02; 0.05] s. Some simulation results are presented in Figs. 1–3.

Design of Hybrid Control System for Nonaffine Plants

217

Fig. 1. Transient processes at small modules of roots.

Figure 1 shows transient processes for the first version of the roots of the polynomial D(p,x,u) and T = 0.05 s. The state variables of the system also have a damping character for other values of the vector of initial conditions x 0 and period T. As can be seen, in this case the nonaffine system is stable for all values of T, up to T = 0.05 s. The steady value of the controlled variable y is exactly equal to the setting value. This follows from Fig. 1c, since at x 2 = 0 this variable is in accordance with the second Eq. (28) y = x 1 . With an increase in the modules of the roots of the polynomial D(p,x,u), the stability margin of the designed system decreases. On Fig. 2. Transient processes are shown for the second variant of the values of the roots of the polynomial D(p,x,u) given in Table 1, as well as for different values of the vector x 0 and period T. As seen in Fig. 2, in this case the designed system remains stable only at T ≤ 0.02. In a stable system, the system error ε0 at g = g0 1(t) is also equal to zero, i.e. ε0 = 0. Based on the graphs shown in Fig. 1 and Fig. 2, we can conclude that the duration of transient processes t pp of stable nonaffine systems designed on the basis of the proposed approach, in the first approximation, corresponds to dependence (33). Indeed, according to this relation, in the first variant of the roots t pp = 3/0.6 = 5 s, and in the second t pp = 3/1.5 = 2 c. These values obviously correspond to the duration of transient processes, the graphs of which are shown in Fig. 1(a, b, c), Fig. 2a, and Fig. 2c. The conclusion that with an increase in the modules of the roots of the polynomial D(p,x,u) the stability margin of the system decreases is confirmed by the transient processes obtained with the third variant of the roots of the polynomial D(p,x,u), shown in

218

A. Gaiduk et al.

Fig. 2. Transient processes: a), b) free system; c) system with external influence.

Table 1. The corresponding graphs for x 10 = -2, x 20 = 1, g0 = 0 and T = 0.001 s are shown in Fig. 3.

Fig. 3. Transient processes at “large modules” of roots of the polynomial D(p,x).

Based on the simulation results obtained, the following conclusions can be drawn. The roots of the characteristic polynomial of the system matrix have the greatest influence on the stability of the designed nonaffine system. The smaller the modules of the roots of this polynomial, the greater the allowable sampling period.

Design of Hybrid Control System for Nonaffine Plants

219

5 Conclusion The proposed approach makes it possible to obtain stable Hurwitz control systems for nonaffine plants using the algebraic polynomial-matrix method for sufficiently small periods of discretization of the variables of the control plant and small modules of the roots of the characteristic polynomial of the matrix of a closed system in its quasilinear model. The area of attraction of the equilibrium position of a closed system is determined by the area of the state space of the plant, in which the condition of controllability of the quasilinear model of the plant is satisfied. In this case, the control is formed by the computing device on the basis of previous control values, as well as discrete values of the measured variables of the object and the setting action. The proposed approach can be applied in design of control systems for nonaffine objects for various purposes. In the future, it is proposed to extend the obtained results to the case of incomplete information about the state variables of control plants and plants with external disturbances. Acknowledgements. This work has been supported by the grant of the Southern Federal University No. SP02/S4_0708Prioritet_06.

References 1. Zhang, J., Zhu, Q., Wu, X., Li, Y.: A generalized indirect adaptive neural networks backstepping control procedure for a class of non-affine nonlinear systems with pure-feedback prototype. Neurocomputing 121(9), 131–139 (2013) 2. Pshikhopov, V.Kh., Medvedev, M.Yu., Gaiduk, A.R., Fedorenko, R.V., Krukhmalev, V.A., Gurenko, B.V.: Position-Trajectory Control System for Unmanned Robotic Airship. In: The 19th World Congress the International Federation of Automatic Control, pp. 8953–8958. IFAC Proceeding, Cape Town, South Africa (2014) 3. Zhou, J., Li, X.: Finite-Time Mode Control Design for Unknown Nonaffine Pure-Feedback Systems. Mathematical Problems in Engineering (2015) 4. Liu, Y.-J., Wang, W.: Adaptive fuzzy control for a class of uncertain nonaffine nonlinear systems. Inf. Sci. 177(18), 3901–3917 (2007) 5. Pshikhopov, V., Medvedev, M., Kostjukov, V., Houssein, F., Kadhim, A.: Trajectory planning algorithms in two-dimensional environment with obstacles. Info. Autom. 21(3), 459–492 (2022) 6. Pshikhopov, V., Medvedev, M.: Multi-loop adaptive control of mobile objects in solving trajectory tracking tasks. Autom. Remote. Control. 81(11), 2078–2093 (2020) 7. Pshikhopov, V., Medvedev, M.: Block design of robust control for a class of dynamic systems by direct lyapunov method. J. Mechan. Eng. Autom. 2, 154–162 (2012) 8. Pshikhopov, V., Medvedev, M., Kupovikh G., Shibanov V.: Position control of mobile robots with multi-contour adaptation. In: Proc. of 18th IEEE International Conference on Autonomous Robot Systems and Competitions (ICARSC). Torres Vedras, Portugal (2018) 9. Kristic, M., Kanellakopoulos, I., Kokoyovic, P.V.: Nonlinear and Adaptive Control Design. Wiley, New York (1995) 10. Petrov, B.N., Emel’yanov, S.V., Utkin, V.I.: Principles for the construction of invariant automatic-control systems with variable structure. Soviet Physics Doklady 154(6), 1294–1296 (1964)

220

A. Gaiduk et al.

11. Krasnova, S.A., Utkin, V.A., Miheev, Y.: Cascade design of state observers. Autom. Remote. Control. 62(2), 207–226 (2001) 12. Druzhinina, M.V., Nikiforov, V.O., Fradkov, A.L.: Methods of nonlinear adaptive control with respect to output. Autom. Remote. Control. 57(2), 153–176 (1996) 13. Pontryagin, L.S., Boltyansky, V.G., Gamkrelidze, R.V., Mikchenko, E.F.: Mathematical theory of the optimal processes. Fizmatgiz, Moscow (1961) 14. Boltjansky, V.G.: Mathematical methods of the optimal control. Nauka, Moscow (1969) 15. Pyatnickiy, E.S.: Controllability of lagrange systems with limited controls. Autom. Remote. Control. 12, 29–37 (1996) 16. Gaiduk, A.R.: Towards design of quasilinear Gurvits control systems. SPIIRAN Proceedings 18(3), 678–705 (2019) 17. Gaiduk, A.R.: Numerical design method of quasilinear models for nonlinear objects. Mekhatronika, Avtomatizatsiya, Upravlenie. 22(6), 283–290 (2021) 18. Gaiduk, A.R.: Nonlinear control systems design by transformation method. Mekhatronika, Avtomatizatsiya, Upravlenie 19(12), 755–761 (2018) 19. Gupta, S.C., Hasdorff, L.: Fundamentals of automatic control. NY Wiley (1970) 20. Gaiduk, A.R.: A polynomial design for nonlinear control systems. Autom. Remote. Control. 64(10), 1638–1642 (2003) 21. Gaiduk, A.R.: Analytical synthesis of controls for nonlinear objects of a certain class. Autom. Remote. Control. 54(2), 227–237 (1993) 22. Çimen, T.: State-Dependent Riccati Equation (SDRE) Control: A Survey. In: Proceedings of the 17th World Congress of the International Federation of Automatic Control, pp. 3761–3775 (2008) 23. Lin, L.-G., Vandewalle, J., Liang, Y.-W.: Analytical representation of the state-dependent coefficients in the SDRE/SDDRE scheme for multivariable systems. Automatica 59, 106–111 (2015) 24. Levine, W.S.: The Control Handbook: Control System Fundamentals, 2nd ed., CRC Press (2011) 25. Gantmakher, F.R.: Theory of Matrices. Nauka, Moscow (1988)

Sliding-Mode Control of Phase Shift for Two-Rotor Vibration Setup Nikolay Kuznetsov1,2

, Boris Andrievsky1,2 , Iuliia Zaitceva1,2,3(B) , and Elizaveta Akimova1

1 Department of Applied Cybernetics, Faculty of Mathematics and Mechanics, St. Petersburg

State University, 28, Universitetsky Pr., Stary Peterhof, St. Petersburg 198504, Russia [email protected] 2 Institute for Problems in Mechanical Engineering of Russian Academy of Sciences, 61, Bol‘Shoy Prospect V.O., St. Petersburg 199178, Russia 3 Department of Automatic Control Systems, St. Petersburg, Electrotechnical University “LETI”, 5, Professora Popova St., St. Petersburg 197376, Russia

Abstract. This paper successfully developed and studied a phase shift control system for a two-rotor vibration mechatronic setup, aiming to maintain the desired revolving speed of the rotors. The sliding mode motion was achieved by utilizing a relay controller in the phase loop, while PI controllers were employed in the velocity control loops. Through numerical study and simulations using the parameters of the Mechatronic Vibration Setup SV-2M, the effectiveness of the proposed velocity and phase shift control laws was demonstrated. The possibility of sliding mode occurrence in the phase shift loop was examined through analytical and numerical analysis, confirming its presence. The relationship between the relative degree of the transfer function of the plant and the possibility of the occurrence of a sliding mode is analyzed based on the locus of a perturbed relay system approach. Simulation results indicated that sliding mode motion appeared after a finite transient time and showcased the dynamical properties of the closedloop system. In conclusion, the findings of this study validate the efficacy of the phase shift control system in achieving the desired rotor speed and demonstrate the feasibility of implementing sliding mode motion in the mechatronic setup. Keywords: Phase Shift Control · Two-Rotor Vibration Mechatronic Setup · Induction Motor · Sliding Mode Motion · Relay Control · PI Control

1 Introduction Vibration technologies are applied in various industries and manufacturing sectors, including ore enrichment, metallurgy, mechanical engineering, chemical industries, production of construction materials, grinding, fine grinding, or surface treatment of various parts. The concept of the feedback control for system performance improvement with the help of a controller is fairly easy to implement as long as it is possible to obtain a satisfactory description of the model systems, there is no influence of nonlinearities or © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 221–232, 2023. https://doi.org/10.1007/978-3-031-43111-1_20

222

N. Kuznetsov et al.

the dimension of space is low states [1–7]. When designing a control system for a vibratory machine (VM), all these factors are present [8–15]. In addition, design parameters can vary. In this case, the use of machine learning has greatly simplified to achieve the set control goals. For a vibration machine, both classical control tasks such as start-up, regulation, and tracking, as well as specific tasks related to vibrations, such as passing through resonance speed, synchronization of vibrators, and modification of attractors, are relevant [4, 16, 17]. The start-up of vibration machine electric drives can be achieved in various ways, providing energy savings and reducing the power of electric motors. The output to the desired rotational speed of the motors can pass through the electromechanical resonance point (the Sommerfeld effect), the overcoming of which is a non-trivial task [8, 11]. In [8], it is shown how the resonant frequencies and oscillation modes change when the center of mass of the technological load is shifted. An algorithm based on the velocity gradient method for overcoming the resonant frequency was proposed in [11]. The main operating modes and control of a vibration system are start-up, passage through resonance, and synchronization. In this case, the movement of the vibration system in the vertical plane is considered, and deviations from planar parallel motion are neglected. To perform operations such as screening, crushing, and vibratory conveying of bulk materials to increase productivity, the synchronous rotation mode of the vibrators is used. The synchronous mode of operation, in the form of self-synchronization that occurs naturally, was discovered and studied by I.I. Blekhman [16, 17]. The conditions for stable operation in the mode of multiple synchronizations, based on the mass ratio and arrangement of rotors, were obtained in the works of I.I. Blekhman, N.P. Yaroshevich, and others. However, adherence to these conditions does not always guarantee stable operation of the vibrators in the multiple synchronization modes, so the development of new algorithmic approaches to solving this problem is a relevant task. The artificial mode of multiple synchronizations is implemented by different (multiple) rotational frequencies of the rotors, forming complex shapes and heterogeneous trajectory fields of the points on the vibrator [18]. This effect expands the technological possibilities of using vibration machines for performing challenging transportation tasks, such as the movement of dusty, wet, and sticky loads, as well as simultaneous screening and separation of bulk materials. Asymmetry of vibrations at multiple rotor speeds creates and enhances vibratory conveying with increased productivity. In [19], a criterion is derived for achieving multi-frequency synchronization and the ratio of the phase difference between exciters based on the approximate equation of motion of the system drive to improve the loosening of granular material. An analysis of the stability of synchronous operation was also carried out in [20], where a fuzzy PID controller was used. It was shown in [21] that the use of sliding control significantly improves the state of synchronization of the vibrating screen, which makes it possible to qualitatively separate cuttings from drilling fluid. Thus, the digitalization of manufacturing technology is a necessary step towards the transition to smart factories, which will subsequently increase productivity, and efficiency of use, and have a positive effect on the environment and the conservation of natural resources. To do this, this article continues the development and research of the phase shift control system of the two-rotor vibration mechatronic setup, maintaining

Sliding-Mode Control of Phase Shift for Two-Rotor Vibration Setup

223

the desired revolving speed of the rotors. The sliding mode motion is ensured using the relay controller in the phase loop, whereas the PI controllers are employed in the velocity control loops. Parameters of the Mechatronic Vibration Setup SV-2M of the Institute for Problems in Mechanical Engineering of RAS (IPME RAS) are taken for numerical study and simulations. The remainder of the paper is organized as follows. A description of the laboratory setup SV-2M is given in Sect. 2. Section 3 represents the proposed velocity and the phase shift control laws. Studying the possibility of sliding mode occurrence for the proposed control law is given in Sect. 4 both analytically and numerically. The simulation results are presented in Sect. 5. Section 6 summarizes the results obtained and outlines the intentions for future work.

2 Servo System Model of Two-Rotor Vibration Setup A vibration complex SV-2M is a nonlinear electromechanical system equipped by two induction motors (IM) with unbalanced rotors. The system operates in two modes: self-synchronization of the vibrators and controlled synchronization mode [22, 23]. All devices of the complex are interconnected into a closed-loop system in which mechanical processes and control processes take place. The dynamics model mechanical part of setup SV-2M can be derived from [24], based on the standard Lagrange formalism. But in [24] is assumed that the control actions are input torques. However, the torques cannot be either directly controlled by the outer equipment or even measured for the existing setup. For certain conditions that are supposed for the present study it is reasonable to use the simplified model, which reflects the motor gains and the principal time constants with a suitable precision, employing an approach of [6, 10, 23, 25], where it was obtained for the SV-2M that at low frequencies (down to 5 Hz), the significant impact of the gravitational (“pendular”) torque to the motor rotation is observed, which cannot be ignored for controller design [5, 24], but at medium and high-frequency ranges (5 – 20 Hz), the so-called “averaging property”, when the fast oscillating components are averaged and for revolving rotors only the “slow” motions can be taken into account [26, 27]. Besides, since the induction motors have their local feedback controllers, the dynamics of the drive systems, including the induction motor and the frequency converter with the feedback local controller can be approximately described by the second-order transfer function from the control signal to angular velocity ω [28–30] as follows: Wd (s) =

b0 kd ω = = 2 2 , u a0 s2 + a1 s + 1 T s + 2ξ Ts + 1

(1)

where b0 , a0 , a1 stand for the drive model parameters, where b0 = kd corresponds to the drive system static gain; T = (a0 )1/2 is the time constant; ξ = a1 (2T )−1 denotes the damping ratio; s ∈ C stands for the Laplace transform variable. Identification of motor model parameters was performed in [6, 15] based on the standard non-recursive least-square estimation (LSE) method [31], see details and results in [15].

224

N. Kuznetsov et al.

3 Robust Phase-Shift Control Algorithm Based on the approach of [15] and the speed-gradient (SG) method of [32, 33], let us use the following control law for control of rotation speed and the phase shift between the rotors for the two-rotor vibration setup: eωl = ωl∗ − ωl , eωr = ωr∗ − ωr ,

(2)

δ˙ωl = eωl , uωl = kI ωl δωl + kPωl eωl ,

(3)

δ˙ωr = eωr , uωr = kI ωr δωr + kPωr eωr ,

(4)

ω = ωl − ωr ,

(5)

    ul = sat0umax uωl − uψ , ur = sat0umax uωr − uψ ,

(6)

where uψ is the “phase-loop” control signal, produced by the following separate algorithm: ψ = ϕr − ϕl , eψ = ψ ∗ − ψ,

(7)

σ = eψ + τM ω,

(8)

ν˙ = γ1 sign(σ ),

(9)

 uψ = −satuψ (ν + γ sign(σ ) |σ |,

(10)

where τM is the reference model time constant; kI and kp are the rotation frequency controller gains; ω∗ denotes the desired rotation frequency, while ψ ∗ stands for the desired phase shift; γ is the relay controller parameter (the relay “shelf” level). The objective is to determine the existence and time of appearance of the sliding mode (SM) in the closed-loop system (1)-(10), considering the presence of two control channels, each involving servo-drives described by (1) with different parameters. The linear time-invariant (LTI) single-input-single-output (SISO) systems controlled using the relay algorithm [32–35]: x˙ (t) = Ax(t) + Bu(t), y(t) = GCx(t),

(11)

u = −γ sign(σ (t)) where x(t) ∈ Rn , u(t) ∈ Rn , σ (t) = GCx(t) ∈ R1 it was proved that a SM occurs on the surface σ (t) = 0, if transfer function W (s) = GC(sI −A)−1 B = N (s)/D(s) of D(s) (12)

Sliding-Mode Control of Phase Shift for Two-Rotor Vibration Setup

225

from control input u to measured output σ is a strictly minimal phase (SMP), i.e. if N (s) is the Hurwitz polynomial, the W (s) relative degree is ρ, defined as ρ = degD(s)−degN (s) is equal to 1, and γ > 0 is sufficiently large (with respect to a given region of initial conditions x(0)). Then for (12) the auxiliary control goal lim σ (t) = 0 is achieved. t→∞ The algorithm represented by (7)-(10) is referred to as the Implicit Reference Model (IRM) algorithm [32, 33, 35]. In the case under consideration, this mentioned model is expressed by the identity σ (t) = 0, and the applied control law, according to the SG method, ensures the decrease in |σ (t)|. The equivalence σ (t) = 0 means fulfillment of equation: ˙ + ψ(t) = ψ ∗ (t), τM ψ(t)

(12)

which can be called “the reference equation” by the analogy with the habitual reference model in Model Reference Adaptive Control (MRAC), as presented by [36].

4 Exploring Sliding Mode Potential 4.1 Analytical Study Firstly, let us check the SMP property for the system with control input uψ as in (6) and output σ, given by (9). To represent the system model in the state-space LTI form under the assumption that the saturations are not active, introduce drives state vectors xl = [ϕl , ωl , εl ]T ∈ R3 , xr = [ϕr , ωl , εl ]T ∈ R3 , where ϕl , ωl , εl denote rotation angles, angular velocities, and accelerations of the left (l) and right (r) rotors respectively; variables ul,r are taken as drive inputs u, and vectors ϕl,r , ωl,r ∈ R2 as drives outputs y. Evidently, the state-space representation of transfer function (1) leads to the following triples (A, B, C) in the state-space form: ⎡ ⎡ ⎤ ⎤

0 1 0 0 ⎦, Bd ,l = ⎣ ⎦, Cd ,l = 1 0 0 , Ad ,l = ⎣ 0 (13) 0 1 0 010 0 −1/a0,l −a1,l /a0,l b0,l /a0,l Similar (13) expressions have matrices for the right drive. Then the transfer functions for the left and right drives from ul,r to yl,r are Wdl (s) = Cdl (sI2 − Adl )−1 Bdl , Wdr (s) = Cdr (sI2 − Adr )−1 Bdr ,

(14)

respectively. Now let us consider the PI-controllers (2)-(4). To this end, let us introduce the controller gain matrices (row-vectors) G = [1, τM ] ∈ R1×2 and K = [kI , kP ] ∈ R1×2 . This leads to the following block matrices in the state-space form for the system with input uψ and output σ, defined as in (7)-(9):



03,3 Bd ,l Ad ,l − Bd ,l KCd ,l ,B = , (15) A= 03,3 Ad ,r − Bd ,r KCd ,r −Bd ,r 1 0 0 −1 0 0 C= . (16) 0 1 0 0 −1 0

226

N. Kuznetsov et al.

Substituting the matrices from (15) to (18) one gets the system matrices in the following expanded form: ⎡

0 0

⎢ ⎢ ⎢ − b0,1 kI ⎢ a0,1 A=⎢ ⎢ 0 ⎢ ⎣ 0 0 

1 0

0 1

1+b0,1 kP − a0,1

a1,1 − a0,1

0 0 0

0 0 0

B= 00

b0,l a0,l

0 0

b − a0,r 0,r

0 0 0 1 0

0 0 0 0 0 −

b0,r kI a0,r

 ,C =





0 0 0 0 1

1+b0,1 kP a0,r

1 0 0 −1 0 0 0 1 0 0 −1 0

a

1,r − a0,r

.

⎤ ⎥ ⎥ ⎥ ⎥ ⎥, ⎥ ⎥ ⎦

(17)

(18)

The state-space system representation with matrices (18), (19) gives the following transfer function from input uψ to output σ Wuσψ (s)

= GC(sI6 − A)

−1

B≡

Nuσψ (s) Duσψ (s)

,

(19)

where   Nuσψ (s) = τM a0,l b0,r + a0,r b0,l s4 + (a0,l b0,r + a0,r b0,l + τM (a1,l b0,r +   a1,r b1,l ))s3 + (a1,l b0,r + a1,r b1,l + τM a0,l b0,r + 2b0,l b0,r kP ) s2 + (b0,l + b0,r + 2b0,l b0,r kP + 2b0,l b0,r kI τM )s + 2b0,l b0,r kI , Duσψ (s) = a0,l a0,r s6 + (a0,l a1,r + .   a1,l a0,r )s5 + a1,l a1,r + a0,l (b0,r kP + 1 + (a0,r (b0,l kP + 1))s4 + (a1,l (b0,r kP + (b0,l kP + 1 + a1,l b0,r kI + 1) + (a1,r (bo,l kP + 1)+ (a0,l b0,r kI + + a0,r b0,l kI )s3 +  2 a1,l b0,l kI )s + b0,l kI b0,r kP + 1 + +b0,r kI b0,l kP + 1 s + b0,l b0,r kI2 ) The transfer function of the system under consideration is not SMP since its relative degree of ρ is equal to 2. However, under some conditions, [33, 37–43], this restriction can be relaxed, and the case ρ = 2 ensures achievement of the auxiliary control goal lim σ (t) = 0 if γ is sufficiently large. t→∞ In [33, 38] a significant contribution by introducing the concept of passifiability is made, which involves achieving passification via feedback for systems related to a specific output, even if it differs from the controlled output [35]. Results of [38] made a significant contribution by introducing the concept of passifiability [33, 38], which involves achieving passification via feedback for systems related to a specific output, even if it differs from the controlled output. This concept is known as extended passifiability (EP). Fradkov not only established the necessary and sufficient conditions for EP of LTI systems through linear output feedback but also demonstrated the connection between these results and high-gain stabilization, providing a framework for passifying systems by utilizing feedback control techniques, focusing on outputs that may not be directly controlled. By expanding the notion of passifiability to include such outputs, the concept of EP broadens the scope of control possibilities. In [38], the following closed-loop system model is considered x˙ (t) = Ax(t) + Bu(t), y(t) = Cx(t),

(20)

Sliding-Mode Control of Phase Shift for Two-Rotor Vibration Setup

u(t) = −Ky1 (t) + Lν(t),

227

(21)

where x ∈ Rn , u ∈ Rm , y ∈ R1 , and A, B, C are matrices of corresponding sizes, y1 = C1 x ∈ Rl1 denotes the second output, ν ∈ Rm is a new input, L ∈ m × m is a real matrix, and detL = 0. Hereafter the case of m = l = l1 = 1 is taken for brevity. For system (20), (21) the following notations are introduced [37]: D(s) = det(sIn − A) ≡ sn + dn−1 sn−1 + . . . + d0 , W (s) = C(sIn − A)−1 B ≡ N (s)/D(s), W1 (s) = C1 (sIn − A)−1 B ≡ N1 (s)/D(s), (s) = B1 (s)/B(s). Introduce the following F-property [36]: system (21),  all  (22) is said to possess the D(iω) F-property if either (Re(iω)= 0) for D(iω) D(iω) ω and Re N (iω) > 0 for Re(iω) = 0) or (Re(iω) ≤ 0 for all ω and Re N (iω) < 0 for Re(iω) = 0, i ∈ C), denotes the imaginary unit, i2 = −1. As stated in [38], if polynomial (s) is Hurwitz of degree n − 1, degN1 (s) = n − 2, the system possesses F-property and dn−1 = 0, then system (20) is strictly passifiable by output feedback (21). To find the necessary and sufficient conditions for the feasibility of high gain passification, the following Theorem is proved: Theorem 1 [38]. System (22) is strictly passifiable by output feedback (23) with all sufficiently large K if and only if polynomial N(s) is Hurwitz of degree n − 1 and either (A) degN1 (s) = n − 1, and the system possesses F-property, or (B) degN1 (s) = n − 2, the system possesses F-property and dn−1 = 0. Variant B of Theorem 1 gives the high-gain stabilizability condition for the relative degree ρ = 2 case of interest and, consequently, opens, in principle, an opportunity for the stable sliding-mode relay control as ρ = 2. The corresponding algebraic conditions in the terms of N(s), D(s) coefficients are given in [33]. Frequency conditions for the existence of an SM, obtained based on the describing functions (DF) method, are given in [37, 39]. [37] analyzed the relationship between the relative degree of a plant’s transfer function and the possibility of SM occurrence in relay systems using the locus of a perturbed relay system (LPRS) approach. The LPRS is defined in [41] as a characteristic of the response of a linear part to an unequally spaced pulse control of variable frequency in a closed loop. Based on the LPRS concept, [37, 42] shows that if ρ = 1 or ρ = 2, and the LPRS will not have intersection points with the real axis except for the origin the ideal SM occurs. In this case, the frequency of chattering is infinity, and the equivalent gain becomes infinity too. Namely, the following Theorem was proved. Theorem 2 [39]. If the transfer function (s) is a quotient of two polynomials (s) and (s) of degrees m and n, respectively, with non-negative coefficients, then for the existence of ideal SM, it is necessary that the relative degree ρ = n − m of (s) be one or two. If ρ = 1, then a conventional ideal SM can occur; if ρ = 2, then the so-called asymptotic second-order SM can occur. As stated in [39], Theorem 2 does not provide a sufficient condition, as a periodic motion of a finite frequency can exist even if the relative degree is one or two [37]. In [37], an example is demonstrated showing that if the LPRS intersects the real axis from below, returns then to the lower half-plane, and finally approaches the origin of the coordinates from below having the real axis as an asymptote, the SM appears in some

228

N. Kuznetsov et al.

vicinity of the quiescent state. In this case, the SM appears only if initial conditions are sufficiently small. 4.2 Numerical Study Let us analyze the system properties numerically for the specific values of the system model parameters, taken from [15]: b0,l = 0.0042s−1 , a0,l = 0.1187s2 , al,l = 0.8110 s, b0,r = 0.0043s−1 , a0,r = 0.1185s2 , al,r = 1.2195 s, τM = 1 s, kI = 240 s, kP = 1680, ω∗ = 40 rad/s for both drives; the natural bounds of control signals saturation are umax = 40000 in (6). To start with, consider the Proportional (P) control law as uψ (t) = −γ σ (t),

(22)

where γ > 0 denotes the controller gain. Then the LTI closed-loop system (20), (22) dynamics are described by the following equation   (23) Duσψ (p) + γ Nuσψ (p) σ (t) = 0, where p = d /dt denotes the time derivation operator. Transfer function (19) of the open-loop system has the following distributions of poles sP and zeros sZ : sP = {−508 ± 650i, − 335 ± 747i, − 0128, − 0127}, sZ = {− 421 ± 706i, − 10, − 0127}, in Fig. 1 (left plot). For γ = 2600, characteristic polynomial of closed-loop system (19), (22) has the following roots: sD = {−383 ± 152i, − 428 ± 71i, − 0783, − 0127}. Figure 1 illustrates the positions of zeros and poles for (19), (23) and the corresponding root locus for γ values ranging from 1 to 2600. The plots clearly indicate that as γ increases, four poles sD of the closed-loop LTI system converge towards the open-loop system zeros sZ , while the remaining two poles move towards infinity in modulus, aligned parallel to the imaginary axis in the left-hand complex plane.

Fig. 1. Zeros and poles positions for (19), (22) (left plot); root locus for (19), (22) (right plot), γ ∈ [1, 2600].

To further analyze the open-loop system (19) and (24) with γ = 2600, Fig. 2 presents the Bode diagram and Nyquist chart. These results affirm the asymptotic stability of the

Sliding-Mode Control of Phase Shift for Two-Rotor Vibration Setup

229

closed-loop system (19) and (24) for all γ > 0. Additionally, as follows from the DF method and the abovementioned results of [35, 39], the presence of the sliding mode (SM) in the closed-loop system with the relay controller uψ = −γ sign(σ ), γ > 0,

(24)

is also ensured. To confirm the conclusion made, let us simulate the complete system, including the plant model (19), tracking the reference action by the drives’ angular velocities and the relay law of phase shift control. The overall control law is (2)-(11). The “square-root” term in (10), inspired by the super-twisting method of [42, 43].

Fig. 2. Bode diagram (left plot) and Nyquist chart (right plot) for open-loop system (19), (22), γ = 2600.

As is seen from time histories of processes, depicted in Fig. 3, the controlled variables ψ(t), ω1 (t), ωr (t) tend to the desirable (reference) values ψ*, ω∗ , and after the short transient, the sliding mode occurs.

Fig. 3. System (19), (24), γ = 5000, ω∗ = 40 rad/s, ψ∗ = π. Time histories of σ(t), uψ (t) (left plot) and ψ(t), ω1 (t), ωr (t) (right plot).

230

N. Kuznetsov et al.

5 Conclusion This paper focuses on the development and study of a phase shift control system for a two-rotor vibration mechatronic setup. The sliding mode motion is achieved through the use of a relay controller in the phase loop, while PI controllers are employed in the velocity control loops. The possibility of sliding mode occurrence in the phase shift loop is examined through both analytical and numerical analysis. The simulation results confirm the appearance of sliding mode motion after a finite transient time and demonstrate the dynamical properties of the closed-loop system. Future work intends to refine and optimize the control system based on the obtained results and explore additional applications and enhancements for the two-rotor vibration mechatronic setup. The “square-root” multiplier in the control law, inspired by the supertwisting method of [39, 40] also deserves consideration in future studies. Acknowledgements. This work was supported in part by the St. Petersburg State University grant Pure ID 75207094, by the Leading Scientific Schools of the Russian Federation, project NSh-4196.2022.1.1.

References 1. Pati, C.S., Kala, R.: Vision-based robot following using PID control. Technologies 5(2), 34 (2017) 2. Aggogeri, F., Borboni, A., Merlo, A., Pellegrini, N., Ricatto, R.: Vibration damping analysis of lightweight structures in machine tools. Materials 10(3), 297 (2017) 3. Tomchin, D.A., Fradkov, A.L.: Control of rotor passing through the resonance zone on the basis of the high-speed gradient method. J. Mach. Manuf. Reliab. 5(55), 66–71 (2005) 4. Nonlinear problems of oscillation theory and control theory. Vibration Mechanics: In: Beletsky, V.V., Indeytsev, D.A., Fradkov, A.L. (eds.) Nauka, SPb (2009). (in Russian) 5. Panovko, G., Shokin, A.E.: Experimental analysis of the oscillations of two-mass system with self-synchronizing unbalance vibration exciters. Vibroengineering Procedia 18(3), 8–13 (2018) 6. Fradkov, A.L., Tomchina, O.P., Andrievsky, B., Boikov, V.I.: Control of phase shift in tworotor vibration units. IEEE Trans. Control Syst. Technol. 29(3), 1316–1323 (2021) 7. Zaitceva, I., Andrievsky, B.: Methods of intelligent control in mechatronics and robotic engineering: a survey. Electronics 11(15) (2022) 8. Gouskov, A., Panovko, G., Shokhin, A.: To the issue of control resonant oscillations of a vibrating machine with two self-synchronizing inertial exciters. In: Sapountzakis, E.J., Banerjee, M., Biswas, P., Inan, E. (eds.) Proc. 14th Int. Conf. on Vibration Problems 2019, LNME, pp. 515–526. Springer, Singapure (2021) 9. Fradkov, A., Tomchina, O., Galitskaya, V., Gorlatov, D.: Multiple controlled synchronization for 3-rotor vibration unit with varying payload. In: 5th IFAC Workshop on Periodic Control Systems, pp. 5–10. Elsevier, Amsterdam (2013) 10. Andrievsky, B., Boikov, V.I., Fradkov, A.L., Seifullaev, R.E.: Mechatronic laboratory setup for study of controlled nonlinear vibrations. In: 6th IFAC Workshop on Periodic Control Systems, pp. 1–6. Elsevier, Amsterdam (2010) 11. Fradkov, A., Gorlatov, D., Tomchina, O., Tomchin, D.: Control of oscillations in vibration machines: start up and passage through resonance. Chaos 26(11), 116310 (2016)

Sliding-Mode Control of Phase Shift for Two-Rotor Vibration Setup

231

12. Zaitceva, I., Andrievsky, B.: Real-time reinforcement learning of vibration machine PIcontroller. In: 6th Scientific School Dynamics of Complex Networks and their Applications, pp. 307–311. IEEE, New York (2022) 13. Andrievsky, B., Zaitceva, I.: Symmetrical control law for chaotization of platform vibration. Symmetry 14(11) (2022) 14. Zaitceva, I., Kuznetsov, N.V., Andrievsky, B.: Approach to identifying areas of uncontrolled oscillations in human-machine systems. In: International Russian Smart Industry Conference, pp. 196–201. IEEE, New York (2023) 15. Andrievsky, B., Zaitceva, I., Barkana, I.: Passification-based robust phase-shift control for two-rotor vibration machine. Electronics 12(4), 1006 (2023) 16. Blekhman, I.I., Yaroshevich, N.P.: Multiple modes of vibration maintenance of rotation of unbalanced rotors. Proceedings of the Academy of Sciences of the USSR. Machine Science 6, 62–67 (1986). (in Russian) 17. Blekhman, I.I.: Vibrational Mechanics. Phys.-math. lit, Moscow (1994). (in Russian) 18. Blekhman, I.I., Fradkov, A.L. (eds.): Control of Mechatronic Vibrational Units. Nauka, St.Petersburg (2001). (in Russian) 19. Li, L., Chen, X.: Multi-frequency vibration synchronization and stability of the nonlinear screening system. IEEE Access 7, 171032–171045 (2019) 20. Jia, L., Wang, C., Liu, Z.: Multifrequency controlled synchronization of four inductor motors by the fixed frequency ratio method in a vibration system. Sci. Rep. 13, 2467 (2023) 21. Zou, M., Fang, P., Hou, Y., Wang, Y., Hou, D., Peng, H.: Synchronization analysis of two eccentric rotors with double-frequency excitation considering sliding mode control. J. Commu. Nonlin. Sci. Numeri. Simul. 92, 105458 (2021) 22. Andrievsky, B.R., Blekhman, I.I., Blekhman, L.I., Boikov, V.I., Vasil’kov, V.B., Fradkov, A.L.: Education and research mechatronic complex for studying vibration devices and processes. Prob. Mecha. Eng. Reliab. Machi. 4, 90–97 (2016). (in Russian) 23. Boikov, V.I., Andrievsky, B., Shiegin, V.V.: Experimental study of unbalanced rotors synchronization of the mechatronic vibration setup. Cybernetics and Physics 5(1), 5–11 (2016) 24. Tomchin, D.A., Fradkov, A.L.: Control of passage through a resonance area during the start of a two-rotor vibration machine. J. Machi. Manuf. Reliabi. 36(4), 380–385 (2007) 25. Andrievsky, B., Fradkov, A.L., Tomchina, O.P., Boikov, V.I.: Angular velocity and phase shift control of mechatronic vibration setup. In: 8th IFAC Symposium on Mechatronic Systems, pp. 436–441. Elsevier, Amsterdam (2019) 26. Blekhman, I.I.: Vibrational Mechanics: Nonlinear Dynamic Effects, General Approach, Applications. World Scientific, Singapore (2000) 27. Blekhman, I.I.: Synchronization in Nature and Technology: Theory and Applications. ASME Press, New York (1988) 28. Khalil, H.K., Strangas, E.G., Jurkovic, S.: Speed Observer and reduced nonlinear model for sensorless control of induction motors. IEEE Trans. Control Syst. Technol. 17(2), 327–339 (2009) 29. Joshi, B., Chandorkar, M.: Two-motor single-inverter field-oriented induction machine drive dynamic performance. Sadhana 39(2), 391–407 (2014) 30. Giri, F.: AC Electric Motors Control: Advanced Design Techniques and Applications. John Wiley & Sons, Ltd., Chichester, UK (2013) 31. Ljung, L.: System Identification: Theory for the User. Prentice Hall, Upper Saddle River, NJ, USA (1999) 32. Fradkov, A.L., Miroshnik, I.V., Nikiforov, V.: Nonlinear and Adaptive Control of Complex Systems. Kluwer, Dordrecht (1999) 33. Fradkov, A.L.: Adaptive control in large-scale systems. Nauka, Moscow (1990). (in Russian)

232

N. Kuznetsov et al.

34. Andrievsky, B., Fradkov, A.: Implicit model reference adaptive controller based on feedback kalman-yakubovich lemma. In: IEEE Conference on Control Applications, pp. 1171–1174. IEEE, Glasgow (1994) 35. Andrievskii, B.R., Fradkov, A.: Method of passification in adaptive control, estimation, and synchronization. Autom. Remote. Control. 67(11), 1699–1731 (2006) 36. Landau, J.D.: Adaptive control systems. the model reference approach. Dekker, New York, NY (1979) 37. Boiko, I.: Analysis of modes of oscillations in a relay feedback system. In: American Control Conference, pp. 1253–1258. IEEE, Boston (2004) 38. Fradkov, A.: Passification of linear systems with respect to given output. In: 47th IEEE Conference on Decision and Control, pp. 646–651. IEEE, Mexico (2008) 39. Boiko, I.: Discontinuous Control Systems. Frequency-Domain Analysis and Design. Birkhäuser, Boston (2009) 40. Boiko, I.M., Kuznetsov, N.V., Mokaev, R.N., Akimova, E.D.: On asymmetric periodic solutions in relay feedback systems. J. Franklin Inst. 358(1), 363–383 (2021) 41. Boiko, I.: Input-output analysis of limit cycling relay feedback control systems. In: American Control Conference, pp. 542–546. IEEE, San Diego (1999) 42. Levant, A.: Sliding order and sliding accuracy in sliding mode control. Int. J. Control 58(6), 1247–1263 (1993) 43. Bartolini, G., Levant, A., Pisano, A., Usai, E.: Adaptive second-order sliding mode control with uncertainty compensation. Int. J. Control 89(9), 1747–1758 (2016)

GBMILs: Gradient Boosting Models for Multiple Instance Learning Andrei Konstantinov , Lev Utkin , Vladimir Muliukha(B) and Vladimir Zaborovsky

,

Peter the Great St. Petersburg Polytechnic University, 29, Polytechnicheskaya Street, St. Petersburg 195251, Russia [email protected]

Abstract. An approach based on using the gradient boosting machine for solving the Multiple Instance Learning (MIL) problem under condition of small tabular data is proposed. The MIL deals with labeled objects called bags which consist of several parts of the objects called instances with unknown labels and each bag label depends on the instance labels. Three modifications of the approach are developed and studied. They are determined by different aggregation functions which combine the intermediate predictions of the instance classes in each bag and allow gradient-based optimization through them. The modifications are based on the following aggregation functions: the Hard Max Aggregation, the Simple Attention Aggregation, and the Ensemble of gradient boosting machines in fusion with the Attention Neural Networks. The former two modifications can use an arbitrary decision tree gradient boosting model, which allows iterative training on loss gradients. The later modification simultaneously optimizes an ensemble of parallel gradient boosting models and the parameters of neural network. Numerical experiments with tabular datasets illustrate the proposed modifications of the approach and their superiority comparing to available MIL approaches accompanied by accuracy improvement up to 8%. Keywords: Multiple Instance Learning · Gradient Boosting Machine · Attention Mechanism · Neural Networks

1 Introduction The Multiple Instance Learning (MIL) is a framework in machine learning which deals with objects called bags such that each bag consists of some number of instances. An illustrative example of using the MIL is histopathology where the histology images are viewed as bags and separate patches as instances of the bag [1, 2]. A lot of applications can be formalized by means of the MIL, for example, the drug activity prediction [3], the protein function annotation [4], histopathology [2], etc. One of the tasks solved by the MIL is to classify new bags based on training data consisting of a set of labeled bags. To solve the task, it is assumed some relationship between labels of instances and bags, for example, positive bags contain at least one positive instance [5]. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 233–245, 2023. https://doi.org/10.1007/978-3-031-43111-1_21

234

A. Konstantinov et al.

A lot of MIL models solving the classification tasks have been proposed [6–8]. We take note of the models based on the attention mechanism [9–12]. These models are the most efficient, but they mainly use neural networks and deal with the image data. However, they may be ineffective when dealing with small tabular data. Therefore, our goal is to devise new methods for the MIL problem, leveraging recent developments in the area of neural networks. One of the powerful and efficient algorithms for dealing with small tabular data is the Gradient Boosting Machine (GBM) [13]. Its efficiency has been demonstrated in many applications. Therefore, we consider how GBM can be incorporated into MIL. As a result, we propose several modifications, called the Gradient Boosting MIL (GBMIL). The modifications are determined by different aggregation functions which combine the intermediate predictions of the instance classes in each bag. In particular, we propose three modifications based on the so-called Hard Max Aggregation (GBMIL-MAX), on the Simple Attention Aggregation (GBMIL-ATT), and the Ensemble of GBMs implemented by means of the Attention Neural Networks (GBMIL-ANN). The GBMIL-ANN can be regarded as a generalization of the GBMIL-ATT. Numerical experiments with well-known datasets Musk1, Musk2 [3], Fox, Tiger, Elephant [14] illustrate the proposed models. The above datasets have numerical features that are used to perform tabular data. The paper is organized as follows. A brief introduction to the MIL and the GBM can be found in Sect. 2. Ideas behind the GBM in the MIL are considered in Sect. 3. Three modifications of the Gradient Boosting MIL are provided in the same section. Numerical experiments illustrating the models are provided in Sect. 4. Concluding remarks can be found in Sect. 5.

2 Preliminaries 2.1 Multiple Instance Learning MIL is a weakly supervised learning problem because a structure of data in MIL is represented in a form of bags with the class labels such that each bag consists of instances (patches) which usually do not have class labels. For every application, there is a relationship between the instance labels and the corresponding bag label. One of the conventional relationships between the instance classes and the bag class is of the form: at least one positive instance makes the bag positive, and negative bags contain only negative instances [15]. Two tasks can be stated in MIL. According to the first task, we classify each instance to annotate it. The second task is to classify new bags by having a training set of bags. Formally, these tasks can be described as follows. There is a set D of N bags {X1 , ..., XN } with labels Yi = f (Xi ) ∈ {0, 1}, such that each bag, say Xi , consists of ni instances, (i) (i) (i) (i) i.e., Xi = {x1 , ..., xni }, where each instance xj consists of m features, xj ∈ Rm ,   (i) (i) ∈ {0, 1}. Here f and g are functions of the bag and and has a label yj = g xj instance classifiers, respectively. It is usually assumed that negative and positive classes are represented by 0 and 1, respectively. Labels Yi are known. In contrast to the bag level,

GBMILs: Gradient Boosting Models for Multiple Instance Learning

235

labels Yi are unknown, but the function g is determined during the training process. This is the first task of MIL. The second task is to annotate a new bag X or to find f (X). The dataset D can be represented as:  N (i) i D = {xk }nk=1 , yi . i=1

Taking into account the conventional relationships between the instance classes and the bag class, the function f is written as:     (i) i (i) i = max{g xk }nk=1 . f {xk }nk=1 To find parameters of the model f , we have to minimize the loss function: L=

N N      1   1   (i) i (i) i = . l yi , f {xk }nk=1 l yi , max{g xk }nk=1 N N i=1

i=1

2.2 A Brief Introduction to GBM The MIL is a weakly supervised learning problem because a structure of data in MIL is represented in a form of bags with the c Suppose that there is a dataset D = {(x1 , y1 ), ..., (xn , yn )}, where xi ∈ Rm , yi ∈ R for the regression and yi ∈ {0, 1, ..., C−1} for the classification. GBM [13] tries to iteratively learn a function F(x) on D, which predicts the target value y˜ of a new observation x, by minimizing the loss function: l(F) = E(x,y) l(y, F(x)). The iterative learning improves predictions by means of adding new base learners hi (x), forming an additive ensemble model of size T : F0 (x) = const, Ft (x) = Ft−1 (x) + γt ht (x), t = 1, ..., T , where ht is the i-th base model, for example, a decision tree; γt is the learning rate. The decision tree implementing hi is built at the i-th iteration on the dataset consisting of residuals (xk , rk ), k = 1, ..., n, where, for example, rk = yk − Fi−1 (xk ). (t) The residual at the t-th iteration ri for the i-th instance is defined as the partial derivative of the expected loss function:  ∂l(yi , z)  (t) . ri = − ∂z z=Fi−1 (xi ) Hence, the best gradient descent step-size γt is: γt = argmin γ

N 

l(yt , Ft−1 (xi ) + γ ht (xi )).

i=1

We study below how GBM can be incorporated into MIL.

236

A. Konstantinov et al.

3 Gradient Boosting for MIL The main motivation of applying the GBM with decision trees as base learners to the MIL is that decision trees are an inherently efficient algorithm for dealing with tabular data. Many MIL models, implemented by means of neural networks and successfully dealing with image data, cannot cope with tabular data in many cases. Therefore, we aim to devise new methods for the MIL problem by applying the GBM. Suppose that we have some differentiable loss function l. The classification function FT is written as follows: FT (x) =

T 

γ hq (x) = FT −1 (x) + γ hT (x),

q=0

where hq (x) is a regression tree defined as: hq (x) = argmin

h∈H

 ri =

N 

hq (xi ) − ri 2 ,

i=1

 ∂l(yi , σ (z))  −  ∂z

z=FT −1 (xi )

.

Here σ (z) is the standard sigmoid function. The GBM can be modified for the MIL to handle multiple input points as: Fˆ T (x1 , . . . , xk ) = G(FT (x1 ), . . . , FT (xk )), where G is a differentiable aggregation function. A new tree at each iteration is built to optimize the loss function for all bags: hq (x) = arg min

h∈H

where: (i) rk

ni N  

  (i) (i)2 hq xk − rk ,

i=1 k=1







∂l yi , σ G z1 , . . . , zni  =   ∂zk

  (i) zj =FT −1 xj

.

(i)

The expression for rk can be rewritten using the chain rule as:



 ∂G z1 , . . . , zni  ∂l(yi , σ (z))  (i) rk = .   ·    (i) (i)  ∂z ∂zk (i) z=F T −1 x1 ,...,xni zj =FT −1 x 

j

GBMILs: Gradient Boosting Models for Multiple Instance Learning

237

3.1 Aggregating Functions It can be seen from the above that an important question is how to choose an aggregating function G. These functions can be regarded as approximations of the maximum operation, which defines the relationship between labels of instances and the corresponding bag containing these instances. Several types of the aggregating functions G have been proposed in literature [6]. They are shown in Table 1 wherein the first column contains abbreviated names of the functions, the second column shows the functions themselves, and the third column contains the partial derivative of the functions. The functions are log-sum-exponential (LSE) [16], generalized mean (GM), noisy-or (NOR) [17], and ISR [17]. Some functions depend on the parameter α. We denote the vector (z1 , . . . , zn ) as z. It should be noted that there is the boosting implementation for the MIL that uses these approximations [17], but it builds weighted weak classifiers, i.e., the approach does not use the GBM framework, which produces the regression weak models. Therefore, we propose two additional aggregating functions: the Hard Max Aggregation (MAX) and the Simple Attention Aggregation (ATT), which are shown in Table 2 wherein I[·] is the indicator function. Table 1. Aggregating functions and the corresponding partial derivatives for the MIL. Partial derivatives

LSE

Functions G(z1 , . . . , zn )  n 1 ln 1  exp(αz ) i α n

GM



α−1 G(z) · (zni ) α

NOR

1 − ni=1 (1 − zi )

ISR

n  ζ zi 1+ζ , ζ = 1−zi i=1

Name

softmax(αz)

i=1

LSE

n 1  (z )α i n i=1



1

α

i=1 (zi )

1 (1 − G(z)) 1−z i  2 1 (1 − G(z)) 1−z i



n 1 1  exp(αz ) i α ln n i=1

softmax(αz)

In order to consider how to use the proposed aggregating functions in the GBM for the MIL, we first find the partial derivative of the loss function for computing residuals (i) rk . A natural loss function l for the classification task is the Binary Cross-Entropy represented as follows: l(y, z) = −(ylog(z) + (1 − y)log(1 − z)). The partial derivative of the loss function applied to the sigmoid of the GBM output is of the form:

238

A. Konstantinov et al. ∂l(y,σ (z)) ∂z

= −σ (z) · (1 − σ (z)) ·



y σ (z)



1−y 1−σ (z)



= −(y(1 − σ (z)) − σ (z) + yσ (z) = σ (z) − y.

.

Table 2. The proposed aggregating functions for the MIL. Name

Functions G(z1 , . . . , zn )

MAX

max{z1 , . . . , zn }

Partial derivatives

n I zi = argmax z i=1

ATT

zT · softmax(α · z)

(softmax(α · z)i · [1 + α · zi − αG(z)])i

3.2 Hard Max Aggregation Residuals for the MAX aggregating function can be calculated as follows:     (i) (i) rk = (σ (max z) − yi ) · I FT −1 xk = max z  =

  (i) σ (max z) − yi , if FT −1 xk = max z, 0,

otherwise.

The above expression is equivalent to the instance-level loss calculated for one instance in a bag, which corresponds to maxFT −1 . However, it is important that we include points with zero residuals to a training set for building the base trees. It can be useful since these points may have already correct predictions which should not be changed by new trees. On the other hand, we can ignore these points during training to implement bagging-like algorithm, where a tree is built at each iteration on a new data subset and support positive bags with more than one positive instance. It yields the new Hard Max GBM modification. 3.3 Simple Attention Aggregation The idea behind attention aggregation without trainable weights is to compute maximum as a weighted sum of input values, where weights show how relatively large each component is. Particularly, when α → +∞, we can get:  zk · (softmax(α · z))k → zargmaxz . z T · softmax(α · z) = α→+∞

k

The ATT residuals are obtained as follows: rki = (σ (G(z)) − yi ) · (softmax(α · z)k · [1 + α(zk − G(z))]), where:

 ni  (i) , G(z) = zT · softmax(α · z). z = FT −1 xj j=1

GBMILs: Gradient Boosting Models for Multiple Instance Learning

239

We can analyze the aggregation by considering the following difference:  zk − G(z) = zk − zq · softmax(αz)q q

= zk −



zq · wq = zk (1 − wk ) −



zq · wq .

q=k

q

Let us compare it with −1/α. Suppose that zk − G(z) less than −1/α. Then we can write for k : zk = max z as follows: zk (1 − wk ) −

 q=k

1 zq · wq ≤ − , α

⎛ ⎞   1 zk ⎝ wq ⎠ ≤ zq · wq − , α q=k

zk ≤



q=k

zq · 

q=k

wq p=k wp

1 . α · (1 − wk )



It is guaranteed that there holds: 

zq · 

q=k

wq p=k

wp

≤ zk .

Hence: zk ≥ zk −

 wq 1 1 ≥ . zq ·  − α · (1 − wk ) w α · − wk ) (1 p p=k q=k

The above is equivalent to: 1 zk − G(z) ≥ − , α when wk < 1, otherwise we obviously obtain: 1 zk − G(z) = 0 > − . α For any other values, even for: k : zk = min z,

 q=k

zq · 

wq p=k

wp

≥ zk ,

value (zk − G(z) − 1/α) may be smaller or greater than zero. This implies that the component residual may be even inverted.

240

A. Konstantinov et al.

It should be noted that the ATT aggregation is a generalization for average MIL pooling, when α = 0. Indeed, there holds:  ˆ α=0 (z) = zT · softmax(0) = 1 zk . G n k

Hence, residuals are of the form: (i)

rˆk =

  1 ˆ σ Gα=0 (z) − yi . ni

If α → +∞, then we obtain the MAX aggregation. Thus, the ATT aggregation allows us to go smoothly from one pooling to another one by changing α. 3.4 Ensemble of GBMs and the Attention Neural Networks We propose to replace the introduced ATT aggregating function with a trainable attention neural network. Similarly, to the multiclass classification, an ensemble of GBMs can be built to optimize a common classification loss. In this case, the loss function incorporates the attention-based aggregation of the instance embeddings, where each embedding is obtained by stacking GBM predictions for the corresponding instance. An ensemble of GBMs can be represented as the following set:   (1) (E) FT = FT (x), . . . , FT (x) , where E is the embedding dimension. The ensemble prediction is the vector:   (1) (E) F T (x) = FT (x), . . . , FT (x) . The aggregation model A takes the instance embedding vectors obtained with F T and transforms them to a bag label prediction:      (i) ∈ [0, 1]. A F T x1 , . . . , F T xn(i)i (i)

(i)

Let e1 , . . . , eni be embedding vectors. Then the Attention Neural Network (ANN) is defined as:   (i) ANN e1 , . . . , en(i)i = ωe + β, where: e=

ni  k=1

(i)

αk · ek , α = softmax

  ni  (i) , score ek , v k=1

and v ∈ RE is a template trainable vector of parameters, ω, β are parameters of the linear neural network layer, projecting an embedding to the one-dimensional space, and “score” is a scoring function, e.g., score(a, b) = aT b.

GBMILs: Gradient Boosting Models for Multiple Instance Learning

241

The proposed ANN aggregation is a generalization of the ATT aggregation. Indeed, if E = 1, v = (α) and ω = 1, then these aggregating functions are equivalent. In addition, this method like Attention-based Deep MIL [18] can be regarded as an embeddinglevel approach, which is preferable in terms of the bag level classification [18]. At the same time, attention weights α allow us to extract information about an instance which significantly affects the prediction and, therefore, to solve the well-known problem of interpreting the obtained predictions.

4 Numerical Experiments We compare GBMIL with the available MIL models by training all the models on widely used MIL datasets Musk1, Musk2 (drug activity) [3] and Fox, Tiger, Elephant [14]. Numbers of bags N , instances n in each bag and the number of features m in instances in the datasets are provided in Table 3. Table 3. A brief introduction about datasets for the MIL classification. Data set

N

n

M

Elephant

200

1391

230

Fox

200

1302

230

Tiger

200

1220

230

Musk1

92

476

166

Musk2

102

6598

166

The following MIL models are used for comparison: mi-SVM [14], MI-SVM [14], MI-Kernel [19], EM-DD [20], mi-Graph [21], miVLAD [22], miFV [22], mi-Net [23], MI-Net [23], MI-Net with DS [23], MI-Net with RC [23], Attention and Gated-Attention [9]. We compare these models with three modifications of GBMIL: GBMIL-ATT, GBMIL-MAX, and GBMIL-ANN. In experiments, we apply Extremely Randomized Trees (ERT) for initialization because they provide better results. According to the ERT algorithm, a split point randomly selected for each feature and then the best split among these is selected [24]. We also use in experiments: sigmoid function with the trainable temperature parameter ω, which is initialized with 10, as indicator approximation; softmax operation with the trainable temperature parameter τ , which is is also initialized with 10; the number T of decision trees is 20; the largest depth h of trees is 5; the dimension E of each embedding vector is 4; the number of epochs is 2000; the batch size is 20 and the learning rate is 0.01. The mean and standard deviation as accuracy measures are computed by using 5-fold cross-validation. Numerical results for datasets Elephant, Fox and Tiger are shown in Table 4.

242

A. Konstantinov et al.

Table 4. Accuracy measures (the mean and standard deviation) for comparison of the MIL classification models and proposed modifications by using datasets Elephant, For and Tiger. Model

Elephant

Fox

Tiger

mi-SVM

0.822 ± N/A

0.582 ± N/A

0.784 ± N/A

MI-SVM

0.843 ± N/A

0.578 ± N/A

0.840 ± N/A

MI-Kernel

0.843 ± N/A

0.603 ± N/A

0.842 ± N/A

EM-DD

0.771 ± 0.097

0.609 ± 0.101

0.730 ± 0.096

mi-Graph

0.869 ± 0.078

0.620 ± 0.098

0.860 ± 0.083

miVLAD

0.850 ± 0.080

0.620 ± 0.098

0.811 ± 0.087

miFV

0.852 ± 0.081

0.621 ± 0.109

0.813 ± 0.083

mi-Net

0.858 ± 0.083

0.613 ± 0.078

0.824 ± 0.076

MI-Net

0.862 ± 0.077

0.622 ± 0.084

0.830 ± 0.072

MI-Net with DS

0.872 ± 0.072

0.630 ± 0.080

0.845 ± 0.087

MI-Net with RC

0.857 ± 0.089

0.619 ± 0.104

0.836 ± 0.083

Attention

0.868 ± 0.022

0.615 ± 0.043

0.839 ± 0.022

Gated-Attention

0.857 ± 0.027

0.603 ± 0.029

0.845 ± 0.018

GBMIL-ANN

0.880 ± 0.041

0.685 ± 0.082

0.925 ± 0.027

GBMIL-MAX

0.850 ± 0.050

0.695 ± 0.048

0.913 ± 0.030

GBMIL-ATT

0.875 ± 0.059

0.690 ± 0.080

0.925 ± 0.048

The best results in tables are shown in bold. Numerical results for datasets Musk1 and Musk2 are shown in Table 5. One can see from Table 4 that GBMIL models constantly outperform other available models: GBMIL-ANN has the accuracy higher by 1.2% than the most accurate available model for the Elephant dataset, GBMIL-MAX shows 6.5% accuracy improvement for the Fox dataset, GBMIL-ANN and GBMIL-ATT improve accuracy by 8% than the best available model for the Tiger dataset. It can be seen from Table 5 that the GBMIL models also outperform most available models on Musk1 and Musk2 datasets, showing the best result for Musk1 dataset.

GBMILs: Gradient Boosting Models for Multiple Instance Learning

243

Table 5. Accuracy measures (the mean and standard deviation) for comparison of the MIL classification models and proposed modifications by using datasets Musk1 and Musk2. Model

Musk1

Musk2

mi-SVM

0.874 ± N/A

0.836 ± N/A

MI-SVM

0.779 ± N/A

0.843 ± N/A

MI-Kernel

0.880 ± N/A

0.893 ± N/A

EM-DD

0.849 ± 0.098

0.869 ± 0.108

mi-Graph

0.889 ± 0.073

0.903 ± 0.086

miVLAD

0.871 ± 0.098

0.872 ± 0.095

miFV

0.909 ± 0.089

0.884 ± 0.094

mi-Net

0.889 ± 0.088

0.858 ± 0.110

MI-Net

0.887 ± 0.091

0.859 ± 0.102

MI-Net with DS

0.894 ± 0.093

0.874 ± 0.097

MI-Net with RC

0.898 ± 0.097

0.873 ± 0.098

Attention

0.892 ± 0.040

0.858 ± 0.048

Gated-Attention

0.900 ± 0.050

0.863 ± 0.042

GBMIL-ANN

0.905 ± 0.040

0.885 ± 0.045

GBMIL-MAX

0.833 ± 0.089

0.845 ± 0.048

GBMIL-ATT

0.863 ± 0.038

0.890 ± 0.038

5 Conclusion Three modifications of the MIL classification models (GBMIL-ANN, GBMIL-MAX, GBMIL-SM) have been proposed. The proposed models provide better quantitative results for tabular datasets in comparison with many existing MIL models, showing the accuracy improvement up to 8%. From the qualitative point of view, the proposed models take advantage by incorporation of Decision Trees ensembles, widely used for small tabular datasets and constantly evolving, which allows to leverage the recent developments in the regression field and apply them for MIL. The ideas behind the modifications can be regarded as starting points for developing a large class of the efficient MIL models which extend the introduced aggregation operations, especially the operations based on attention mechanisms. The development and the study of these models can be viewed as directions for further research. Acknowledgments. This work is supported by the Russian Science Foundation under grant 21– 11-00116.

244

A. Konstantinov et al.

References 1. Van der Laak, J., Litjens, G., Ciompi, F.: Deep learning in histopathology: the path to the clinic. Nature Medicine 27, 775–784 (2021) 2. Yamamoto, Y., Tsuzuki, T., Akatsuka, J.: Automated acquisition of explainable knowledge from unannotated histopathology images. Nature Communications 10, 1–9 (2019) 3. Dietterich, T., Lathrop, R., Lozano-Perez, T.: Solving the multiple instance problem with axis-parallel rectangles. Artificial Intelligence 89, 31–71 (1997) 4. Wei, X.S., Ye, H.J., Mu, X., Wu, J., Shen, C., Zhou, Z.H.: Multiple instance learning with emerging novel class. IEEE Trans. Knowle. Data Eng. 33 (2019) 5. Srinidhi, C., Ciga, O., Martel, A.L.: Deep neural network models for computational histopathology: A survey. Medical Image Analysis 67, 101813 (2021) 6. Babenko, B.: Multiple instance learning: Algorithms and applications. Technical report. University of California, San Diego (2008) 7. Cheplygina, V., de Bruijne, M., Pluim, J.: Not-so-supervised: A survey of semi-supervised, multi-instance, and transfer learning in medical image analysis. Medical Image Analysis 54, 280–296 (2019) 8. Yao, J., Zhu, X., Jonnagaddala, J., Hawkins, N., Huang., J.: Whole slide images-based cancer survival prediction using attention guided deep multiple instance learning network. Medical Image Analysis 65, 1–14 (2020) 9. Ilse, M., Tomczak, J., Welling, M.: Attention-based deep multiple instance learning. In: Proceedings of the 35th International Conference on Machine Learning, PMLR, Vol. 80, pp. 2127–2136 (2018) 10. Jiang, S., Suriawinata, A., Hassanpour, S.: Mhattnsurv: Multi-head attention for survival prediction using whole-slide pathology images, arXiv: 2110.11558 (2021) 11. Konstantinov, A., Utkin, L.: Multi-attention multiple instance learning. Neural Computing and Applications 34, 14029–14051 (2022) 12. Rymarczyk, D., Kaczynska, A., Kraus, J., Pardyl, A., Zielinski, B.: ProtoMIL: multiple instance learning with prototypical parts for fine-grained interpretability, arXiv:2108.10612 (2021) 13. Friedman, J.: Stochastic gradient boosting. Computational statistics & data analysis 38, 367– 378 (2002) 14. Andrews, S., Tsochantaridis, I., Hofmann, T.: Support vector machines for multipleinstance learning. In: Proceedings of the 15th international conference on neural information processing systems, NIPS’02, pp. 577–584. MIT Press, Cambridge, MA, USA (2002) 15. Carbonneau, M.A., Cheplygina, V., Granger, E., Gagnon, G.: Multiple instance learning: a survey of problem characteristics and applications. Pattern Recognition 77, 329–353 (2018) 16. Ramon, J., Raedt, L.D.: Multi instance neural networks. In: Proceedings of the ICML-2000 Workshop on attribute-value and relational learning, pp. 53–60 (2000) 17. Viola, P., Platt, J., Zhang, C.: Multiple instance boosting for object detection. In: Advances in neural information processing systems, Vol. 18, pp. 1–8 (2005) 18. Ilse, M., Tomczak, J., Welling, M.: Attention-based deep multiple instance learning. In: International conference on machine learning, PMLR, pp. 2127–2136 (2018) 19. Gartner, T., Flach, P., Kowalczyk, A., Smola, A.: Multi-instance kernels. In: Proceedings of ICML, Vol. 2, pp. 179–186 (2002) 20. Zhang, Q., Goldman, S.: Em-dd: An improved multiple-instance learning technique. In: Proceedings of NIPS, pp. 1073–1080 (2002) 21. Zhou, Z.H., Sun, Y.Y., Li, Y.F.: Multi-instance learning by treating instances as non-iid samples. In: Proceedings of ICML, pp. 1249–1256 (2009)

GBMILs: Gradient Boosting Models for Multiple Instance Learning

245

22. Wei, X.S., Wu, J., Zhou, Z.H.: Scalable algorithms for multi-instance learning. IEEE Trans. Neur. Netw. Learn. Sys. 28, 975–987 (2017) 23. Wang, X., Yan, Y., Tang, P., Bai, X., Liu, W.: Revisiting multiple instance neural networks. Pattern Recognition 74, 15–24 (2018) 24. Geurts, P., Ernst, D., Wehenkel, L.: Extremely randomized trees. Machine learning 63, 3–42 (2006)

Approach to Numerical Solution of Nonlinear Optimal Feedback Control Problems Samir Guliyev1,2(B) 1 Institute of Control Systems, 68, B. Vakhabzadeh street, Baku AZ1141, Azerbaijan

[email protected] 2 Azerbaijan State Oil and Industry University, 20, Azadlyg avenue, Baku AZ1010, Azerbaijan

Abstract. In the paper, we consider optimal feedback control problems for dynamic, in the general case, nonlinear systems with lumped parameters based on continuous and discrete feedback on the object’s state. To calculate the values of the feedback control’s parameters, we propose to use the measured values of observable components of the phase vector or the object’s output at the current and some previous (past) moments of time in order to compensate for the inability to measure all the components of the object’s phase state. As a result of this kind of formation of the dependence of the parameters of the synthesized control on a part of the object’s state, the process under consideration will be described by ordinary differential equations with time-constant delay arguments in the phase state. The feedback control problem is solved numerically by reducing it to a finite-dimensional optimization problem. To this end, we derive formulas for the gradient of the objective functional of the reduced problem with respect to the optimizable parameters are zonal values of the feedback parameters. These formulas make it possible to formulate necessary first-order optimality conditions, as well as to use them for numerical solution to model problems using first-order optimization methods. Keywords: Feedback Control · Gradient of Functional · Numerical Optimization Methods · Delay Differential Equations · Zonal Control Actions

1 Introduction Of constant practical interest among control problems are the problems of synthesizing optimal controls for complex technical objects, for example, manipulation robots, rockets, etc. Feedback control problems within the framework of automatic control and regulation systems have been studied by many authors [1–6]. Interest in this class of problems has increased over the past decades due to the development of technical, computational, and measuring tools for monitoring and controlling technical objects and technological processes. In most cases, linear systems were considered, and in the case of nonlinear systems, the corresponding linearized systems were used. For linear objects with relatively small dimensions of the phase and control vectors, effective state feedback control methods have been developed, which have been widely used in practice in © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 246–257, 2023. https://doi.org/10.1007/978-3-031-43111-1_22

Approach to Numerical Solution

247

automatic control and regulation systems, as well as in intelligent measuring systems. The works [7, 8] present fairly compact and complete surveys of major results in the field of feedback control and regulation systems. Note that an important role in the development of feedback control systems and their widespread practical implementation belongs to modern technical means of measuring and computing technology, which make it possible to carry out a large amount of measuring and computing work in real time. The main differences and advantages of the approach proposed in this paper to constructing a feedback control system from the known classical methods are as follows. The approach makes it possible to study objects with both linear and nonlinear dynamics from a unified standpoint. The approach is robust (stable) to errors (noises) in measurements of the current object’s state. This is explained by the fact that the parameters of the synthesized values of control actions depend not on the current measured values of the object’s state but on the subset (zone) of the set of possible states to which the current state belongs. Feedback control parameters (coefficients of dependence of the control on the state) in each zone are constant and determined from the optimality condition of the given control quality functional. The cases of discrete and continuous feedback are studied, for each of which a piecewise-constant and linear functional dependence of the parameters of zonal control actions on the object’s state or output is considered. The constancy of the parameters of zonal control actions (zonal controls themselves or zonal amplification coefficients with a linear dependence of the control on the object’s state or output) ensures the robustness of the control system and feasibility of synthesized control actions with a sufficiently high accuracy and improves technical performance of the equipment involved in the control loop. Examples of objects for which the considered classes of control actions are the most appropriate are pumping and compressor stations in pipeline transport systems of hydrocarbon raw materials, wells in oil and gas fields, and many other technological processes and technical objects. On the one hand, a frequent operational change in the modes of controllable actions for these objects is practically impossible, and on the other hand, it leads to rapid wear of the equipment. For each class of control actions considered, we obtain formulas for the gradient of the objective functional with respect to the optimizable parameters of synthesized controls. These formulas make it possible to formulate the necessary optimality conditions and to construct numerical solution schemes based on first-order iterative optimization methods. We present the results of numerical experiments obtained by solving test problems.

2 Problem Statement and Solution Scheme Here we formulate the feedback control problems for nonlinear objects with lumped parameters with various types of feedback on the input and output of the object and on various classes of zonal control functions with inaccurately given information on the values of the object’s parameters. Let the controllable object (process) be described by the following nonlinear system of differential equations: x˙ (t) = f(x(t), u(t), p), t ∈ (0, T ].

(1)

248

S. Guliyev

Here x(t) ∈ Rn denotes the continuous, piecewise continuously differentiable function of the phase state; T is the duration of the control process; u(t) ∈ U is the r-dimensional vector of piecewise-continuous control actions, the domain of admissible values of which is a closed convex set U ⊂ Rr ; p is the m-dimensional vector of the object’s parameters constant in time, the exact values of which are not known, but they can take values from some predetermined set P; there is given the density (weight) function of the assumed values defined on P: (2) The vector function f(x, u, p) is continuously differentiable with respect to the first two arguments and continuous with respect to the third argument. Control of the process (1) is carried out taking into account the presence of feedback from the object, while the state vector x(t) can be measured in whole or in part. Assume that observations of the state of the object can be carried out continuously or at discrete moments of time. Further assume that the initial state of the object x0 = x(0) is not specified exactly,   but there is a set X0 of possible initial states with the density (weight) function ρX0 x0 defined on X0 : (3) For each given specific initial condition x0 ∈ X0 and specific value of the parameters p ∈ P, we estimate the quality of control over the time interval [0, T ] using the following functional: (4)   where x(t) = x t; x0 , p, u is the solution to the system (1) under initial conditions x(0) = x0 ∈ X0 , control u(t) ∈ U, and the values of the parameters p ∈ P. The process completion time T can be either a given value, as in the case of functional (4), or an optimizable value, as, for example, in optimal fast-action problems. Considering that the initial state and parameter values are not specified exactly but are determined within the accuracy of density functions on the corresponding sets, we evaluate the quality of control by the following average value of the functional (4) over all possible initial states x0 ∈ X0 and parameter values p ∈ P: (5) Note that during the operation of the object, it may be possible to measure the state of ∼ only part x (t) ∈ R of the components of the phase vector x(t),  ≤ n, continuously in ∼

time. Denote by X ⊆ Rn and X⊆ R the set of all possible states of the object itself and ∼ the set of all possible states of the components x (t) of the object state x(t), respectively, whose dynamics is described by system (1) for different initial states x0 ∈ X0 , parameter values p ∈ P, and controls u(t) ∈ U for t ∈ (0, T ]. As a rule, for many processes, based ∼

on experience and the results of studying the functioning of objects, the set X or X of

Approach to Numerical Solution

249

possible states is known in advance. Control of the dynamics of process (1) is carried out ∼ taking into account the presence of feedback on the object’s current state x(t) or x (t), or on the object’s output y(t), which is determined by the nonlinear function of its state x(t): y(t) = G(x(t)), y ∈ Rν ,

(6)

where the ν-dimensional observation vector function G(x) is continuously differentiable with respect to each variable on the set X and, as a rule, ν ≤ n. We assume that process (1) is completely observable locally. Feedback (retrieval of information on the phase state or the object’s output) can be carried out both continuously for t ∈ (0, T ] and at discretely given moments of time τj ∈ [0, T ], j = 0, 1, 2, . . . , N . To form the current values of the control actions using feedback, the concept of “zonal control” is introduced. The values of control actions u(t) in the control process will be assigned as described below. Let Y ⊂ Rν be the set of values of the observable output vector y(·) from (6) for various possible values of the ∼

state x(·) ∈ X. Let us divide the sets Y and X into L disjoint open subsets (so-called zones) Yi ⊂ Rν and i ⊂ R , respectively, so that: L Y= Yi , Yi ∩ Yj = ∅, i = j, i, j = 1, 2, . . . , L, (7) i=1



X=

L i=1

i , i ∩ j = ∅, i = j, i, j = 1, 2, . . . , L,

(8)

where Yi and i are the closures of the sets Yi and i , respectively. The boundaries between any two adjacent (having a common boundary) zones Yi and Yj are defined by known continuous, almosteverywhere differentiable functions hi,j (y)   = −hj,i (y) = 0, and we assume that Yi ⊂ y : hi,j (y) < 0 or Yi ⊂ y : hj,i (y) ≥ 0 . For the first type of feedback control, the values of the control actions u(t) at the current moment of time in the process of controlling the dynamics of the object (1) will be assigned depending on which of the subsets Yi the observed current value of the output vector y(t) belongs to. For the second type of feedback control, to form the control actions u(t) at the current ∼ moment of time, in addition to the state value x (t), we use previously measured values ∼ x (t−τj ), j = 1, 2, . . . , s. The choice of delay times τj , j = 1, 2, . . . , s, is done depending on the rate of change of the object’s state; namely, at high speeds, it is desirable to choose the quantities τj sufficiently small. The choice of these quantities also depends on the accuracy of the measurements; namely, with low accuracy, they are chosen large enough so that the dynamics of the change in the object’s state would exceed the measurement accuracy. Then the value of the control at the current moment of time t is built as a linear ∼ dependence on the values of the components of the phase variable x (t) measured at the moments of time t and t − τj , j = 1, 2, . . . , s: ∼









u(t) = γ0i x (t) + γ1i x (t − τ1 ) + γ2i x (t − τ2 ) + · · · + γsi x (t − τs ), x (t) ∈ i , (9) where γ0i , γ1i , γ2i , …, γsi are constant matrices of dimension (r × ) defining the opti∼

mizable feedback parameters when the components x (t) of the phase state belong to

250

S. Guliyev

the ith zone i . Such controls will be called zonal. The problem of assigning to each zone Yi or i , i = 1, 2, . . . , L, control actions optimal in the sense of functional (5) will be called the feedback zonal control problem. We consider the following variants of the feedback zonal control problem, which differ in the nature of the observation and the type of dependence of the control on the observable state or output vector. Problem 1. There are given discrete moments of observation time ξj ∈ [0, T ], j = 0, 1, 2, . . . , N , such that ξ0 = 0and ξN =  T, at which it is possible to measure the value of the object’s output state y ξj = G x ξj ∈ Y. The frequency of these observations should be such, that while the object’s state belongs to any zone, at least one observation is made. If this condition is not met, the zones through which the trajectory y(t) did not pass from any starting point, as well as the zones in which no state measurements were made, will not be assigned of the zone control parameters. The control   the values values u(t) constant at t ∈ ξj , ξj+1 are determined depending on the last measured value of the observation vector for the object’s current output, namely, depending on which subset (zone) Yi , i = 1, 2, . . . , L, of the space Y the last measured (observed) output state belonged to. Thus, each subset Yi has its own constant control value:        (10) u(t) = ϑ i = const ∈ U, y ξj = G x ξj ∈ Yi , t ∈ ξj , ξj+1 .   If the observed output value y ξj belongs to the boundary of any zones, then we use the value of the zonal control of the adjacent zone into which the trajectory has passed. Problem 1 consists in finding admissible zonal control values ϑ i , i = 1, 2, . . . , L, according to (8) that optimize the value of the functional (5). The dimension of the optimizable finite-dimensional vector in this case is L × r. Problem 2. Control actions are determined by linear functions from the results of observing the parameters of the object’s output state at the given discrete moments of time ξj ∈ [0, T ], j = 0, 1, 2, . . . , N :          (11) u(t) = α1i · y ξj + β1i , y ξj = G x ξj ∈ Yi , t ∈ ξj , ξj+1 .   Here α1i is the r × ν matrix and β1i r-dimensional vector constant for t ∈ ξj , ξj+1 . Problem 2 consists in determining the zonal values α1i and β1i , i = 1, 2, . . . , L. The dimension of the optimizable vector is L × r × (ν + 1). Problem 3. Continuous measurements of the object’s output vector are carried out, and the control actions assume zonal values: u(t) = θ i = const ∈ U, y(t) = G(x(t)) ∈ Yi , t ∈ [0, T ], i = 1, 2, . . . , L.

(12)

Problem 3 consists in determining the admissible zonal control values θ i , i = 1, 2, . . . , L, that optimize the value of the functional (5). The dimension of the optimizable vector is L × r. Problem 4. Continuous measurements of the vector of observations of the object’s output state are conducted. The control is determined by a linear function of the measured output values: u(t) = α2i · y(t) + β2i , y(t) = G(x(t)) ∈ Yi , t ∈ [0, T ],

(13)

Approach to Numerical Solution

251

It is required to find admissible values α2i and β2i , i = 1, 2, . . . , L, that optimize the value of the functional (5). The number of optimizable parameters is r × L × (ν + 1). ∼

Problem 5. It is assumed in (9) that the values τj are sufficiently small, and for x (t) ∈  ∼  i , the condition x t − τj ∈ i is satisfied for at least one index j ∈ {1, 2, . . . , s}. To build the subsets i constructively, we proceed as described below. Let each of the ∼ components x˜ s (t) of the phase vector x (t) under all possible initial conditions x0 ∈ X0 , parameters p ∈ P, and controls u(t) ∈ U belong to some finite interval: (14) ωk ≤ x˜ k (t) ≤ ωk , t ∈ [0, T ], k = 1, 2, . . . , .   We divide each of the intervals ωk , ωk by points xk,j , j = 0, 1, .., k , such that xk,0 = ωk and xk,k = ωk , into k subintervals. Let us introduce the notation for -dimensional parallelepipeds:

(15) i = x ∈ R : xk,ik −1 ≤ xk ≤ xk,ik , ik = 1, 2, . . . , k , k = 1, 2, ...,  ,









where i = (i1 , i2 , . . . , i ) is the m-dimensional multiindex. Substituting the feedback ∼ control in (9) using the values of the observed components x (t) at the current and past moments of time into the system (1), we obtain:

∼ ∼ ∼ ∼ x˙ (t) = f x(t), γ0i x (t) + γ1i x (t − τ1 ) + γ2i x (t − τ2 ) + · · · + γsi x (t − τs ), p (16) ∼

for x (t) ∈ i , i = 1, 2, . . . , M , where M is the total number of zones. The system of differential Eqs. (16) is a system with constant delay arguments in the phase variable [13]. Thus, the original optimal control problem (1)-(5), taking into account the zonal feedback with part of the components ofthe phase variablein the form (9), is reduced to finding the feedback parameters γi = γ0i , γ1i , γ2i , . . . , γsi , i = 1, 2, . . . , M , that are (r × )-dimensional constant matrices, under which the solutions of the initial-value problems with respect to the system of delay differential Eqs. (16) with the set of initial 0 conditions  set of parameters P minimize the functional (5) depending on  1 2X and Mthe γ = γ , γ , . . . , γ only, i.e., F = F(γ, T ). The dimension of the resulting parametric optimal control problem is equal to r ×  × M × (s + 1), i.e., γ ∈ Rr××M ×(s+1) . All five formulations of the feedback zonal control problems considered above lead to finite-dimensional optimization problems. To solve the first, third, and fifth problems, in the case of a simple construction of the set of admissible controls U (for example, a parallelepiped, a ball, a polyhedron, etc.), it is effective to use first-order numerical optimization methods – gradient projection methods or conjugate gradient projection methods. If the domain of admissible controls has a complex boundary and the projection operator on it is not constructive, then to solve all five problems, one can use sequential unconditional optimization methods (for example, internal and external penalty functions methods) [14]. In the statements of the formulated problems, it is possible to consider the case when the time T is given, and it is also possible that the time T is optimizable. To construct iterative procedures based on numerical optimization methods, it is important to have exact formulas for the components of the gradients of the objective

252

S. Guliyev

functional in the space of optimizable parameters of the zonal controls. For this purpose, we derive formulas for the gradients of the functionals of the problems under consideration. The derivation of these formulas is based on the technique of calculating the increment of the objective functional obtained by incrementing an optimizable parameter, as well as the technique of obtaining the necessary first-order optimality conditions for discontinuous systems and systems with variable structure [15–17]. Using these results, we can prove theorems, each of which contains formulas for the components of the gradient of the objective functional with respect to the optimizable parameters in problems 1–5. Theorem 1. The components of the gradient of the objective functional on the class of controls (11) are determined by the formulas:

where i = 1, 2, ..., L; ψ(t; x0 , p, u), t ∈ [0, T ], is the solution to the adjoint initialvalue problem:

where δ(·) is the Dirac delta function. Theorem 2. For the  classes of feedback controls (10)–(13), if the process completion time T = T X0 , P is optimizable, then the derivative of the functional (5) with respect to the time T is determined by the formula:

  where ψ t; x0 , p, u is the solution of the adjoint problem corresponding to the case of the considered class of feedback controls. Note that to prove these theorems, we use either the technique of obtaining necessary first-order optimality conditions for discontinuous systems [15, 18] or the technique of obtaining necessary optimality conditions of the first order on the class of piecewiseconstant and piecewise-linear functions [9, 16]. To prove Theorem 2, one can use, for

Approach to Numerical Solution

253

example, the scheme applied in [2] for problems with a non-fixed value of the control process completion time. The quality of the control system with the use of zonal control actions is significantly affected by the structure of the division of the entire set of possible output values Y and ∼

the set of phase states X into zones Yi and i , i = 1, 2, ..., L, respectively. It is clear that increasing the number of zones due to their refinement can only lead to a decrease in the functional value. On the other hand, increasing the number of zones leads to the situation where control actions may change their values more often in time, and, therefore, the robustness property of the control system degrades, which often leads to rapid wear and failure of actuators. Conversely, increasing the size of the zones, i.e., decreasing their number, degrades the controllability of the object, and with a small number of them, the object may become completely uncontrollable. On the other hand, this increases the value of the objective functional, i.e., the quality of control deteriorates. Given the above, an important role is played by the choice of both the number and specifically the structure of the zones Yi and i , i = 1, 2, ..., L. The following approach is recommended for choosing a rational number of zones and determining the zones Yi and i , i = 1, 2, ..., L, themselves. For this purpose, the initial value L is selected, and any zones Yi and i , i = 1, 2, ..., L, that satisfy conditions (7) and (8) are assigned. Having solved the feedback control problem for any strategy of organizing the control, a comparative analysis of the obtained values of the parameters of control actions for all neighboring zones is carried out. If the parameters of the control actions in any adjacent zones differ by a sufficiently small value, then these zones can be combined into one, thus reducing the number L. If the control actions in adjacent zones differ significantly, then each of these adjacent zones should be divided, for example, into two zones, thus increasing the number L and again solving the feedback control problem. The increase in the number of zones should be carried out until the value of the objective functional ceases to change significantly.

3 Results of Numerical Experiments First, we consider the numerical solution to the feedback control problem for a nonlinear system with feedback on the object’s entire state. Example 1. Let us apply the obtained formulas to the construction of the considered variants of synthesized controls in the example of the following problem. The dimensions of the vectors are as follows: m = 1, i.e., p ∈ P ⊂ R; n = 2, i.e., x = (x1 , x2 ) ∈ R2 ; r = 1, i.e., u ∈ U = [−1, 1] ⊂ R. The system (1) has the following form:     x˙ (t) x2 (t) x˙ (t) = 1 = ,t > 0 (17) x˙ 2 (t) −sinx1 (t) − px2 (t) + u(t) The set P consists of three points:P = {2.9, 3.0, 3.1}. The set X0 consists of 80 points xj0 ∈ R2 , j = 1, 2, . . . , 80, uniformly distributed on a circle of radius 2.



 x T ; x0 , p, u , T = R · min(x12 T ; x0 , p, u + x22 T ; x0 , p, u − 0.0001;0 ]2 ,

254

S. Guliyev

R is a sufficiently large positive number; f 0 (x(t), u(t), p) ≡ 1. Then the objective functional to be minimized, averaged over all possible values of the parameters p ∈ P and initial points x0 ∈ X0 , for the problem under consideration will have the following form:



 1 80 3  0 T xj , pi , u +  x T ; xj0 , pi , u , T → min. (18) j=1 i=1 u∈U 240

Here, the function T xj0 , pi , u determines the time of bringing the object from the point F(u, T ) =

xj0 ∈ X0 to the given neighborhood of the origin for the value of the parameter pi ∈ P. The entire phase space of all possible states of the vector x(t) is divided into L = 72 zones. Zones are obtained using circles of radii 0.1, 0.5, 1.0, 1.5, and 2.0 and straight lines that divide each quarter of the x1 x2 plane into 18 parts. The zones are numbered counterclockwise, starting from the first quarter, in order of increasing radius, forming six rings. The initial control values in all zones were taken to be equal to 0. The value of the parameter R was successively increased from 102 to 106 . The optimization parameter values were chosen as follows: accuracies for one-dimensional minimization δ = 10−5 and for multidimensional minimization ε = 10−4 . The parameter ξ in problems 1 and 2 varied during numerical experiments. The results of solving the first four aforementioned problems are shown in Fig. 1. 1. In each zone, the control u takes a constant value. Thus, the vectors ϑ =  1 Problem ϑ , ϑ 2 , ..., ϑ 72 and T = (T1 , T2 , ..., T80 ) are optimizable in the problem. Observations of the process state are conducted at moments of time ξi = ξi−1 + ξ, ξ = const, i = 1, 2, . . . . Figure 1(a) shows the optimal trajectories of the system. Problem 2. In each zone, the control u depends linearly on the state of the system

1 , α 2 , . . . , α 72 at the current time. Thus, the optimizable vectors are α1,x1 = α1,x 1,x1 1,x1 1

  1 2 1 2 72 72 andα1,x2 = α1,x2 , α1,x2 , . . . , α1,x2 ,β1 = β1 , β1 , . . . , β1 . The state of the process was observed at the moments of time ξi = ξi−1 +ξ, ξ = const, i = 1, 2, . . . . Figure 1(b) shows the optimal trajectories of the system. Problem 3. In each zone, the control u takes a constant value. Thus, the vectors   θ = θ 1 , θ 2 , ..., θ 72 and T = (T1 , T2 , ..., T80 ) are optimizable in the problem. The state of the process was observed continuously. Figure 1(c) shows the optimal trajectories of the system. Problem 4. In each zone, the control u depends linearly on the state of the system

1 , α 2 , . . . , α 72 at the current time. Thus, the optimizable vectors are α2,x1 = α2,x 2,x1 2,x1 1

  1 , α 2 , . . . , α 72 , β = β1 , β2 , . . . , β72 . The state of the process and α2,x2 = α2,x 2 2,x2 2,x2 2 2 2 2 was observed continuously. Figure 1(d) shows the optimal trajectories of the system. Next, we consider the numerical solution to the feedback control problem for the nonlinear system (17) with feedback on the object’s output. Example 2. The object’s output state y(t) is observable, which is given by the following non-linear function: y(t) = G(x(t)) = x1 (t) + x2 (t) + x1 (t) · x2 (t), y(t) ∈ R.

(19)

Approach to Numerical Solution

255

Fig. 1. Phase portrait of optimal trajectories of the system (17) in problems 1–4 for the feedback state problem.

The sets P and X0 and the objective functional are the same as in the first example. For the set of possible values of the output y(t), the segment [−5.5, 5.5] was chosen as a result of numerical experiments on calculating the state of the system (17) for various admissible values x0 ∈ X0 , p ∈ P, and u ∈ U. The problem was solved numerically, with the set of possible output values Y divided into L = 112 zones. Zones are obtained by dividing the set by points {yi = −5.5 + i · 0.1, i = 0, 1, 2, ..., 110}, i.e., Yi = (−5.5 + 0.1 × (i − 1), −5.5 + 0.1 × i). The initial values of the control actions as well as the values of the parameters of the numerical methods involved in the solution to the problems were chosen the same as in Example 1. The results of solving the first four afore-mentioned problems are shown in Fig. 2 in the form of optimal trajectories of the system. It is of interest to compare the solutions of the feedback zonal control problem with the object’s output feedback and state feedback considered in the previous sections. In both problems, the object is described by the same system (17) with the same objective functional (18). As can be seen from Figs. 1(a–d) and 2(a–d), the optimal trajectories in all problems are quite the same, but nevertheless, for almost all possible initial points x0 ∈ X0 , the times of bringing the object to the neighborhood of the origin of the phase plane are different. The values of the objective functionals also differ; namely, the value of the functional with feedback on the object’s output exceeds the value of the functional with feedback on the phase state for all variants of the classes of control functions by approximately 10–15%. It should be noted, of course, that the values of the parameters of the zonal control functions in terms of output and state are completely different, as

256

S. Guliyev

are the chosen zones (subsets) Yi , i = 1, 2, ..., L, the sets of possible output values Y and the zones (subsets) Xi , i = 1, 2, ..., L, from the space of possible states X.

Fig. 2. Phase portrait of optimal trajectories of the system (17) in problems 1–4 for the feedback output problem.

4 Conclusion The paper has proposed an approach to the optimal feedback control of objects described by nonlinear systems of ordinary differential equations. The difference from the classical methods of optimal feedback control, in which the values of control actions depend on the current state of the process, is as follows: Here, the values of control actions depend on the subset (zone) of states to which the process’s current state belongs. The proposed approach is more resistant (stable) to errors in measurements of the current state. With respect to the parameters of the zonal values of the control actions, a finite-dimensional optimization problem is obtained, for whose numerical solution there are ready-made standard software packages. The results of carried out numerical experiments are presented on the example of solving some test problems.

References 1. Desoer, C.A., Vidyasagar, M.: Feedback Systems: Input-Output Properties 1. Academic Press, New-York (1975)

Approach to Numerical Solution

257

2. Bryson, A.E.: Applied Optimal Control, Estimation and Control, 1st edn. CRC Press, Boca Raton (1975) 3. Egorov, A.M.: Fundamentals of Control Theory. Fizmatlit, Moscow (2004). (In Russ.) 4. Yemelyanov, S.V., Korovin, S.K.: New Types of Feedback. Nauka, Moscow (1997). (In Russ.) 5. Ray, W.H.: Advanced Process Control. Butterworth, Stoneham (1989) 6. Quincampoix, M., Veliov, V.M.: Optimal control of uncertain systems with incomplete information for the disturbances. SIAM J. Control. Optim. 43(4), 1373–1399 (2004) 7. Polyak, B.T., Scherbakov, P.S.: Robust Stability and Control. Nauka, Moscow (2002). (In Russ.) 8. Polyak, B.T., Khlebnikov, M.V., Rapaport, L.B.: Mathematical Theory of Automatic Control. Lenard, Moscow (2019). (In Russ.) 9. Aida-zade, K.R., Guliyev, S.Z.: A task for nonlinear system control synthesis. Autom. Control. Comput. Sci. 39(1), 15–23 (2005) 10. Aida-zade, K.R., Guliyev, S.Z.: Zonal control synthesis for nonlinear systems under nonlinear output feedback. J. Autom. Inf. Sci. 47(1), 51–66 (2015) 11. Guliyev, S.Z.: Synthesis of control in nonlinear systems with different types of feedback and strategies of control. J. Autom. Inf. Sci. 45(7), 74–86 (2013) 12. Guliyev, S.Z.: Synthesis of zonal controls of nonlinear systems under discrete observations. J. Autom. Control Comput. Sci. Allerton Press 45(6), 338–345 (2011) 13. Bellen, A., Zennaro, M.: Numerical Methods for Delay Differential Equations, 1st edn. Oxford University Press, Oxford (2003) 14. Nocedal, J., Wright, S.: Numerical Optimization, 2nd edn. Springer, New-York (2006) 15. Aida-zade, K.R., Guliyev, S.Z.: On a class of inverse problems for discontinuous systems. Cybern. Syst. Anal. 4, 915–924 (2008) 16. Aida-zade, K.R., Handzel, A.V.: An approach to lumped control synthesis in distributed systems. Appl. Comput. Math. 6(1), 69–79 (2007) 17. David, E.S., Anitescu, M.: Optimal control of systems with discontinuous differential equations. Numerische Math. 114(4), 653–695 (2010) 18. Aida-zade, K.R., Guliyev, S.Z.: On numerical solution of one class of inverse problems for discontinuous dynamic systems. Autom. Remote Control Pleiades Publishing 73(5), 786–796 (2012)

Model-Based Policy Optimization with Neural Differential Equations for Robotic Arm Control Andrey Gorodetskiy1(B) , Konstantin Mironov1,2 , and Aleksandr Panov1,2,3 1 Center of Cognitive Modeling, Moscow Institute of Physics and Technology, 9, Institutskij

per., Dolgoprudny 141701, Russia [email protected] 2 Artificial Intelligence Research Institute, Moscow 105064, Russia 3 Federal Research Center “Computer Science and Control”, 9, 60-letiya Oktyabrya pr., Moscow 117312, Russia

Abstract. Applying learning-based control methods to real robots presents hard challenges, including the low sample efficiency of model-free reinforcement learning algorithms. The widely adopted approach to tackling this problem uses an environment dynamics model. We propose to use the Neural Ordinary Differential Equations to approximate transition dynamics as this allows for finer control of a trajectory generation process. NODE offers a continuous-time formulation that captures the temporal dependencies. We evaluate our approach on various tasks from simulation environment including learning 6-DoF robotic arm to open the door, which represents particular challenges for policy search. The NODE model is trained to predict movement of the arm and the door, and is used to generate trajectories for the model-based policy optimization. Our method shows better sample efficiency on this task comparing to the model-free and model-based baseline. It also shows comparable results on several other tasks. The application of NODE to model-based reinforcement learning enables more precise modeling of robotic system dynamics and enhances the sample efficiency of learning-based control methods. The empirical evaluation on various tasks demonstrates the efficacy of our approach, offering promising prospects for improving the performance and efficiency of real-world robotic systems. Keywords: Machine Learning · Deep Learning · Manipulation · Reinforcement Learning · Model-Based Control

1 Introduction Robotic manipulation is often connected to uncertainties and non-deterministic relations within the environment. The dynamics of the manipulators in deterministic and predictable conditions are well explored. Accurate control methods are efficient for these conditions. However, collaborative manipulators often have to operate with various objects of the non-deterministic environment with the unknown parameters. These operations may be executed using policies obtained via deep reinforcement learning (RL), which allows finding high-performance controller from interaction data [1, 2]. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 258–266, 2023. https://doi.org/10.1007/978-3-031-43111-1_23

Model-based Policy Optimization

259

Operating multilink manipulators is challenging for RL due to high number of degree of freedom in the system, high accuracy requirements and continuous-time model of the process [3]. In this work we aim to overcome these challenges by adding continuous-time environment model to policy-based control. There are several potential advantages of using the Neural Ordinary Differential Equations (NODE) [4] architecture for modeling the environment dynamics in modelbased policy optimization [5–7]. First, it can provide a more flexible and expressive model of continuous-time dynamics [8] and allows for the state trajectory to be described as a continuous function, which can better capture the complex relationships. Second, it can be used to model dynamics with varying time scales [9, 10]. Traditional discrete-time models, such as those based on recurrent neural networks, require the time step to be discretized into fixed-size intervals, which can lead to limitations in modeling dynamics with varying time scales. The NODE architecture has been shown to achieve high performance on a variety of tasks, including image classification and time-series prediction [4]. This suggests that it may be a promising approach for modeling environment dynamics in model-based reinforcement learning algorithms. We propose a new method for policy learning, where model-free RL algorithm is trained on samples from world model, represented by NODE. The method is designed for the challenging task of robotic door opening. Many previous works explored NODE in different settings, but very few considered the task of modeling action-conditioned dynamics [8, 11]. Our contribution consists in, first, use of NODE for modelling continuous-time dynamics of the system under the influence of control inputs, and, second, applying NODE-based action-conditioned dynamics model in model-based RL setting. The rest paper is organized as follows. In section II we briefly discuss existing works related to model-based policy optimization, neural ordinary differential equations and robotic door opening. Then in short section III we present background statement and notation. Section IV present a description of the proposed approach. In section V we discuss the results of numerical experiments, while section VI includes concluding remarks.

2 Related Work Model-based policy optimization is a specific approach to model-based reinforcement learning that focuses on optimizing the agent’s policy using a model of the environment [12–15]. Recent advances in model-based policy optimization have focused on improving the accuracy and efficiency of learned dynamics models, as well as developing new algorithms that can leverage these models to improve sample efficiency and overall performance. E.g. [16] uses model predictive control in model-based reinforcement learning setting, while [6] learns a world model and uses it for planning. In [17] authors propose MBPO algorithm that builds a model of the environment to optimize the policy in a sample-efficient manner by alternating between collecting data from the environment and optimizing the policy using the learned model. The algorithm has been shown to achieve state-of-the-art performance on a variety of continuous control tasks. [18, 19] trained one-step policy with the use of dynamics model, represented by Recurrent neural network (RNN).

260

A. Gorodetskiy et al.

The use of ODE framework for making model predictions gained massive attention after the work on Neural Ordinary Differential Equations [4]. This framework was used to build dynamics model for model-based RL algorithms [8], including continuous-time setting [7]. A scope of works focuses on modeling dynamics of mechanical systems [20, 21]. Modeling action-conditioned transition dynamics was explored in [11]. As an example use case for manipulation with requirement of dynamic models we consider the task of robotic door opening. The ability to open standard doors in living and office environments is useful for the mobility of robotic systems in human-intended environments. Door and robot may be considered together as a complex underactuated system. The kinematics and dynamics of such a system is represented by a complex model, involving forward kinematics of the robot and inverse kinematics of the door. This model may also include unknown parameters (e.g. size of the door). This property make door opening a theoretically interesting task for evaluating RL approaches. Therefore it was included into meta-world RL benchmark [22]. We may highlight the following significant works applying RL for this task. [23] apply parallelized Normalized Advantage Function Algorithm for training several real robots to solve door opening task. Another work on real-robot RL for door opening is [24], where door-opening task required 17.5 h of real world learning. Some other works consider learning for door opening task in simulation. [25] concentrate on active policy learning with image input and without exact reward models. Experiments on meta world door opening task [22] are provided, while for real world experiments other setups were chosen. [26] apply genetic algorithm to find appropriate parameters of the Deep Deterministic Policy Gradient (DDPG) algorithm. Aforementioned works consider model-free RL for door opening. As an example of utilizing model-based RL we can mention [27], which consider door opening as a sequence of separated tasks (handle approaching, handle rotation, etc.), where each task has own dynamics. The purpose of model learning was to define appropriate embeddings for various subtasks. In our work we consider door opening as an end-to-end task and aim to define learnable model of its continuous dynamics.

3 Background Reinforcement learning task is formulated as a Markov Decision Process (MDP). MDP is a tuple (S,A,τ ,ρ,o), where S – set of states, A – set of actions, τ – transition dynamics, ρ – reward function,  – observation function. Agent starts in environment in some initial state s and have access to corresponding observation o = (s). Then it chooses action a = π (a | s) following policy π, environment transitions to the next state s’ ∼ τ (s,a), and agent observes reward r = ρ(s’). The goal of agent is to find optimal policy π ∗ that maximizes the expected sum of rewards agent would observe during episode. Model-based reinforcement learning algorithms learn some model E of environment dynamics τ and then use it to search for policy of value functions in a sample-efficient manner, instead of trying to learn them directly from interaction data between agent and environment. In our work we use NODE as a dynamics model.

Model-based Policy Optimization

261

4 Method We use model-based policy optimization scheme similar to [17] to train policy entirely on synthetic samples from the environment model, while concurrently optimizing both from scratch (Algorithm 1).

Algorithm 1 General model-based policy optimization 1: Initialize policy π, dynamics model pθ, buffer for environment samples BE, buffer for model samples Bπ 2: for N epochs do 3: sample n transitions from environment under π; add to BE 4: Train model pθ on BE 5: for M model rollouts do 6: Sample ot uniformly from the last q records in BE 7: Perform k-step model rollout starting from st = g(ot) using policy π; add observed transitions to Bπ 8: end for 9: for G gradient updates do 10: Update policy parameters on batch from Bπ 11: end for 12: end for

For policy training, we use off-policy SAC [28] algorithm. Policy π is updated on data from buffer Bπ, which stores only transitions, sampled from the environment model E (Fig. 1). To generate trajectories from environment model, we set it into initial state, and generate a short rollout of fixed length using current policy. Initial states distribution is uniform distribution over the fixed-size chunk of buffer BE, which contains the most recent transitions observed from environment. To compute reward for predicted transitions, we use true reward functions for each env. To successfully learn a collection of co-dependent models, it’s important to balance their sampling and update rates. We fix these as hyperparameters. We build a model of environment dynamics using action-conditioned NODE. It consists of multilayer perceptron (MLP) fθ, observer g and differential equation solver (Eq. 1). MLP learns to predict state derivative ˙s provided state s and action a. Differential equation solver integrates fθ to obtain state trajectory from the initial state s0 and a sequence of actions: s˙ = fθ (s, a),

s = g(o)

(1)

Observer g is implemented by handcrafted functions that use a subset of agent observation’s features to reconstruct states from vector observation o and do not use latents. We use it to simplify prediction task by constructing exhaustively descriptive, minimal, interpretable representations separately for each environment. To interpolate action sequence during the integration of state dynamics we use zero-order hold. The model is trained using RMSprop by minimizing MSE between predicted and observed trajectories. All trajectories observed from environment are stored in buffer

262

A. Gorodetskiy et al.

Fig. 1. Sampling scheme.

BEM. We uniformly sample batches of subsequences of fixed length from this buffer to update the model on them. As these subsequences are typically short, we do not use adjoint method and compute gradient using automatic differentiation. Our method differs from previous works in the choice of dynamics model and training schedule. We use action-conditioned NODE to model environment dynamics. Instead of using encoder, we provide dynamics model with vector states, reconstructed from observations by environment-specific hand coded function, which has no trainable parameters. This function extracts relevant parts of observation which determine their own derivative and are more favorable to predict.

5 Experiments We train and evaluate the agent in multiple environments and deterministically evaluate it. Environments we use are based on Reacher-v2 from Gym [29], Cheetah-run from DMC [30], and DoorOpen from Metaword [22]. They are built on MuJoCo simulator. Our version of Reacher allows for joint position control to facilitate environment model learning, and DoorOpen uses modified reward function to facilitate policy learning. In all environments we assume that the observed quantities are given and we do not solve the problem of their estimation. In some of the environments our method exhibit critic divergence, which leads to considerable value overestimation and failure of the learned policy. We hypothesize this is due to specific sampling scheme used to draw transitions from the environment model. This scheme draws a batch of short rollouts from initial states that are chosen from the observed ones. Since the environment model is imperfect, the last state of each of these short rollouts has a considerable chance not to be encountered as a starting state

Model-based Policy Optimization

263

in any transition sampled from environment model. Hence, critic will not be updated in any of these “boundary” states that will accumulate “fake” value from critic updates in other points. The values of boundary states will propagate towards values of observed states, leading to catastrophic overestimation. We validate our hypothesis on simple 1-dimensional environment (Fig. 2) and observe that this effect actually takes place.

Fig. 2. Demonstration of parasitic value propagation from boundary states in SAC algorithm. Environment is a 1-dimensional segment where agent can walk. The farthest state from the start yields the maximal reward and is reachable in 3 steps. Average value of state value function spikes when the maximum number of steps allowed (l) is 3.

We train and evaluate NODE model on trajectory prediction task using Reacher environment. We collect 1000 trajectories of length 50 using uniform policy and train model on this fixed dataset (Fig. 3a). After 500 iterations, the joint angles prediction error, averaged over prediction horizon of 4 steps, drops below 1 degree. Then we evaluate our MBRL method in Reacher using the same dynamics model configuration (Fig. 3b). Results show that our method outperforms SAC in terms of sample-efficiency, as expected, while successfully solving the task. Dreamer-v3 shows lower performance on Reacher. We compare our method with Dreamer-v3 [31] in Cheetah environment (Fig. 4). We provide all compared methods with the same vector observations. In this environment our method learns faster on the budget of 10^6 environment samples. However, given additional samples, Dreamer-v3 outperforms our method. We evaluate the proposed method on the door opening task (Fig. 5) in DoorOpen environment. The agent observes position of the end-effector, as well as position and orientation of the handle. Our method demonstrates higher performance than Dreamer-v3. Dreamer-v3 is failed to solve the door opening task with default hyperparameters. In the challenging DoorOpen environment our method outperforms Dreamerv3 and solves the task. In the Cheetah environment our method shows inferior to Dreamer-v3 performance, but nevertheless learns successful gate.

264

A. Gorodetskiy et al.

Fig. 3. Method evaluation in the Reacher environment. (a, left)The prediction error of joint angles is consistently below 1 degree over a horizon of 4 steps. (b, right) Our method achieves higher sample efficiency in Reacher environment, requiring 3x fewer samples than SAC.

Fig. 4. In the Cheetah-run environment. Our method demonstrates higher sample-efficiency. However, given 2 106 samples, Dreamer-v3 shows superior performance compared to our method. All compared methods use identical vector observations.

Fig. 5. (left) Our method demonstrates superior performance in the DoorOpen environment compared to Dreamer-v3 with XL (xlarge) model size and default hyperparameters. Notably, the Dreamer-v3 algorithm fails to solve the door opening task. (right) Frame from DoorOpen environment.

Model-based Policy Optimization

265

6 Conclusion This paper presents the method for model-based policy optimization that successfully solves the door opening task. The main differences from previous works are the training schedule and the use of Neural Ordinary Differential Equations for modeling continuoustime process dynamics. We evaluate the proposed method on different robotic control tasks and compare to Dreamer-v3 model-based algorithm and SAC model-free algorithm. Our method outperforms these methods for door-opening and show comparable results on other tasks.

References 1. Kalashnikov, D., et al.: Qt-opt: Scalable deep reinforcement learning for vision-based robotic manipulation. arXiv preprint arXiv:1806.10293 (2018) 2. Kalashnikov, D., et al.: Scalable deep reinforcement learning for vision-based robotic manipulation. In: Conference on Robot Learning, pp. 651–673 (2018) 3. Lillicrap, T.P., et al.: Continuous control with deep reinforcement learning. arXiv preprint arXiv:1509.02971 (2015) 4. Chen, R.T., Rubanova, Y., Bettencourt, J., Duvenaud, D.K.: Neural ordinary differential equations. In: Advances in Neural Information Processing Systems, vol. 31 (2018) 5. Ha, D., Schmidhuber, J.: World models. arXiv preprint arXiv:1803.10122 (2018) 6. Hafner, D., et al.: Learning latent dynamics for planning from pixels. In: International Conference on Machine Learning, pp. 2555–2565 (2019) 7. Yildiz, C., Heinonen, M., Lahdesmaki, H.: Continuous-time model-based reinforcement learning. In: International Conference on Machine Learning, pp. 12009–12018 (2021) 8. Du, J., Futoma, J., Doshi-Velez, F.: Model-based reinforcement learning for semi-markov decision processes with neural odes. Adv. Neural. Inf. Process. Syst. 33, 19805–19816 (2020) 9. Rubanova, Y., Chen, R.T., Duvenaud, D.K.: Latent ordinary differential equations for irregularly-sampled time series. In: Advances in Neural Information Processing Systems, vol. 32 (2019) 10. Kidger, P., Morrill, J., Foster, J., Lyons, T.: Neural controlled differential equations for irregular time series. Adv. Neural. Inf. Process. Syst. 33, 6696–6707 (2020) 11. Alvarez, V.M.M., Rosca, R., Falcutescu, C.G.: Dynode: Neural ordinary differential equations for dynamics modeling in continuous control. arXiv preprint arXiv:2009.04278 (2020) 12. Ivashko, A., Safonov, G.: Machine learning model for determination of the optimal strategy in an online auction. Inf. Autom. 22(1), 146–167 (2023). https://doi.org/10.15622/ia.22.1.6 13. Hung, N., Loi, T., Huong, N., Hang, T.T., Huong, T.: AAFNDL – an accurate fake in-formation recognition model using deep learning for the Vietnamese language. Inf. Autom. 22(4), 795– 825 (2023). https://doi.org/10.15622/ia.22.4.4 14. Osipov, V., Kuleshov, S., Miloserdov, D., Zaytseva, A., Aksenov, A.: Recurrent neural networks with continuous learning in problems of news streams multifunctional pro-cessing. Inf. Autom. 21(6), 1145–1168 (2022). https://doi.org/10.15622/ia.21.6.3 15. Favorskaya, M., Nishchhal, N.: Verification of marine oil spills using aerial images based on deep learning methods. Inf. Autom. 21(5), 937–962 (2022). https://doi.org/10.15622/ia. 21.5.4 16. Nagabandi, A., Kahn, G., Fearing, R.S., Levine, S.: Neural network dynamics for modelbased deep reinforcement learning with model-free finetuning. In: 2018 IEEE International Conference on Robotics and Automation (ICRA), pp. 7559–7566 (2018)

266

A. Gorodetskiy et al.

17. Janner, M., Fu, J., Zhang, M., Levine, S.: When to trust your model: model-based policy optimization. In: Advances in Neural Information Processing Systems, vol. 32 (2019) 18. Hafner, D., Lillicrap, T., Ba, J., Norouzi, M.: Dream to control: Learning behaviors by latent imagination. arXiv preprint arXiv:1912.01603 (2019) 19. Hafner, D., Lillicrap, T., Norouzi, M., Ba, J.: Mastering atari with discrete world models. arXiv preprint arXiv:2010.02193 (2020) 20. Zhong, Y.D., Dey, B., Chakraborty, A.: Symplectic ode-net: Learning hamiltonian dynamics with control. arXiv preprint arXiv:1909.12077 (2019) 21. Greydanus, S., Dzamba, M., Yosinski, J.: Hamiltonian neural networks. In: Advances in Neural Information Processing Systems, vol. 32 (2019) 22. Yu, T., et al.: Meta-world: a benchmark and evaluation for multi-task and meta reinforcement learning. In: Conference on Robot Learning, pp. 1094–1100 (2020) 23. Gu, S., Holly, E., Lillicrap, T., Levine, S.: Deep reinforcement learning for robotic manipulation with asynchronous off-policy updates. arXiv preprint arXiv:1610.00633 (2016) 24. Zhu, H., Gupta, A., Rajeswaran, A., Levine, S., Kumar, V.: Dexterous manipulation with deep reinforcement learning: efficient, general, and low-cost. In: 2019 International Conference on Robotics and Automation (ICRA), pp. 3651–3657 (2019). https://doi.org/10.1109/ICRA. 2019.8794102 25. Singh, A., Yang, L., Hartikainen, K., Finn, C., Levine, S.: End-to-end robotic reinforcement learning without reward engineering. arXiv preprint arXiv:1904.07854 (2019) 26. Sehgal, A., La, H., Louis, S., Nguyen, H.: Deep reinforcement learning using genetic algorithm for parameter optimization. In: 2019 Third IEEE International Conference on Robotic Computing (IRC), pp. 596–601 (2019). https://doi.org/10.1109/IRC.2019.00121 27. Huang, Y., Xie, K., Bharadhwaj, H., Shkurti, F.: Continual model-based reinforcement learning with hypernetworks. In: 2021 IEEE International Conference on Robotics and Automation (ICRA), pp. 799–805 (2021). https://doi.org/10.1109/ICRA48506.2021.9560793 28. Haarnoja, T., Zhou, A., Abbeel, P., Levine, S.: Soft actor-critic: offpolicy maximum entropy deep reinforcement learning with a stochastic actor. In: International Conference on Machine Learning, pp. 1861–1870 (2018) 29. Brockman, G., et al.: Openai gym. arXiv preprint arXiv:1606.01540 (2016) 30. Tunyasuvunakool, S., et al.: dm control: Software and tasks for continuous control. Softw. Impacts 6, 100022 (2020) 31. Hafner, D., Pasukonis, J., Ba, J., Lillicrap, T.: Mastering Diverse Domains through World Models. arXiv preprint arXiv:2301.04104 (2023)

Monitoring the State of Robotic Systems Based on Time Series Analysis Viktor Semenov(B) St. Petersburg Federal Research Center of the Russian Academy of Sciences, 39, 14th Line, St. Petersburg 199178, Russia [email protected]

Abstract. In view of the close integration of robotic systems into industrial and technological systems, critical infrastructure objects, as well as a significant number of possible entry points, the task of monitoring operational safety for robotic systems is more complex than ensuring information security in classical information systems. The paper presents a method for monitoring the state of robotic systems based on time series analysis. The developed method differs from the existing ones by the combined approach of using an ensemble of parallel classifiers and Fishburn weight coefficients in the security event management system. The time series is composed of a set of informative features, characterizing the functioning of a robotic system. Values for previous discrete time points are ranked using significance weights. The method was approved on a data set of a real industrial system. Due to parallel computing, it was possible to significantly increase the speed of determining the state of robotic systems. The identification precision due to the combined approach increased by 1.45% compared to the best results presented in scientific papers, the recall increased by 4.45% and amounted to 99.85% for both indicators. The results of the study can be applied in monitoring the safety of robotic systems. Keywords: State Analysis · Robotic Systems · Functional Safety · Identification of Anomalies · Decision Trees

1 Introduction The merging of information technologies and industrial processes, observed in recent years in the development of technological infrastructure, has led to the transformation of the principles of building production facilities and the widespread use of robotic systems (RS). At the present stage of RS development, there is an increase in the degree of intelligence of control systems, their autonomy and adaptability, along with this, the volume of processed information transmitted from various sensors is rapidly increasing [1]. These systems are complex and distributed, which also leads to a number of problems associated with their performance and safety [2]. Thus, it is necessary to ensure constant monitoring of the state of the RS, including security assessment, while it is extremely important to take into account the time parameters of potential safety incidents. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 267–278, 2023. https://doi.org/10.1007/978-3-031-43111-1_24

268

V. Semenov

Among the main vulnerabilities of the RS, one can distinguish: the possibility of listening to channels, sending “external” packets, physical access of an attacker to the RS, insufficient standardization of intelligent routing algorithms that take into account the state of the network, etc. Figure 1 shows typical destructive effects on RS elements at different levels: physical, network and application levels.

Fig. 1. Security threats at different levels of robotic systems.

Management of security incidents, including their identification, fixation, prediction, is an ongoing complex process of monitoring and analyzing the results of security events and other data. This kind of monitoring is a complex task of dealing with threats and violations of information security, as well as technological failures and failures, complicated by the heterogeneity of industrial network devices and protocols, the amount of data and the speed of their receipt. An additional problem is the need to adapt monitoring tools in dynamically changing conditions. Proceeding from this, a promising task is to develop methods for monitoring the state of RS in order to ensure their comprehensive information security and functional safety, as well as sustainable operation in the presence of information threats and attacks. Existing methods and technologies are more focused on classical computer or information systems [3–5], which limits the possibility of their application in RS. The solutions currently used [6, 7] do not have sufficient functionality to ensure effective realtime monitoring, which causes a number of problems in ensuring information security and functional safety related to the analysis of the state of RS. In this regard, there is an objective need to develop and adapt the methods of mathematical support for specialized information systems integrated into the RS in order to counteract external and internal destructive influences.

Monitoring the State of Robotic Systems

269

This work is a logical continuation of works [8, 9].

2 Problem Statement Consider the problem of classifying time series that characterize the state of information security and functional safety of RS. Let there be a time series X = {{x 1 (t 1 ), x 2 (t 1 ), …, x S (t 1 )}, {x 1 (t 2 ), x 2 (t 2 ), …, x S (t 2 )}, …, {x 1 (t m ), x 2 (t m ), …, x S (t m )}}, each tuple of which corresponds to a set of characteristics of information or physical processes of the RS at a discrete point in time; {c1 , …, cl } is the set of RS states, l is the number of identified RS states; C = {C 0 , C 1 } is the set of considered states. Each RS element can be in a dangerous (C 1 ) or safe (allowed) (C 0 ) state. C 0 = {c1 , c2 , …, ck }, C 1 = {ck+1 , ck+2 , …, cl }, k is the number of RS safe states. It is required to determine the state of the RS (label of class c), to which the input element of the time series belongs. To train the model, it is necessary to obtain a characteristic of the ongoing information and physical processes of each considered state of the RS. The training sample can be formed, for example, using software for automated analysis of network traffic. The RS state class label c at a discrete time t is described using the method proposed in the paper using a number of the most informative features selected according to [8]:        c = μ a1 xt,1 , xt,2 , . . . , xt,s , a2 xt,1 , xt,2 , . . . , xt,s , . . . , an xt,1 , xt,2 , . . . , xt,s , (1) c ∈ C, xt,i ∈ Df , s , where V is a set of computational nodes, U is a set of ribs. V = {vj } = {}, where j is the node identifier, pj is the node performance. The user operation is described as an acyclic graph G’ = < T,C >, which vertexes are assigned to tasks, and ribs are assigned to data transmission between them. where T is the set of subtasks, C is the set of information connections. T = {t i } = {}, where i is the subtask identifier, wi is the computational complexity of the subtask, d i is the data volume transferred to the network. It must be mentioned that the tasks of data transmission are described in this way as well. The problem solution is the following tasks assignment:    1 ... ...   (8) A =  . . . . . . . . . . ... ... 1  Assume the case when every participant demands the best individual gain. So, we have m objective functions, which represent “individual good” criteria. The objective functions can be represented in the following way: P1 (τ ) = e−λj τ 2 ... Pm (τ ) =

e−λj τ 2

kD1 10

;

kD1 10

;

(9)

where m is the number of nodes which participate in computation process. To transfer m-criteria function to the one-criterion function, we use the multiplicative convolution due to that fact that every node reliability function cannot be compensated with the reliability of the other nodes. The usage of the multiplicative convolution approach allows to assign some preference weights to the chosen nodes: m P0 (τ ) = Pi (τ )ξi , (10) i=1

where ξi determines the preference of the particular node, which, for example, we’d like to offload.

4 Some Assumptions and General Technique for Resource-Saving Multiobjective Task Distribution The main particularity of the problem formulated is that it relates to the structuralparametric optimization ones. Indeed, as we consider data routes and workload of the nodes, first we have to form the nodes community which processes data, and after that to analyze all possible routes, which can be used, and find the best one.

Resource-Saving Multiobjective Task Distribution

283

So, the general problem is decomposed into two consequently solving problems: the problem of data processing nodes set choice and the one of routes forming with the objective function estimation (10). It is seen that the both of problems are np-hard and hardly can be solved with the usage of classical optimization methods, i.e., brunch-andbound technique in an acceptable time. At the same time there are a lot of so-called metaheuristic techniques, such as simulated annealing, genetic algorithms, particle swarm optimization and others. The basic content of iteration, for example, of simulated annealing method contains the following steps [12]: • to form a solution in a random manner; • to compare the solution with previously formed “current solution”; • if new solution is better, then it becomes the new current solution with the probability, which is determined by the particular probability law; • to decrease the temperature variable. Another particularity of the optimization problem considered is that we have to estimate the workload of the nodes, which can process more than one task or transmit data. The separation between data processing and transmission is the first assumption of the modeling technique proposed, because the mixture of data processing and transmission clutters the estimation model. Then, the workload is calculated based on available time. According to our model task processing/transmission time depends on the selected node computational power, as well as on the remainder of the time for task performing and data transmission according to the time constraint for all the tasks to process. Therefore, the workload depends on the node computational power and time constraints. To estimate the workload of the nodes, we have to esteem the time first, which is available to solve the assigned task. The way of time periods estimations is the second assumption of this study. The technique of estimations is as follows: • Consider the task graph and find the critical path; • For each task, which belongs to the critical path, the available time estimation is got as T 0 /N, where N is the number of critical path tasks. Every node on the critical path is marked with its own time estimation. • For each task, which is not on critical path, the available time is esteemed as (T 0 T marked )/N 1 , where T marked is the sum of marked tasks time periods, N 1 is the number of unmarked nodes. • Then the network graph is under consideration. Every data transmission path is analyzed, and when Task i sends data to Task j , and on the network graph there are nodes between the data source and data sink, time estimation for tasks T i and T j is corrected with paying attention to data transmission as follows: every unmarked network node pair, which participates in the process of tasks performing gets the time to process its tasks depending on the data transmission route nodes number, i.e.: T marked /N route . • So all the network nodes are weighed with the time periods, which are available for computational tasks performing. One can see that these estimations for time periods are quite rough. However, such time distribution does not conflict with the common trend: with the increase of data route

284

A. Klimenko and A. Barinov

length more and more time from overall time constraint is used for data transfer, and less time is available for tasks performing. This increases the workload of all nodes on the data transmission route.

5 Experimental Comparison of the Greedy and Multiobjective Solutions According to (5), each node can choose its most beneficial strategy: to process the task or to transmit the data somewhere else. As it was mentioned, data routes are not considered in this greedy approach as well as the reliability features of other nodes do not. Variable x as a time share to transmit the data is substituted with the following, as we can esteem it with the information about the data channels throughput: tprocess = t0 −

k−1 i=0

ti ,

(11)

where t i is the data transfer times in the route, t 0 – overall time constraint. Consider the following task and network graphs (Figs. 1, 2, 3 and 4) (Table 1).

Fig. 1. Experiment 1.

Fig. 2. Experiment 2.

The experimental comparison of the greedy (individual) tasks distribution and the one based on the multiobjective structural-parametric problem solution shows that in the most cases solutions don’t contradict, but solution based of multiobjective optimization problem allows to get a precise result for the group of nodes while the greedy approach

Resource-Saving Multiobjective Task Distribution 1

2

100

100

3

4

100

1

5000

10

100

5

285

Network graph

100

2

Task graph

250

Fig. 3. Experiment 3.

1

2

100

200

3

4

300

1 10

5000

400

2

5

Network graph

500

Task graph

250

Fig. 4. Experiment 4.

Table 1. The results of greedy and multiobjective techniques comparison. Greedy approach

Multiobjective approach

Optimal tasks distribution 1

All nodes prefer to process the task, maximum difference between reliability function value on the node 3

Tasks 0,1 are assigned to the node 1

Optimal tasks distribution 2

All nodes prefer to process the task, maximum difference between reliability function value on the node 1

Tasks 0,1 are assigned to the node 1

Optimal tasks distribution 3

All nodes prefer to process the task, maximum difference between reliability function value on the node 1

Tasks 0,1 are assigned to the node 1

Optimal tasks distribution 4

All nodes prefer to process the task, maximum difference between reliability function value on the node 1

Tasks 0,1 are assigned to the node 1

considers only one node state. The result of experiment 1 shows that with greedy approach there can be situations when the individual preferences of the node contradict

286

A. Klimenko and A. Barinov

with preferences of others and lead to the situation, when one node wins, but others are in situation, which is much worse than they have had before.

6 Experimental Multiobjective Tasks Distribution Consider the following task and network graph (Fig. 5) We presuppose that node 0 is the starting one and plays the role of fog robotic server. Task 0 is always assigned to the node 0 (Fig. 6).

Fig. 5. Task graph.

Fig. 6. Network graph.

The first solution is random. After the simulated annealing has been used, with the forming of sets of nodes, the following improved tasks distribution has been provided (Fig. 7). The results of the experiment illustrate the tasks distribution improvement in the aspect of nodes workload (Fig. 8). One can see that with the optimization conducted some nodes are out of data routes and have no additional workload, so the data processing time increases significantly because of extra time emergency. Also, one can see that multiobjective tasks distribution and the usage of multiplicative convolution are beneficial for all nodes as individuals, which participate the computational process (workload decrease up to 39% comparing with the random distribution).

Resource-Saving Multiobjective Task Distribution

287

Fig. 7. Tasks distribution.

Fig. 8. The workload values with random and optimized tasks distribution.

7 Conclusion Although there is a wide range of papers devoted to the resource allocation and tasks distribution in robot groups, very seldom publications are devoted to the reliability issue of the computational nodes as the integral parts of robots, although those nodes with low computational complexity can process the large amounts of sensor data. Also, in the robot groups sensor data transmission is used as well, although there can be various cases when it is more profitable to process the data by means of robot group. Such lack of common strategy in computational tasks distribution with paying attention to the data transmission tasks, lead to the robot computational resources spending and to the decrease of their potential lifetime because of reliability function uncontrolled deterioration. In this paper we state the problem of resource-saving task distribution. We consider the problem as the multiobjective optimization one, when every node has its own objective function and tries to improve its own workload state. Then this multiobjective function is transformed to the one-criterion optimization problem by the multiplicative convolution. Formulated problem is a structural-parametric one, and can be decomposed into a range of parametric problems, which are solved with the usage of simulated annealing.

288

A. Klimenko and A. Barinov

The main results of this paper are as follows: • multiobjective optimization problem statement for resource-saving tasks distribution within the robot group; • some techniques of problem solution, including rough time estimations for the task and network graphs; • experimental comparison of selected cases of the greedy strategy results and of the multiobjective optimization problem ones. The main conclusion is that “greedy” results do not conflict with the optimization problem solution in the most cases, however, sometimes greedy approach leads to the significant deterioration of the solution for the nodes, which participate in data transmission. • experimental research of tasks distribution among robot nodes according to the techniques proposed shows the positive effect on the individual reliability function values because of the decrease of the data transmission routes and, as a consequence, potential tasks processing time increase. The latter allows to distribute the task computational complexity though time and so to decrease the robot node workload.

References 1. Gul, O.M.: Energy harvesting and task-aware multi-robot task allocation in robotic wireless sensor networks. Sensors 23(6), 3284 (2023) 2. Zhenwei, Z., Zhao, X., Tao, B., Ding, H.: Distributed gossip-triggered control for robot swarms with limited communication range. IEEE Trans. Industr. Electron. 70(12), 12511– 12521 (2023) 3. Wang, S., Wang, Y., Li, D., Zhao, Q.: Distributed relative localization algorithms for multirobot networks: a survey. Sensors 23(5), 2399 (2023) 4. Xu, X., Chai, Z., Xiong, Z., Wu, J.: A scalable resource management architecture for industrial fog robots. In: Liu, X.J., Nie, Z., Yu, J., Xie, F., Song, R. (eds.) ICIRA 2021. LNCS (LNAI), vol. 13013, pp. 67–77. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-89095-7_7 5. Alirezazadeh, S., Correia, A., Alexandre, L.A.: Optimal algorithm allocation for robotic network cloud systems. Robot. Auton. Syst. 154, 104144 (2022) 6. Matrouk, K., Matrouk, A.: Mobility aware-task scheduling and virtual fog for offloading in IoT-fog-cloud environment. Wireless Pers. Commun. 130, 801–836 (2023) 7. Maciel, P., et al.: A survey on reliability and availability modeling of edge, fog, and cloud computing. Reliable Intell. Environ. 8, 227–245 (2022) 8. Baranwal, G., Vidyarthi, D.: TRAPPY: a truthfulness and reliability aware application placement policy in fog computing. J. Supercomput. 78, 7861–7887 (2022) 9. Montoya-Muñoz, A., Silva, R., Caicedo, M., Fonseca, N.: Reliability provisioning for fog nodes in smart farming IoT-fog-cloud continuum. Comput. Electron. Agric. 200, 107252 (2022) 10. Klimenko, A.: Model and method of resource-saving tasks distribution for the fog robotics. In: Ronzhin, A., Meshcheryakov, R., Xiantong, Z. (eds) Interactive Collaborative Robotics. ICR 2022. Lecture Notes in Computer Science, vol. 13719, pp. 210–222. Springer, Cham. (2022). https://doi.org/10.1007/978-3-031-23609-9_19 11. Melnik, E., Klimenko, A.: A condition of reliability improvement of the system based on the fog-computing concept. In: Journal of Physics: Conference Series, vol. 1661, p. 012007 (2020). https://doi.org/10.1088/1742-6596/1661/1/012007 12. Manickam, R., Venkateswaran, C., Ramu, K., Prasanth, V., Mathivanan, G.: Application of Simulated Annealing in Various Field (2022). https://doi.org/10.46632/mc/1/1/1

Reliability of Robot’s Controller Software Eugene Larkin1

, Tatiana Akimenko1(B) , Alexey Bogomolov2 and Vadim Sharov1

,

1 Tula State University, 92, Lenina pr., Tula 300012, Russia

[email protected] 2 St. Petersburg Federal Research Center of the Russian Academy of Sciences, 39, 14th Line

V.O., St. Petersburg 199178, Russia

Abstract. The reliability of robots’ digital control systems, based on Von Neumann type computer platform is investigated. An approach is proposed for assessing the reliability of a robot control program, based on the analysis of software as a model of a computational process unfolding in real physical time. It is shown that in systems of suchlike class the importance of software in ensuring the reliability of robot operation as a whole increases significantly. It is determined that software failure potential is being laid to the control program at its designing stage due to neglect of such digital controllers properties, as time delays when data processing. The model of control program failures emerging, caused by time factor, is worked out. For delays between transactions estimation the semi-Markov model of poling procedure is used, that permit to estimate probabilities of exceeding data skew, pure lag, and sampling period the threshold, beyond which the control system failure takes place. Using a stochastic matrix, describing poling procedure, probability of failure, caused by transactions order disturbances is estimated. Probabilities and sampling period obtained are used for simulation of failure flow generator, describing reliability of control program. Keywords: Control Program · Poling · Sampling Period · Transaction · Semi-Markov Process · Stochastic Matrix · Failure Flow

1 Introduction The main feature of robots, as object under control, is their structural complexity, which is understood as a great number of controllable units, as well as links between them [1–3]. Structural complexity of the object under control implies complexity of software, which may be considered as an integral part of the digital controller, ensuring characteristics requirements of the control system as a whole [4–7]. This, in turn, leads to tightening demands for reliability of software embedded on the controller. Leaving aside hardware aspects of controller reliability, soft failures may be considered as a consequence of program developer errors, arising at all stages of program creation, from specification of demands to data processing in the control system, till the embedding of the completed control program into the controller and acceptance test of the system [8–12]. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 289–299, 2023. https://doi.org/10.1007/978-3-031-43111-1_26

290

E. Larkin et al.

In any digital control system algorithm and the program are models of processing data, receiving from sensors subsystem, to compute data, transmitted to actuators. So, errors in the software mean that either an inadequate, or an imperfect model of the computational process was created, which may or may not lead to the failure of the control system as a whole. This is why, when analyzing program reliability one should take into account not only its algorithm semantics, but also the fact, that the program is interpreted by the controller sequentially, command by command, and the interpretation unfolds in the real physical time. Data processed by the controller is random one, so both the choice of the direction of data processing at program branching points, and the physical time of commands execution, are random values. This fact should be taken into account when developing the data processing model and transforming it to reliability model. For modes of both types the theory of semi-Markov processes is widely used [13, 14]. Therefore, the analysis of the reliability of control programs can be reduced to the transformation of a semiMarkov model, which describes the operation of a digital controller, into a semi-Markov model, which describes its failures due to errors made during programming. Methods for forming a software reliability model for robot control systems are not well developed, which explains the relevance and importance of the research in this domain.

2 Simulation of Robot’s Digital Control System The flowchart of the controllable robot is shown in Fig. 1. [1–7]. System includes robot itself, which is characterized by K-element state vector x(t) = robot state on the Interface is transformed [x1 (t), ..., xk (t), ..., xK (t)]. Measured   to K-element data xc (n) = xc,1 (n), ..., xc,k (n), ..., xc,K (n) , elements of which together with elements of K-element vector of desired robot states f (n) =  by controller soft. The result f1 (n), ..., fk (n), ..., fK (n) are processed   of data processing is the data vector uc (n) = uc,1 (n), ..., uc,k (n), ..., uc,K (n) , which on the interface is transformed to K-element action vector u(t) = [u1 (t), ..., uk (t), ..., uK (t)] of analogue signals, applied to actuators of robot.

Fig. 1. The control system functional diagram.

States x(t) and actions u(t) are linked with non-linear differential equation system:  V x[0(x)] (t), . . . , x[m(x)] (t), . . . , x[M (x)] (t),

Reliability of Robot’s Controller Software

 u[0(u)] (t), . . . , u[m(u)] (t), . . . , u[M (u)] (t), t = 0,

291

(1)

where V = (V1 , ..., Vk , ..., VK )θ is the vector function of vector argument; Vk is the scalar component of vector function; t is the time; x[m(x)] (t) is the m(x)-th order derivative of x(t) with respect to time; u[m(u)] (t) is the m(u)-th derivative of u(t) with respect to time; 0(x) ≤ m(x) ≤ M (x) and 0(u) ≤ m(u) ≤ M (u) are orders of time derivatives; θ is the transpose sign. Vector function (1) describes robot operation dynamics, including sensor and actuator subsystems operation, as well as cross links between control contours, which exist on the physical level. The controller processes data according to the program, in which differentiation and integration operations are substituted by differencing and sunning operations correspondingly. Difference equation takes the form: W{uc [n − 0(u)], . . . , uc [n − m(u)], . . . , uc [n − M (u)], xc [n − 0(x)], ..., xc [n − m(x)], ..., xc [n − M (x)], n, f(n)} = 0,

(2)

where W = (W1 , ..., Wk , ..., WK )θ is the vector function of vector arguments; Wk is the scalar component of the vector function; uc [n − 0(u)] is calculated according the program current value of control data vector; uc [n − m(u)] is the value of control vector, calculated on previous stages, m(u) discrete times ago; xc [n − 0(u)] is the current value of data vector describing current robot state; xc [n − m(u)] is the value of robot state existed m(x) discrete times ago; M (u) and M (x) are maximum values of the difference Eq. (2) order In accordance with the expression (2) the control program successively, elementby-element receives current information xc (n) about the robot state, and successively, element-by-element computes vector uc (n) and transmits it to actuators. Interpretation of control program is carried out cyclically and unfolds in real physical time.

3 Causes of Software Failures Let us define as reference point, designated the beginning of the cycle, the moment of element x1 (t) digital image xc,1 (n) formation, and assume, that all other transactions are generated correspondingly to xc,1 (n) with a delay, which is defined by concrete poling procedure, realized in control algorithm. Delays τ1(x),k(x) , 2(x) ≤ k(x) ≤ K(x), between inputs of first and other elements of vector x(t) born input data skew. Delays τ1(u),k(u) , 2(u) ≤ k(u) ≤ K(u), between outputs of first and other elements of vector u(t) born output data skew. Both skews influence the control process quality characteristics and should be considered when software failure analysis. Time intervals τk(x),l(u) between inputting codes xc,k (n) and outputting codes uc,l (n), where 1 ≤ k, l ≤ K give pure lags in control contours, and also should be taken into account when software failure analysis [15–19]. Besides, non-compliance with Nyquist conditions of sampling theorem when organizing poling procedure leads to software failure [20–22].

292

E. Larkin et al.

Common transactions flow, generated by poling procedure, is shown in Fig. 2, where round point arrows designate input data to controller, squire point arrows designate output data from controller. Start of poling procedure is designated as ttr,n = nτ1(x),1(x) . It is linked with the moment of inputting xc,1 (n) of vector xc (n). The cycle period is designated as τtr = τ1(x),1(x) it lasts from the moment of inputting xc,1 (n) till the xc,1 (n + 1) inputting moment.

Fig. 2. Transactions flow, generated by control program.

For time intervals estimation the model of controller, as a physical device, operated in a physical time, should be built. For simplicity the model may be represented by transactions operators only. It is assumed that all computations are executed by the proper transaction operator. An adequate model of controller operation when control program interpretation is the semi-Markov process, represented by Z × Z semi-Markov matrix [13, 14]       (3) h(t) = hξ,ζ (t) = gξ,ζ (t) ⊗ pξ,ζ , where pξ,ζ are probabilities of direct switching from the state ξ to the state ζ; gξ,ζ (t) is the time density of sojourn the process in the state ξ, when a priory is known, that next state will be the ζ one; ⊗ is the sign of direct multiplication of matrices. ξ, ζ ∈ {1(x), ..., k(x), ..., K(x) , 1(u), ..., k(u), ..., K(u)}; Z = 2K, ; is the common quantity of semi-Markov process states, equal to common quantity of transactions, generated by control program. Due to the poling procedure is the cyclic one, and during the cycle all transactions should be updated, the structure of semi-Markov process is represented by a complete strongly connected graph with loops, shown in Fig. 3a. Semi-Markov process (3) is the ergodic one, the absorbing  or semi-absorbing states are absent in the process. On  probabilities of matrix pξ,ζ and time distribution densities the following restrictions are lay. min arg [gξ,ζ (t)]max ξ,ζ

0 < Tξ,ζ

Z ζ =1



, 1 ≤ ξ, ζ ≤ Z;

pξ,ζ = 1; 1 ≤ ξ ≤ Z;

max ∫=1 Tξ,ζ

min Tξ,ζ

gξ,ζ (t)dt,

Reliability of Robot’s Controller Software

293

Fig. 3. Structures of semi-Markov processes: a is the initial one; b is the process for estimation of time interval between transactions; c is the process for estimation of polling procedure period. min and T max are uppermost and lowermost borders of density g (t) domain. where Tξ,ζ ξ,ζ ξ,ζ For estimation of time interval value from transaction ξ till transaction ζ the ergodic semi-Markov process, shown in Fig. 1a should be transformed into the process,      , (4) h (t) = gi,j (t) ⊗ pi,j

shown in Fig. 3b, in which state ξ and ζ are starting and absorbing ones correspondingly. The transforming is reduced to recalculation of probabilities according to the formula: 0, when i = ζ, or j = ξ ;  = pi,j pi,j 1−pi,ξ in all other cases. where 1 ≤ i, j ≤ Z. Time delay is estimated as follows [23] ∞   i  c  · Iζ gξ,ζ L h (t) (t) = I rξ · L−1 i=1

(5)

where L[...] and L−1 [...] are direct and inverse Laplace transforms, correspondingly; [24–26]; I rξ is the Z -element row vector, ξ-th element of which is equal to one, and other elements are equal to zero; I cζ is the Z -element column vector ζ-th element of which is equal to one, and other elements are equal to zero. The probability of system failure due to exceeding of delay τξ,ζ its critical value τ˜ξ,ζ is as follows [27]:  qτ,ξ,ζ = ∫∞ τ˜ξ.ζ gξ,ζ (t)dt,

(6)

where ξ, ζ ∈ {1(x), ..., k(x), ..., K(x) , 1(u), ..., k(u), ..., K(u)}. For estimation of transaction cycle period value semi-Markov process (3) should be transformed into the process   (7) h (t) = hξ,ζ (t) ,

294

E. Larkin et al.

where h (t) is the (Z + 1) × (Z + 1) semi-Markov matrix. ⎧ ⎨ hξ,1 (t), when ζ = Z + 1 hξ,ζ (t) = 0, when ξ = Z + 1, or ζ = 0; ⎩ hξ,ζ (t) in all other cases. Structure of the semi-Markov process (7) is shown in Fig. 3c. With use (7) the period of transactions cycle may be estimated as follows ∞   i  c · I Z+1 , gtr (t) = I r1 · L−1 L h (t) (8) i=1

where I rξ is the (Z + 1)-element row vector, first element of which is equal to one, and all other elements are equal to zeros; I cZ+1 is the (Z + 1)-element column vector, (Z + 1)-th element of which is equal to one, and all other elements are equal to zeros. The probability of system failure due to exceeding of cycle period τtr pre-determined critical value τ˜tr may be estimated as qtr = ∫∞ τ˜tr gtr,1,1 (t)dt.

(9)

Programming links missing, leading to transactions order violation is fairly common error when coding control algorithm of a complex multi-contour object that is a robot. When transactions order violation part of vector x(t) elements get processed with algorithm (2) not in the current, but in the next cycle, that may lead or may not lead to system failure. To estimate the probability of failures caused by errors of this type, the task of routing walks through the states of semi-Markov process would be solved. For solving task, the stochastic matrices concatenation operation should be defined. Stochastic matrices 1 p and 2 p concatenation is the matrix:    1 2 2 1 (10) p& p = pξ,ψ & pψ,ζ , ψ=1

  where & is concatenation operation notation; 1 p = 1 pξ,ψ is the ×  stochastic   matrix; 2 p = 2 pψ,ζ is the  × Z stochastic matrix. As it follows from (10), stochastic matrices concatenation operation is performed according to rules of matrices multiplication [28, 29], but instead of multiplication there is concatenation operation, and instead of summation there is operation of strings joint. For   &i  &(i−1)    = pξ,ζ & pξ,ζ the squire stochastic matrix pξ,ζ concatenation degree pξ,ζ  &Z may be defined. As a result of concatenation exponentiation pξ,ζ , strings of Z probabilities designations are formed, located on routes with Z length, leading from the state, with number of the stochastic matrix row to the state with number of the stochastic matrix column Indices of probabilities nomination strings define routes of wandering through semi-Markov process (3) states, and hence determine the transactions sequence in semi-Markov process realizations. From the point of view of software failure, under interest are strings, which begin and end in the first state of semi-Markov process (3), due to other strings do not form required cycles.

Reliability of Robot’s Controller Software

295

&Z  In common case component with index 1,1 of matrix pξ,ζ is as follows: P1,1 =

j j=1

P1,1,j ,

(11)

where P1,1,j is the j-th string of probabilities nominations, which starts and finishes at the state 1; J = (Z − 1)(Z−1) is common number of strings. From J strings P1,1,j of (11), only strings with non-recurring indices describe proper transactions, because during cycle all robot’s sensors and actuators must be quested. Common number J˜ = (Z − 1)! of such string is defined as quantity of Hamiltonian cycles starting and finishing at the state 1. From J˜ Hamiltonian cycles only Jˆ ≤ J˜ satisfy to restrictions on the wandering routes, imposed by semantic of algorithm (2). For estimation of probability of Hamiltonian cycle P 1,1,ˆj = p ˆ &p ˆ ˆ &...&p ˆ ˆ &...&p ˆ forming probabilities from the 1,ξ j

ξ j ,ψ j

i j ,m j

ζ j ,1

string should be transformed to cortege   &p     &...&p     &...&p   1,ξ ˆj ξ ˆj ,ψ ˆj i ˆj ,m ˆj ζ ˆj ,1

p

 → p

 , 1 ˆj

 , 2 ˆj

p

..., p ˆ , ..., p

 Z ˆj

l j

 ,

→ (12)

   where 1 ˆj ≤ l ˆj ≤ Z ˆj is the index, defining number of elements in the cortege. Probability of failure because or transactions ordering violation is equal to qord = 1 −

 ˆj  p l ˆj l ˆj =1 ˆj

Z

Jˆ ˆj=1

(13)

4 Model of Failure Flux Forming In the roughest version, failures emerging due to investigated above causes can be considered as independent ones. Due to this assumption the probability of software failure is equal to sum of probabilities, i.e.: q=

Z ζ =1

Z ξ =1

qτ,ξ,ζ + qtr + qord .

(14)

With use (14) and (8) the semi-Markov model of the software failures generator may be formed:   (1 − q) · gtr (t) q · gtr (t) . (15) hϕ (t) = 0 0 The structure of the model (15) is shown in Fig. 4. The loop in the structure describes the absence of failures during the cycle, arc from the state α1 to the state α2 means the failure of control program.

296

E. Larkin et al.

Fig. 4. Model of software failures generator.

Time density from switching on the controller till software failure is as follows:   ∞   i  0 gϕ (t) = (1, 0) · L−1 · L hϕ (t) . (16) i=1 1 As it follows from (16) time between failures represents by stochastic sum of times, between one, two, etc. cycles. In accordance with [30–33] failure flux may be considered as Poisson one, in which time is distributed according to exponential law:   t 1 , (17) exp − gϕ (t) ∼ = Tϕ Tϕ ∞

where Tϕ = ∫ t · gϕ (t)dt is the time to failure. 0

Parameter Tϕ can be change to a big side, if robot operating mode allows correction the embedded control program.

5 Example Let’s consider software failure due to missing, when control program design, features of digital control system operation, discussed above. The robot’s controllable equipment, including sensors and actuators, is described with the following differential equation system, performed in the form (18): (3) (2) (1) 2, 5 · 10−3 · x1 (t) + 0, 1 · x1 (t) + x1 (t) = u1 (t) + u2 (t); (18) (3) (2) (1) 6, 25 · 10−4 · x2 (t) + 0, 05 · x2 (t) + x2 (t) = −2, 5u1 (t) + 2, 5u2 (t), where x1 (t), x2 (t) are state parameters of robot equipment; u1 (t), u2 (t) are control actions. Control actions, applied to equipment actuators, being computed by controller, is as follows:  u1 (t) = η(t) − x1 (t); (19) u2 (t) = η(t) − x2 (t), where η(t) is the Heaviside function. Transient processes in the control system are shown in Fig. 5. Curves in Fig. 5 represent dynamics of the system without lags, data skews, and other errors, introduced by digital controller. Overshooting in the system does not exceed ten percents, and the control overtime is equal to 3 s. Figure 5b shows disturbances emerging

Reliability of Robot’s Controller Software

297

at the sampling stage, when Nyquist conditions are violated by poling procedure. In this case in the system occur fluctuations around equilibrium point, which amplitude the greater, the less inertial is the object under control. In the system with delays in feedback circuit overshooting increases up to twenty percent, and control overtime increases to 4 s, which in some cases, may be considered as the system failure. Transactions order violation means that data describing object under control state arrive for processing after the sampling period, that born both data skew, and feedback signal delay (Fig. 5d). In turn, both these factors lead to software failure.

Fig. 5. Transient processes in the digital control system.

6 Conclusion In such a way, the approach to evaluation of robot’s control program reliability, based on analysis of the software as a model of computing process, unfolding in real physical time, is proposed. It is shown how the model of computing process may be transformed into the software failure generator model, which parameters are defined by the soft, embedded onto robot’s digital controller. To avoid considered type system failures during robot with installed control program exploiting, it may be recommended at the stage of software design the setting strong limits on the program runtime parameters, especially on the sampling period and delays in feedback contours, for a possibly wide range of robot states. Further investigations in the domain may be directed on working out methods of robot’s control software failure-free design.

298

E. Larkin et al.

Acknowledgments. The research was carried out within the grant № 22–26-00808 of Russian Scientific Foundation.

References 1. Siciliano, B., Khatib, O.: Springer Handbook of Robotics. Springer, Berlin (2016) 2. Mehmszow, U.: Mobile Robotics: A Practical Introduction. Springer, London (2003) 3. Kamaldar, M., Mahjoob, M.J., Yazdi, M.H., Vahid-Alizadeh, H., Ahmadizadeh, S: A control synthesis for reducing lateral oscillations of a spherical robot. In: IEEE International Conference on Mechatronics, pp. 546–551 (2011) 4. Landau, I.D., Zito, G.: Digital Control Systems. Design. Identification and Implementation. Springer, London (2006) 5. Aström, J., Wittenmark, B.: Computer Controlled Systems: Theory and Design. Tsinghua University Press, Beijing (2002) 6. Larkin, E.V., Nguyen, V.S., Privalov, A.N.: Simulation of digital control systems by nonlinear objects. In: Dang, N.H.T., Zhang, YD., Tavares, J.M.R.S., Chen, BH. (eds.) Artificial Intelligence in Data and Big Data Processing. ICABDE 2021. Lecture Notes on Data Engineering and Communications Technologies, vol. 124, pp. 711–721. Springer, Cham (2022). https:// doi.org/10.1007/978-3-030-97610-1_56 7. Fadali, M.S., Visioli, A.: Digital Control Engineering: Analysis and Design. Academic Press, Cambridge (2012) 8. Dubrova, E.: Fault-Tolerant Design. Springer, New York (2013) 9. Sánchez-Silva, M., Klutke, G.-A.: Reliability and Life-Cycle Analysis of Deteriorating Systems. Springer, Switzerland (2016) 10. O’Conner, P., Kleyner, A.: Practical Reliability Engineering. Willey, Hoboken (2012) 11. Koren, I., Krishna, M.: Fault Tolerant Systems. Morgan Kaufmann Publishers, San Francisco, CA (2007) 12. Zhang, Y., Jiang, J.: Bibliographical review on reconfigurable fault-tolerant control systems. Annu. Rev. Control. 32(2), 229–252 (2008) 13. Howard R.A.: Dynamic Probabilistic Systems. VOL. 1 Markov Models, Vol. 2 Semi-Markov and Decision Processes. Courier Corporation (2012) 14. Janssen, J., Manca, R.: Applied Semi-Markov processes. Springer, Cham (2006) 15. Pospíšil, M.: Representation of solutions of delayed difference equations with linear parts given by pairwise permutable matrices via Z-transform. Appl. Math. Comput. 294, 180–194 (2017) 16. Sanz, R., García, P., Fridman, E., Albertos, P.: Robust predictive extended state observer for a class of nonlinear systems with time-varying input delay. Int. J. Control 93(2), 217–225 (2020) 17. Wu, M., He, Y., She, J.H., Liu, G.P.: Delay-dependent criteria for robust stability of timevarying delay systems. Automatica 40(8), 1435–1439 (2004) 18. Zhang, X.M., Wu, M., He, Y.: Delay dependent robust control for linear systems with multiple time-varying delays and uncertainties. Control Decision 19, 496–500 (2004) 19. Wu, R., Fan, D., Iu, H.H.C., Fernando, T.: Adaptive fuzzy dynamic surface control for uncertain discrete-time non-linear pure-feedback mimo systems with network-induced time-delay based on state observer. Int. J. Control 92(7), 1707–1719 (2019) 20. Hayes, M.H.: Statistical digital signal processing and modeling. Willey, Hoboken (2009) 21. Yeh, Y.C., Chu, Y., Chiou, C.W.: Improving the sampling resolution of periodic signals by using controlled sampling interval method. Comput. Electr. Eng. 40(4), 1064–1071 (2014)

Reliability of Robot’s Controller Software

299

22. Meyer-Baese, U.: Digital Signal Processing. Springer, Heidelbrg (2004) 23. Larkin, E.V., Bogomolov, A.V., Privalov, A.N.: A method for estimating the time intervals between transactions in speech-compression algorithms. Automatic Documentation Math. Linguist. 51, 214–219 (2017) 24. Pavlov, A.V.: About the equality of the transform of Laplace to the transform of Fourier. Issues Anal. 23, 21–30 (2016) 25. Li, J., Farquharson, C.G., Hu, X.: Three effective inverse Laplace transform algorithms for computing time-domain electromagnetic responses. Geophysics 81(2), E113–E128 (2016) 26. Schiff, J.L.: The Laplace transform: Theory and Applications. Springer, New York (1991) 27. Kobayashi, H., Mark, B.L., Turin, W.: Probability, Random Processes, and Statistical Analysis: Applications to Communications, Signal Processing, Queueing Theory and Mathematical Finance. Cambridge University Press, Cambridge (2011) 28. Petersen, P.: Linear Algebra. Springer, New York (2012) 29. Cherney, D., Denton, T., Thomas, R., Waldron, A.: Linear Algebra. Davis, California (2013) 30. Grigelionis, B.: On the convergence of sums of random step processes to a Poisson process. Theory Probab. Appl. 8(2), 177–182 (1963) 31. Lu, H., Pang, G., Mandjes, M.: A functional central limit theorem for Markov additive arrival processes and its applications to queuing systems. Queuing Syst. 84(3), 381–406 (2016) 32. Larkin, E., Bogomolov, A., Gorbachev, D., Privalov, A.: About approach of the transactions flow to Poisson one in robot control systems. In: Ronzhin, A., Rigoll, G., Meshcheryakov, R. (eds.) ICR 2017. LNCS (LNAI), vol. 10459, pp. 113–122. Springer, Cham (2017). https:// doi.org/10.1007/978-3-319-66471-2_13 33. Larkin, E., Privalov, A., Bogomolov, A., Akimenko, T.: Model of digital control system by complex multi-loop objects. In: AIP Conference Proceedings, vol. 2700, no. 1. AIP Publishing (2023)

Ontological Approach to the Organization of Computing in Distributed Monitoring Systems with Mobile Components Based on a Distributed Ledger Eduard Melnik

and Irina Safronenkova(B)

Federal Research Center, The Southern Scientific Center of the Russian Academy of Sciences, 41, Chekhov st., Rostov-on-Don 344006, Russia [email protected]

Abstract. The implementation of mobile components into monitoring systems has significantly expanded their scope of application. A distributed ledger (DL) allows to synchronize copies of data received from geographically distributed sources. This makes it possible for mobile components to work with the “nearest” copy of the ledger in terms of data access, which reduces the time and energy costs associated with data access. This paper considers issues related to the efficiency of distributed monitoring systems with mobile components based on the concept of fog computing and DL technologies. The problem of organizing the computational process in terms of placement and workload reallocation is highlighted. It is proposed to use the ontological approach in order to reduce the search space of candidate nodes for the placement of computational load. The use of other known methods of placing the computational load is limited by the characteristics of the environment and the system. The core of ontological approach is the ontological model built with the subject domain in mind. Experimental studies have shown that the application of the ontological approach can reduce the search space by 80%. The obtained data are consistent with the results of the experiment conducted for another subject area. Keywords: Monitoring Systems · Mobile Components · Distributed Ledger · Ontological Model · Latency · Workload Relocation

1 Introduction The concept of fog computing implies workload relocation closer to the edge of the network in order to reduce the load on the information infrastructure and to reduce the latency of the system [1–3]. This concept fits well with the basic requirements for monitoring systems, for which response time is one of the most important parameters of their functioning [4]. The implementing of mobile components (UAVs, autonomous robots, devices on board the vehicle, mobile devices) in monitoring systems has greatly expanded their © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 300–310, 2023. https://doi.org/10.1007/978-3-031-43111-1_27

Ontological Approach to the Organization of Computing

301

scope of application. Distributed data storage in such systems based on a distributed ledger allows to increase the convenience of access to data and in many cases to reduce the load on the network. A DL allows synchronizing copies of data obtained from geographically distributed sources. This enables mobile components to work with the “closest” copy of the ledger in terms of data access, which reduces the time and energy costs associated with data access [5–7]. At the same time, when organizing the computational process, it is necessary to take into account the dynamism of the computing environment implemented on the basis of fog and edge devices, taking into account the peculiarities of the functioning of mobile components. One of the important stages of such organization is the stage of solving the problem of placement and relocation of computational load, which is performed repeatedly [8–10]. The number of relocations affects the following important parameters: the time to solve a functional task and power consumption. The latter is especially relevant for distributed systems with mobile components. It is possible to reduce the number of workload reallocations by selecting such an initial placement of computational task on nodes that would meet a certain level of reliability of computational nodes of the system. In this case, we will understand the reliability as the ability of a computational node to perform an assigned amount of calculations at a certain moment of time. Thus, we can say that providing a certain level of reliability of computational nodes that perform the problem solving, allows to reduce the time of solving the functional task and reduce power consumption, which is an important criterion for the effective functioning of the distributed system with mobile components. In other words, improving the quality of the initial placement of the task by computational nodes allows to increase the efficiency of operation of distributed monitoring systems with mobile components based on a DL. This paper proposes to use an ontological approach to select the most “suitable” candidate nodes for a particular problem, both for the initial placement and for possible subsequent workload relocations. This approach reduces the search space and shortens the time to make a decision [11]. The use of other known methods of placing the computational load is limited by the characteristics of the environment and the system. The methods and algorithms proposed in [12, 13] do not take into account the influence of the solution time of the problem of moving the workload on the efficiency of solving the user problem. The strategies and approaches proposed in [14–16] have high computational complexity and are not oriented to solving problems of high dimensionality or complex topology. When implementing a LDG-method [17], the time to obtain a solution may be unacceptably long due to numerous information exchanges. The constraint forming method for solving the problem of computational load reallocation in a distributed system [18] does not take into account the presence of mobile components and DL nodes.

2 Ontological Approach to the Organization of Computations in Distributed Systems The idea of using ontologies to reduce the search space is not new. For example, it was proposed to use domain ontology models to reduce the search space when solving the workload relocation problem in a distributed system [11, 18].

302

E. Melnik and I. Safronenkova

The method within the ontological approach, aimed at reducing the search space and including an ontological model of the subject area, is performed in 6 steps (Fig. 1): • Step 1. Input data, containing information about the transferred subgraph of the problem and about the resources of the candidate nodes. • Step 2. Carry out ontological analysis procedure. • Stage 3. Forming limited space of candidate nodes. • Step 4. Simulation procedure of computational workload allocation on the limited set of nodes. • Step 5. Solving the problem of workload relocation on the limited set. • Step 6. Obtaining the result of the problem of workload relocation.

Fig. 1. Ontological method for reducing the search space of candidate nodes.

The ontological analysis procedure is based on the domain ontological model, which reflects the information about the computing nodes involved in the distribution of workload of the system and in the direct solution of the functional task of the system, as well as information about the information flows inherent in this system. Accordingly, to form a domain ontological model describing the specifics of functioning of distributed systems with mobile components, implemented using fog computing and DL technologies, it is necessary to conduct a comprehensive analysis of this subject area.

3 Development of a Domain Ontological Model 3.1 Methodology for the Development of a Domain Ontological Model The development of a domain ontological model involves performing the following steps, according to [19]: • • • • • • •

identifying the domain and scope of the ontology; consideration of options for reuse of existing ontologies; listing important terms in the ontology; definition of classes and class hierarchy; definition of class properties - slots; definition of slot facets; creation of instances.

Ontological Approach to the Organization of Computing

303

3.2 Analysis of the Subject Area of Operation of a Distributed Monitoring System with Mobile Components Based on a Distributed Ledger To design a domain ontological model it is necessary to carry out a comprehensive analysis of the area under consideration, in this case it is the area of operation of a distributed monitoring system with mobile components based on a DL. The concept of a scheme for collecting, processing, transmitting and storing data in such a distributed system has the form shown in Fig. 2 [20].

Fig. 2. The concept of functioning of a distributed monitoring system with mobile components based on a DL.

The data arriving at the input of the distributed system consists of a set of data collected and, in most cases, preprocessed by mobile components, as well as data that are copies of the DL data. Once the data arrive, their subsequent processing takes place, which is the process of solving a set of complex tasks (T1 , T2 ,…TN ). To solve complex tasks we split them into subtasks and then assign them to computational nodes, and in some cases such nodes can be DL nodes (L1 , L2 ,…,LN ), which is not desirable, and the priority of assigning subtasks to nodes is taken into account in the proposed approach, which will be described below. A group of computing nodes, calculating a particular task, passes the result of the obtained solution (output) data to the “closest” node of the DL, as shown in Fig. 3.

304

E. Melnik and I. Safronenkova

For example, the result of solving problem T1 will be transmitted to nodes L1 and L2 of the DL, while the results of solving subtasks T2 and T3 will be transmitted to nodes L1 and L2 , respectively.

Fig. 3. The computational process schemes.

Let us consider various scenarios for the implementation of the computational process in a monitoring system with mobile components based on distributed ledger. Scenario 1. Let the computational process (solution of some functional problem T2 ) involve only “edge” components of the system. In this case, the result of solving some problem is transmitted to the “closest” node of distributed ledger is node L1 . Scenario 2. Let a group of computational nodes, for which node of distributed ledger L1 is “closest”, solve some computational task is T2 . In the case where the computational node (group) needs in accordance with the mission to collect data, geographically remote from the node of distributed ledger, which was for him the “closest” is L1 until this node has not changed its position in space, then in terms of reducing the time costs of data transfer, the node must transfer data to the node of distributed ledger is L3 , which is for him at the moment is the “closest” (Fig. 4). Then the node of DL L3 synchronizes the data on those nodes of DL that need the data. Thus, to synchronize the data, the nodes of DL located in the “fog” layer are included in the computational process. Based on the scenarios described above, we can conclude that two levels can be distinguished in the process of organizing the computational process: • The “edge” layer level, when the computational process is performed at the level of mobile components and other “edge” devices; • The level of the “fog” layer, when to organize the computational process it is necessary to include DL nodes, which are not “closest” to the group of computing devices involved in solving the problem. The following key concepts and their characteristics were identified in the process of analyzing the subject area:

Ontological Approach to the Organization of Computing

305

Fig. 4. Computational process according to Scenario 2.

• the structure of the task graph; • the method of cutting the task graph: longitudinal (cutting into graph branches) and transverse (cutting into tiers of the graph); • information flows: volume of transmitted data, frequency of transmitted data; • communication channels: latency, bandwidth; • node-candidate: energy resource, bandwidth, distance from the source/receiver of data, mobility of the candidate node, “edge” level latency; • DL candidate node: energy resource, bandwidth, distance from data source/receiver, candidate node mobility, link latency, “fog” level latency. 3.3 Domain Ontological Model of Distributed Monitoring System Functioning with Mobile Components Based on DL Technologies Based on this analysis and the ontology development methodology described above, a domain ontological model for the operation of a distributed monitoring system with mobile components based on DL technologies was developed. The developed model is shown in Fig. 5. On the basis of the developed domain ontological model, which allows to describe the features of functioning of the distributed monitoring system with mobile components based on a DL, in accordance with the basic principles of distributed computing, taking into account the characteristics of the functioning environment, a software model has been obtained. The specified software model allowed us to conduct a series of experiments.

306

E. Melnik and I. Safronenkova

Fig. 5. Domain ontological model for the operation of a distributed monitoring system with mobile components based on DL technologies.

4 Experimental Studies The aim of the experimental study: to investigate the effectiveness of the developed software model depending on the initial number of candidate nodes for different cutting methods (longitudinal and transverse). The effectiveness criterion of the developed software model will be understood as a reduction in the number of candidate nodes after the ontology analysis procedure in comparison with the initial number of candidate nodes. The effectiveness of the proposed method is calculated by the formula (1): =

Finit − Ffin × 100%, Finit

(1)

Ontological Approach to the Organization of Computing

307

where Finit is initial number of candidate nodes to place computational load; Ffin is final number of candidate nodes to place computational load after ontology analysis procedure. By the depth of the fog layer we will mean the maximum distance, measured in hops, from the leader node to the computational node to which some of the computational load can be transferred. Scenario 1 consists of computational experiments on the implementation of the ontological analysis procedure on the set of computational nodes, the depth of which lies in the range from 1 to 5 hops (see Fig. 6, 7). Scenario 2 consists of computational experiments on the implementation of ontology analysis on the set of computational nodes, the depth of which is between 1 and 30 hops (see Fig. 8, 9). The varied parameter is also the initial number of candidate computational nodes, Finit . F_fin_{1;5}_l – selected nodes after ontological analysis at layer depth {1;5} hop for the longitudinal method of graph cutting. F_fin_ {1;30}_l – selected nodes after ontological analysis at layer depth {1;30} hop for the longitudinal method of graph cutting. F_fin_{1;5}_t – selected nodes after ontological analysis at layer depth {1;5} hop for the transversal method of graph cutting. F_fin_{1;30}_t – selected nodes after ontological analysis at layer depth {1;30} hop for the transversal method of graph cutting.

Fig. 6. Diagram showing the number of selected nodes after the ontology analysis procedure, when Finit = 100, the depth lies in the range from 1 to 5 hops, different methods of task graph cutting.

According to formula (1), the effectiveness of the proposed method can be estimated at 80%, which means reducing the search space by 5 times. The overall effect of reducing the decision time for the selection of candidate nodes depends on the time complexity of the optimization algorithm used.

308

E. Melnik and I. Safronenkova

Fig. 7. Diagram showing the number of selected nodes after the ontology analysis procedure, when Finit = 100, the depth lies in the range from 1 to 30 hops, different methods of task graph cutting.

Fig. 8. Diagram showing the number of selected nodes after the ontology analysis procedure, when Finit = 1000, the depth lies in the range from 1 to 5 hops, different methods of task graph cutting.

Fig. 9. Diagram showing the number of selected nodes after the ontology analysis procedure, when Finit = 1000, the depth lies in the range from 1 to 30 hops, different methods of task graph cutting.

Ontological Approach to the Organization of Computing

309

5 Results The obtained results of the experimental studies demonstrate the promising use of the ontological approach to the selection of candidate nodes for placement and redistribution of computational load in distributed monitoring systems with mobile components based on the concept of fog computing and DL, and also agree well with the results obtained in earlier studies [13, 14].

6 Conclusion In this paper, we examined issues related to the organization of the computational process in monitoring systems with mobile components, implemented using the concept of fog computing and distributed ledger technologies, in terms of solving the problem of placement and redistribution of the computational load. To solve this problem, it was proposed to use ontological approach to reduce the search space, which has shown its effectiveness in solving large dimension optimization problems. Domain ontological model of distributed monitoring systems that takes into account the specifics of mobile components functioning, as well as the characteristics of the environment and data access organization on the basis of a distributed ledger was developed. Experimental studies have shown the applicability of the ontological method and the adequacy of the developed ontological model to the considered subject area. Acknowledgments. This study is supported by the GZ SSC RAS N GR project 122020100270–3.

References 1. Moura, J., Hutchison, D.: Fog computing systems: state of the art, research issues and future trends, with a focus on resilience. J. Netw. Comput. Appl. 169, 102784 (2020) 2. Movahedi, Z., Defude, B., Hosseininia, A.M.: An efficient population-based multi-objective task scheduling approach in fog computing systems. J. Cloud Comput. 10(1), 1–31 (2021) 3. Abdali, T.A.N., Hassan, R., Aman, A.H.M., Nguyen, Q.N.: Fog computing advancement: concept, architecture, applications, advantages, and open issues. IEEE Access 9, 75961–75980 (2021) 4. Abreha, H.G., Cano, C.J.B., De La Oliva, A., Cominardi, L., Saloa, A.A.: Self-Adaptive Monitoring in Fog Computing by Leveraging Machine Learning. Technical Report (2020) 5. Distributed Ledger Technology, Blockchains and Identity. A Regulatory Overview. https:// www.gsma.com/identity/wp-content/uploads/2018/09/Distributed-Ledger-Technology-Blo ckchains-and-Identity-20180907ii.pdf. Accessed 30 May 2023 6. Lücking, M., et al.: The merits of a decentralized pollution-monitoring system based on distributed ledger technology. IEEE Access 8, 1–17 (2020) 7. Clark, N., Maglaras, L., Kantzavelou, I., Chouliaras, N., Ferrag, M.A.: Blockchain technology: security and privacy issues. In: Patnaik, S., Wang, T.-S., Shen, T., Panigrahi, S.K. (eds.) Blockchain Technology and Innovations in Business Processes. SIST, vol. 219, pp. 95–107. Springer, Singapore (2021). https://doi.org/10.1007/978-981-33-6470-7_6 8. What is fog computing? What is Fog Computing? - Definition from IoTAgenda, techtarget.com Accessed 30 May 2023

310

E. Melnik and I. Safronenkova

9. Hurbungs, V., Bassoo, V., Fowdur, T.P.: Fog and edge computing: concepts, tools and focus areas. Int. J. Inf. Technol. 13, 511–522 (2021) 10. Li, G., Yan, J., Chen, L., Wu, J., Lin, Q., Zhang, Y.: Energy consumption optimization with a delay threshold in cloud-fog cooperation computing. IEEE Access 7, 159688–159697 (2021) 11. Klimenko, A.B., Safronenkova, I.B.: A Technique of workload distribution based on parallel algorithm structure ontology. Adv. Intell. Syst. Comput. 1046, 37–48 (2019) 12. Yin, L., Luo, J., Luo, H.: Tasks scheduling and resource allocation in fog computing based on containers for smart manufacturing. IEEE Trans. Industr. Inf. 14(10), 4712–4721 (2018) 13. Jamil, B., Shojafar, M., Ahmed, I., Ullah, A., Munir, K., Ijaz, H.: A job scheduling algorithm for delay and performance optimization in fog computing. Concurrency Comput. Pract. Experience 32(7), e5581 (2020) 14. Sun, Y., Lin, F., Xu, H.: Multi-objective Optimization of resource scheduling in fog computing using an improved NSGA-II. Wireless Pers. Commun. 102, 1369–1385 (2018) 15. Wang, J., Li, D.: Task scheduling based on a hybrid heuristic algorithm for smart production line with fog computing. Sensors 19(5), 1023 (2019) 16. Ghenai, A., Kabouche, Y., Dahmani, W.: Multi-user dynamic scheduling-based resource management for Internet of Things applications. In: 2018 International Conference on Internet of Things, Embedded Systems and Communications (IINTEC), Hamammet, Tunisia, pp. 126–131 (2018) 17. Melnik, E.V., Klimenko, A.B., Ivanov D.Y.: The model of the problem of forming communities of information and control system devices in fog computing environments. In: XIII All-Russian Meeting on Management Problems-2019, pp. 2979–2984. Tula State University, Tula (2019) 18. Klimenko, A., Safronenkova, I.: An ontology-based approach to the workload distribution problem solving in fog-computing environment. In: Silhavy, R. (ed.) CSOC 2019. AISC, vol. 985, pp. 62–72. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-19810-7_7 19. Noy, N., McGuinness, D.: Ontology development 101: a guide to creating your first ontol-ogy. Stanford knowledge systems laboratory Technical report KSL-01–05 and Stanford Medical Informatics Technical report SMI-2001–0880 (2001) 20. Kapustyan, S. G., Orda-Zhigulina, D. V., Orda-Zhigulina, M. V., Prakapovich, R. A., Sychev, U. A.: Model of multi-robotic complex at the base of distributed registry for monitoring and diagnostics system. In: Silhavy, R. (ed.) CSOC 2021. LNNS, vol. 228, pp. 659–669. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-77448-6_64

Improved Model of Greedy Tasks Assignment in Distributed Robotic Systems Anna Klimenko(B) Institute of IT and Security Technologies, RSUH, 25-2, Kirovogradskaya Street, Moscow 117534, Russia [email protected]

Abstract. The problem of computations efficiency estimation is a topical one nowadays because of cost and various constraints, including energy consumption, resource spending, data transmission constraints, etc. Taking into account the tight connection between distributed robotic systems and IoT concepts, including fog and edge, the problem of computational resource spending is considered as one of the efficiency criteria. In the current paper the improved model for computational tasks distribution efficiency estimation is presented and discussed. As the failure rate of the node depends on the workload, we consider the strategy, when each node can choose its regime - to transmit or to process data. The decision depends on the estimation inequality, which includes such parameters as computational complexities of data processing, data transmission and time share of the data transmission in the overall time constraint for the tasks performing. The model developed allows to implement the greedy strategy of tasks distribution, in which every robotic device chooses the best individual state and differs from the previously presented model by more precise estimations of the data transmission. Also, some selected experimental results are presented, pros and cons of such greedy approach are discussed. Keywords: Distributed Robotics · Tasks Distribution · Reliability · Fog Robotics

1 Introduction The latest trends in robotics, such as increase of the robotic things interconnection level, an application of machine learning techniques at the robotic edge, tasks offloading and the virtualization, make robotics and IoT much more connected than ever. Such intersection of domains brings the well-known problems of IoT edge-cloud continuum to the distributed robotics. Having the robot limited on-board computational capacity, the problems of energy consumption, optimal resource allocation and tasks distribution take place as well, especially for complex computational tasks – mapping, learning, routing. Such problems are considered in [1–3]. In this paper a problem of computational resource spending for the distributed robotic systems, including fog- and edge ones, is considered. As every robot in the system has its own computing power, it can be considered as computational node/device with its © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 311–321, 2023. https://doi.org/10.1007/978-3-031-43111-1_28

312

A. Klimenko

own performance, failure rate, reliability function and average residual life. As is shown in [4], the non-optimal tasks distribution causes the average residual life to decrease and, consequently, the decrease of the expedient exploitation time of the device. It can be explained in a way as if some robotic nodes in the distributed robotic system process more data than others, it is presupposed that their lifetime shortens faster than it could. So, the problem of tasks distribution through the robotic system is topical and supposed to be solved. The major contribution of this paper is an improved model of tasks distribution efficiency estimation, which differs from the previously presented model by the new parameter addition. This parameter is the throughput of the data route, and its usage allows to estimate the time, which is used for the data transfer and data processing more precisely.

2 Robotics and Distributed Computing Domains Intersection A lot of problems in contemporary robotics are solved in a way as ones of distributed computing are. For example, the study [5] is one of the early ones in this field and uses the robotic cluster to solve complex tasks with the performance objective function. In this work SLAM problem is solved by the distributed robotic system, combined to the cluster of computational nodes. Multi-robot SLAM is solved in [6], and edge-cloud SLAM algorithm is proposed in [7] as well. SLAM problem in the aspect of edge- and cloud computing, including the tasks offloading issues, is considered in [8]. Routing is another complex computational problem, which often relates to the TSP or M-TSP [9]. Earlier the distributed computing was used to increase the accuracy and decrease the solution time of the TSP [10]. One more example of routing is [11], as well as [12]. In the study [13] computationally heavy routing task for the robot group is solved in a distributed manner. Yet, another emerging area of the edge- and fog- concepts application in the distributed robotics relates to the machine learning. Some studies consider offloading strategies in general [14–16], while some works are devoted to the robots learning processes [17–20]. Some studies also relate to the distributed tasks processing, including the following [21, 22]. So, the following general conclusions can be made: • The areas of robotics, fog- and edge- computing have been intersected and now are coupled tightly. • Numerous problems of robotics are solved with the usage of distributed computing mechanisms, including learning, SLAM, routing, and many others. • Despite of intensive usage of robots with limited computational capacity, no attention has been paid to the problem of resource-saving. The latter relates to the average residual life/reliability/failure rate of a computational robot part, i.e., affects the time of expedient device exploitation. Our previous study considers the model for the task distribution in the resourcesaving way [21].

Improved Model of Greedy Tasks Assignment

313

Reliability function value depends on the failure rate of the computational node, while failure rate is connected to the device temperature and, consequently, workload: T

λ = λ0 · 2 10 ,

(1)

where λ is a resulting failure rate, λ0 is the failure rate under conditions of unloaded device, ΔT is the temperature difference between the temperature of unloaded device and the temperature of loaded one. Also, the coefficient can be determined, which connects the node temperature and the workload: kD

λ = λ0 · 2 10 .

(2)

Consequently, the reliability function is determined as follows: P(t) = e−λt = e−λt·2

kD/ 10

,

(3)

where D is the node workload. Average residual life is estimated as follows: R=

1 . λ

(4)

Consider the node workload for the node 3 (Fig. 2) as follows (for the case when the node transits some data): D=

wi , pj ttransfer

(5)

where wi is the computational complexity of the task, t transfer is the time needed for the data transfer. Consider the data processing shift discussed in [21], where the data transmission to the cloud takes place (Fig. 1).

Fig. 1. Data processing shift illustration.

With the data processing shift the workload of the processing node is: D=

wi . pj tprocess

(6)

314

A. Klimenko

Obviously, with the data shift tprocess is bigger than ttransfer , and so D decreases. And, finally the following “greedy” rule has been formed in [21]: wreceice + wprocess + wsend < 2

wreceive , x

(7)

where x is the time fraction for the time of data transfer process of the particular node. However, at least one drawback of this model can be considered, that is the absence of estimations of time fraction x, which is needed for data transfer. Of course, this model is sufficient to prove the expediency of the data processing near the data source, yet, the accuracy of the estimations does not allow to use this model, for example, when we have a route consisted of robots, as well as we cannot get accurate estimations of how long the data will be transferred through this route. In this paper an improved model is presented, which considers the network throughput parameter. It allows to get more precise estimation of expediency of the data processing on the particular node.

3 Tasks Distribution Efficiency Estimation Consider the robotic network as a set of connected robots, which have to transfer some data to the base station for processing. It is obvious that sometimes data have to be send through some intermediate robotic nodes – or, vice versa – to be processed somewhere in the robotic group, while the base station gets the results only. When data are processed somewhere in the robotic group, this data processing can be distributed among the robots with low computational capacity – actually, it reduces the robotic group latency. Computing resources of robots are spent much faster, and there is a need to search a kind of equilibrium between computational resource spending and the system latency. However, there is a possibility to save some extra-time and use it for data processing distribution to prevent the computational resource spending, and, as a result, there must be the estimation method to choose what it is better for the particular robot: to transmit data or to process. As is mentioned earlier, the workload of the computational nodes can be estimated as is in Eqs. (5, 6). Consider the route for data transfer, presupposing the fact that the velocity of the data transfer by the node is constrained by the throughput of the network channel. So, if we have some nodes in the route, and the data transmits without accumulation on the nodes, we can esteem the time needed for data transmission to the processing node, as well as we can esteem the time for data processing. Assuming the t 0 as the tasks processing completion time, the following describes the data transmission time and the data processing time: k−1  tmax , (7) tprocess = t0 − i=0

where tmax is the minimum of the data transfer velocity in the route. Then, the workload of the data processing on the node k is as follows: Dk =

res wreceive + wop + wsend , pk (t0 − (k − 1)tmax

(8)

Improved Model of Greedy Tasks Assignment

315

where wreceive is the computational complexity of the data receive procedure; W op is the res is the computational complexity computational complexity of the data processing; wsend of the sending of the data processing result; Pk is the performance of the node k; t0 is the time constraint; tmax is the maximum time of the data transfer in the route. The workload of the node, which just transmits the data is as follows: D k =

2wreceive . pk tmax

(9)

Taking into account the Eqs. (2, 3), it is expedient to process the data on the node k, when the following inequality is fair: Dk < D k ,

(10)

res wreceive + wop + wsend 2wreceive < . (t0 − (k − 1)tmax tmax

(11)

So, for the node k, having the information about the minimum node throughput, the data processing becomes expedient, when the inequality (11) is fair. In case when every node accumulates data and then sends it further, the Eq. (11) transforms into the following one: res wreceive + wop + wsend 2wreceive < , i−1 ti (t0 − j=1 tj )

(12)

where ti is the time of data transfer of the selected node in the route, tj is the time of data transfer of the previous node in a route. It must be mentioned that according to the inequality (11) the maximum increase of the reliability function is when the nearest to the data source node processes a task. It can be checked, if we transform the Eq. (11) into the non-linear integer optimization problem, with the objective function as is shown in Eq. (13): res wreceive + wop + wsend 2wreceive → max. − tmax (t0 − (k − 1)tmax .

(13)

So, the inequalities (11, 12) allow to estimate the node’s proficiency in case of data transmission or, vice versa, data processing. If a node estimates its preferences in this way, it takes a task for a processing, or transmits data somewhere else to the nodes, which are farer from data source. Such approach forms a kind of “egoistic” strategy, in which every node choses the best individual strategy without paying attention to other nodes state and workload. It must be mentioned as well, that within such strategy the node can choose from “data transmission” and “data processing” only and has very limited knowledge about the set of nodes, which participate in computational process. However, such “egoistic” approach is quite suitable for single-node offloading, or when there is a need to select the processing node for the consequently upcoming tasks. Next section contains the experimental results for estimation of developed estimation application prospects.

316

A. Klimenko

4 Experimental Results To check the possibility of inequality (11) application to the tasks assignment in the resource-saving manner, consider the following data transfer routes within the robot group/distributed robotic system (Figs. 2, 3).

Fig. 2. Network route structure №1.

Fig. 3. Network route structure №2.

Consider the node 0 as a data source, nodes 1–5 as nodes which are capable of data processing and weighed with their performances. The aim of the experiment is to check if our estimation allows to assign the data processing task in an optimal way according to the given parameters. 4.1 Experiment 1. Small Data Volume to Transmit, Route №1 We consider the following variable values: wreceive = 50; wop = 200; wsend = 50; t0 = 320 (modeling units); the ribs of the graph is weighed with the following: 100, 100, 50, 35, 10 (modeling units), which determine the data transfer velocity between the nodes. For this case the following values of Eq. (11) are got (Table 1). One can see that with the given values of variables nodes 2, 3, 4 suppose that it is more profitable for them to transfer the data somewhere else than to process it, taking into account data transfer delays. So, data processing can be assigned to the node 1 or to the node 5. Despite that fact that node 5 is the most distant one from the data source, it is suited for data processing due to its performance, which is the highest within the node set. 4.2 Experiment 2. Large Data Volume to Transmit, Route №1 We consider the following variable values: wreceive = 5000; wop = 200; wsend = 50; t0 = 500 (modeling units); the ribs of the graph is weighed with the following: 100, 100, 50, 35, 10 (modeling units), which determine the data transfer velocity between the nodes. One can see that this case is for the large data volume processing and transfer (Table 2). The results conducted are quite interesting because of the nodes overloading in the data transmission regime. One can see that if the data are transmitted with the

Improved Model of Greedy Tasks Assignment

317

Table 1. Results of experiment 1. 1

Data transfer workload, node1

Data processing workload, node1

Data transfer R, node1

Data processing R, node1

0.0625

0.02

0.01875

9330.33

9370.838

2

Data transfer workload, node 2

Data processing workload, node 2

Data transfer R, node 2

Data processing, R, node 2

−0.0303

0.026667

0.05

9117.225

8408.964

3

Data transfer workload, node 3

Data processing workload, node 3

Data transfer R, node 3

Data processing R, node 3

−0.1470

0.04

0.042857

8705.506

8619.728

4

Data transfer workload, node 4

Data processing workload, node 4

Data transfer R, node 4

Data processing, R, node 4

−0.1587

0.014815

0.028571

9499.516

9057.237

5

Data transfer workload, node 5

Data processing workload, node 5

Data transfer, R, node 5

Data processing, R, node 5

1.42857

0.012

0.01

9707.306

9659.363

Table 2. Results of experiment 2. 1

Data transfer workload, node1

Data processing workload, node1

Data transfer, R, node 1

Data processing, R, node 1

89.5

2

0.21



4829.682

2

Data transfer workload, node 2

Data processing workload, node 2

Data transfer, R, node 2

Data processing, R, node 2

120.20

2.666667

0.35



2973.018

3

Data transfer workload, node 3

Data processing workload, node 3

Data transfer, R, node 3

Data processing, R, node 3

217.79

4

0.21



4829.682

4

Data transfer workload, node 4

Data processing workload, node 4

Data transfer R, node 4

Data processing R, node 4

423.44

1.481481

0.081395



7542.022

5

Data transfer workload, node 5

Data processing workload, node 5

Data transfer R, node 5

Data processing R, node 5

975.58

1

0.024419

312.5

9188.536

maximum velocity of the channel (and this allows to save time for data processing), the computational devices are overloaded, though all of the nodes successfully process the same large data volumes. Such situation is a good proof for fog concept application and its benefits from the resource-saving.

318

A. Klimenko

4.3 Experiment 3. Small Data Volume to Transmit, Route №2 We consider the following variable values: wreceive = 50; wop = 250; wsend = 50; t0 = 500 (modeling units); the ribs of the graph is weighed with the following: 10, 35, 50, 100, 100 (modeling units), which determine the data transfer velocity between the nodes. One can see that this case is for the large data volume processing and transfer (Table 3). Table 3. Results of experiment 3 1

Data transfer workload, node 1

Data processing workload, node 1

Data transfer, R, node 1

Data processing, R, node 1

3.7444

0.008889

0.0014

9696.631

9951.597

2

Data transfer workload, node 2

Data processing workload, node 2

Data transfer, R, node 2

Data processing, R, node 2

1.6386

1.638655

0.004706



9946.823

3

Data transfer workload, node 3

Data processing workload, node 3

Data transfer, R, node 3

Data processing, R, node 3

0.5641

0.564103

0.02

9330.33

9851.362

4

Data transfer workload, node 4

Data processing workload, node 4

Data transfer, R, node 4

Data processing, R, node 4

0.1358

0.01

0.011475

9659.363

9610.097

5

Data transfer workload, node 5

Data processing workload, node 5

Data transfer, R, node 5

Data processing, R, node 5

−0.1475

0.01

0.011475

9659.363

9610.097

It is seen here that such strategy can produce not a good solution: while node 5 supposes that it will better for it to transfer data somewhere else, this can lead to the solution loss, because the farer node has less time for the data processing than its predecessor. Also, node 2 is overloaded with the data transfer regime, because of high channel throughput and, according to our assumption, the need to transmit the data as fast as it is possible. 4.4 Experiment 4. Large Data Volume to Transmit, Route №2 We consider the following variable values: wreceive = 5000; wop = 250; wsend = 50; t0 = 500 (modeling units); the ribs of the graph is weighed with the following: 10, 35, 50, 100, 100 (modeling units), which determine the data transfer velocity between the nodes. One can see that this case is for the large data volume processing and transfer (Table 4). With the need to transfer a large amount of data, it is seen that with the high performance of the nodes it is profitable to process the data. Nodes 3, 4, 5 are overloaded for the data transmission regime, while data processing is preferable.

Improved Model of Greedy Tasks Assignment

319

Table 4. Results of experiment 4 1

Data transfer workload, node 1

Data processing workload, node 1

Data transfer, R, node 1

Data processing, R, node 1

433.844

0.888889

0.0212

459.292

9291.607

2

Data transfer workload, node 2

Data processing workload, node 2

Data transfer, R, node 2

Data processing, R, node 2

224.477

0.470588

0.023297

1957.466

9224.333

3

Data transfer workload, node 3

Data processing workload, node 3

Data transfer, R, node 3

Data processing, R, node 3

121.685

2

0.065432



7971.038

4

Data transfer workload, node 4

Data processing workload, node 4

Data transfer, R, node 4

Data processing, R, node 4

86.9135

1

0.17377

312.5

5475.822

5

Data transfer workload, node 5

Data processing workload, node 5

Data transfer, R, node 5

Data processing, R, node 5

82.6229

1

0.17377

312.5

5475.822

So, as a result, the greedy approach of tasks assignment can be formulated. The initial conditions are as follows: • The unit, which holds the control center functions, have to assign the new task to the one of the nodes within the given distributed robotic system; • The system of robots is distributed geographically such as the routes of data transmission can be formed; • The amounts of data to transfer, the computational complexities of tasks, the network throughput are known as well. The scenario of the tasks assignment is as follows: 1. The robot, which receives some sensor data, requests its neighbours for free resources to perform a new task; 2. For all answers the estimations (11–12) are made; 3. The robot with the highest estimation value is chosen. The similar strategy can be applied to the situation when some tasks must be offloaded from one robot to another. The robot requests its neighbours and offloads its task to that one, which has the maximum result of estimation inequality (11).

5 Conclusion In this paper the previously developed model for resource-saving tasks assignment is enhanced and specified with the new parameter related to the time of data transfer. This new model does not conflict with the previous one, clarifying the way to estimate the data transmission time.

320

A. Klimenko

Greedy approach to the tasks assignment, as experimental results have shown, is applicable to the process of tasks offloading or single/consequent tasks assignment in the distributed robotic systems. Yet, as every greedy approach, the developed model and strategy have some drawbacks: • There can be solutions, when the refusal of the particular robot to perform the task leads to the loss of potentially acceptable task distribution; • as only one robot is considered for making a solution, to process or to transmit data, other robots workload and resources are out of consideration, and potentially this leads to the uncontrolled resource spending among other robots, if, for example, there are more than one data transmission routes. The latter is an important feature of the developed approach: individual profit does not guarantee the common one. Besides, such strategy does not take into account the length of data transfer route: with the growth quantity of robots in the routes, more robots have to perform an additional task of data transfer, which is quite complex in case of large sensor data volumes. It leads to the following decision making: to use n nodes with workload D1, or to use n-k nodes with workload D2, D2 > D1. Such issues demand further consideration and research.

References 1. Avgeris, M.: dynamic resource allocation and computational offloading at the network edge for internet of things applications. PhD thesis (2021) 2. Afrin, M., Jin, J., Rahman, A., Gasparri, A., Tian, Y.-C., Kulkarni, A.: Robotic edge resource allocation for agricultural cyber-physical system. IEEE Trans. Netw. Sci. Eng. 9(6), 3979– 3990 (2022). https://doi.org/10.1109/TNSE.2021.3103602 3. Natsuho, S., Ohkawa, T., Amano, H., Sugaya, M.: Power consumption reduction method and edge offload server for multiple robots. In: Zhang, L.-J. (ed.) EDGE 2021. LNCS, vol. 12990, pp. 1–19. Springer, Cham (2022). https://doi.org/10.1007/978-3-030-96504-4_1 4. Melnik, E., Klimenko, A.: A condition of reliability improvement of the system based on the fog-computing concept. J. Phys. Conf. Ser. 1661, 012007 (2020). https://doi.org/10.1088/ 1742-6596/1661/1/012007 5. Gouveia, B.D., Portugal, D., Silva, D.C., Marques, L.: Computation sharing in distributed robotic systems: a case study on SLAM. IEEE Trans. Autom. Sci. Eng. 12, 410–422 (2015) 6. Zhong, S., Qi, Y., Chen, Z., Wu, J., Chen, H., Liu, M.: DCL-SLAM: a distributed collaborative LiDAR SLAM framework for a robotic swarm. arXiv:2210.11978 (2022). https://arxiv.org/ abs/2210.11978 7. Lv, T., Zhang, J., Chen, Y.: A SLAM algorithm based on edge-cloud collaborative computing. J. Sens. 2022, 1–17 (2022). https://doi.org/10.1155/2022/7213044 8. Huang, P., Zeng, L., Chen, X., Luo, K., Zhou, Z., Yu, S.: Edge robotics: edge-computingaccelerated multi-robot simultaneous localization and mapping. IEEE Internet Things J. 9, 1 (2022) 9. Liu, C., Zhang, Y.: Research on MTSP problem based on simulated annealing. In: ICISS 2018: Proceedings of the 2018 International Conference on Information Science and System, pp. 283–285 (2018). https://doi.org/10.1145/3209914.3234638

Improved Model of Greedy Tasks Assignment

321

10. Nishi, T., Mori, Y., Konishi, M., Imai, J.: An asynchronous distributed routing system for multi-robot cooperative transportation. In: 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS, pp, 1730–1735 (2005). https://doi.org/10.1109/IROS.2005. 1545268 11. Camisa, A., Testa, A., Notarstefano, G.: Multi-robot pickup and delivery via distributed resource allocation. IEEE Trans. Robot. 39, 1106–1118 (2022) 12. Guo, Y., Wang, Y., Qian, Q.: Intelligent edge network routing architecture with blockchain for the IoT. Chin. Commun. 1–14 (2023) 13. Seisa, A., Satpute, S., Nikolakopoulos, G.: A Kubernetes-based edge architecture for controlling the trajectory of a resource-constrained aerial robot by enabling model predictive control (2023) 14. Wu, S., Xue, H., Zhang, L.: Q-learning-aided offloading strategy in edge-assisted federated learning over industrial IoT. Electronics 12(7), 1706 (2023) 15. Zhao, P., Yang, Z., Mu, Y., Zhang, G.: Selfish-aware and learning-aided computation offloading for edge-cloud collaboration network. IEEE Internet Things J. 10(11), 9953–9965 (2023) 16. Yang, Z., Zhong, S.: Task offloading and resource allocation for edge-enabled mobile learning. Chin. Commun. 20, 326–339 (2023) 17. Felbrich, B., Schork, T., Menges, A.: Autonomous robotic additive manufacturing through distributed model-free deep reinforcement learning in computational design environments. Constr. Robot. 6, 1–23 (2022) 18. Esteves, L., Portugal, D., Peixoto, P., Falcao, G.: Towards mobile federated learning with unreliable participants and selective aggregation. Appl. Sci. 13, 3135 (2023). https://doi.org/ 10.3390/app13053135 19. Jayaratne, M., Alahakoon, D., Silva, D.: Unsupervised skill transfer learning for autonomous robots using distributed growing self organizing maps. Robot. Auton. Syst. 144, 103835 (2021). https://doi.org/10.1016/j.robot.2021.103835 20. Gamboa, J., Alonso-Martin, F., Marques, S., Sequeira, J., Salichs, M.: Asynchronous federated learning system for human-robot touch interaction. Expert Syst. Appl. 211, 118510 (2023) 21. Klimenko, A.: Model and method of resource-saving tasks distribution for the fog robotics. In: Ronzhin, A., Meshcheryakov, R., Xiantong, Z. (eds.) Interactive Collaborative Robotics. ICR 2022. Lecture Notes in Computer Science, vol. 13719. Springer, Cham (2022). https:// doi.org/10.1007/978-3-031-23609-9_19 22. Meshcheryakov, R.: Information processing methods in Ergatic robotic systems: In: International Conference Engineering and Telecommunication (En&T), Dolgoprudny, Russian Federation, pp. 1–4 (2021). https://doi.org/10.1109/EnT50460.2021.9681750

Construction of a Three-Dimensional UAV Movement Planner When the Latter Moves in Conditions of Difficult Terrain Vladimir Kostyukov(B)

, Igor Evdokimov, and Vladislav Gissov

Joint Stock Company “Scientific-Design Bureau of Robotics and Control”, Taganrog 347900, Russia [email protected]

Abstract. The known methods of planning the routes of movement of robotic platforms based on cellular decomposition of the area of movement in a threedimensional formulation are severely limited in speed. Therefore, the construction of high-speed planning algorithms in a three-dimensional mapped environment is an urgent task. This article proposes a method for planning the movement of robotic platforms in this environment, combining the use of the well-known Dijkstra algorithm for constructing a two-dimensional projection curve with subsequent projection reconstruction and multi-stage correction of the target spatial piecewise polyline curve. The restoration of the original spatial curve by its twodimensional projection onto the horizontal plane is performed on the basis of a given discrete elevation map of the motion area, and the specified adjustment is made taking into account the requirements, firstly, the minimality of the total length of the final piecewise polyline, and, secondly, taking into account the specified known kinematic limitations of the apparatus. The algorithm for the synthesis of a spatial curve is detailed for the common case when obstacles are represented in the form of rectangular cylinders with polygonal generators. The effectiveness of the developed global scheduler algorithm is confirmed by the results of numerical modeling. Keywords: Methods of Planning Robot Movements · Complex Mapped Environment · Smoothing of Piecewise Linear Trajectory · Kinematic Limitations of the Robot

1 Introduction The construction of a software trajectory of a robotic platform moving in a threedimensional mapped environment can be carried out using a wide range of displacement planning algorithms, among which cellular decomposition methods can be distinguished, primarily Dijkstra algorithms, A* [1, 2], D* [3]. Dijkstra’s algorithm allows us to obtain the optimal path up to the quantization error of the grid structure. Methods A*, D* are not optimal. A common disadvantage of algorithms based on the cellular decomposition method is a high demand for memory, as well as a sharp increase in computational complexity © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 322–333, 2023. https://doi.org/10.1007/978-3-031-43111-1_29

Construction of a Three-Dimensional UAV Movement Planner

323

with an increase in the number of partition cells. To reduce the memory requirements of the A* algorithm and its computational complexity, such modifications as the algorithm with iterative deepening [4], A* with memory restriction [5], hierarchical A* [6], A* with dynamic change of edge weights [7, 8] have been developed. Nevertheless, when solving the problems of path planning in three dimensions, even existing modifications of these algorithms are not enough to ensure the required performance. Algorithms based on the methods of potential fields [9–12] allow to increase the speed by orders of magnitude, but their disadvantage is the possibility of the existence of local minima of the potential function, which entails the problem of providing an exit from the corresponding areas of looping. There are known approaches [12, 13] that offer certain algorithms for detecting and exiting local minima zones, but what they have in common is the lack of a sufficient experimental base confirming their operability. This problem can be leveled to a large extent by applying the method of unstable modes based on the use of the dynamics of the object when planning to bypass sudden obstacles, and the stability of the corresponding control algorithms in which such a scheduler is implemented is shown [14, 15]. However, this method is highly dependent on the adequacy of the used dynamic model of a moving object and the accuracy of setting its parameters. The second problem of the algorithms of the potential fields method is the formation of a non-smooth trajectory as a result of their application, often including areas with oscillations [16–19]. To solve this problem, a method of two-stage smoothing of trajectories with harmful (noise) oscillations in two dimensions is proposed in [12]. In the same work, an algorithm is proposed for finding a special time parametric representation of the target trajectory. The approach to planning movements in three dimensions developed in this article combines the use of Dijkstra’s algorithm for constructing a two-dimensional projection curve and a number of algorithms presented below for restoring the target spatial curve and correcting it in order to minimize its length in compliance with the existing kinematic limitations of the apparatus.

2 Problem Statement Suppose there is some terrain map and a height map linked to it, constructed with a given discreteness. The initial S and target G points of the route are given. There are also obstacles presented in the form of cylindrical areas with polygonal sections, and the guides of these cylinders are all orthogonal to the horizontal plane. Restrictions are set on: a) the height of the flight above the terrain; b) the kinematic characteristics of the device: horizontal and vertical components of linear velocities, pitch angle, maximum curvature of the spatial trajectory. It is required to construct the target trajectory of the vehicle from the starting point to the target one, taking into account the above restrictions, so that the length of this curve is as small as possible. At the same time, the optimization problem is not supposed to be solved in a strict formulation. At the output of the scheduler, nodes of the resulting spatial piecewise polyline curve and linear velocity vectors in them should be formed.

324

V. Kostyukov et al.

Figure 1 shows a fragment of a three-dimensional terrain map, where green parallelepipeds represent relief elements taking into account a given elevation map, a rectangular cylinder with a polygonal section shows an obstacle. The target trajectory of movement from the starting point (located in the lower left corner) to the target point is also shown.

Fig. 1. To the problem statement.

3 Generalized Global Scheduler Algorithm When constructing the global trajectory of the UAV in conditions of complex terrain, we will start from the possibility of constructing an appropriate two-dimensional trajectory is a projection curve along which the projection of the origin of the coordinates of the local system of the device moves on a conditional horizontal plane. Let’s consider a generalized algorithm of the global scheduler. The three-dimensional target spatial trajectory is constructed in two stages. At the first stage, a two-dimensional projection of the desired spatial trajectory is constructed using Dijkstra’s algorithm. The resulting piecewise polyline trajectory is further minimized in length using a special algorithm. Such optimization can be performed using a modified Ramer-Douglas-Pecker method. The second stage of the global algorithm involves the generation of the corresponding spatial piecewise polyline according to its two-dimensional projection, as well as its necessary subsequent correction. For each node of the curve, the corresponding target height is found, taking into account the heights of those relief elements whose twodimensional projections border on this point. After obtaining the resulting optimized piecewise polyline, it is necessary to smooth it. To do this, a modified Dubins method can be used, which consists in replacing the circular arc of parts of rectilinear segments adjacent to each given internal node of the polyline, as well as constructing a time parametric representation of the smoothed trajectory, taking into account the peculiarities of the change in the trajectory velocity of the apparatus when passing curved sections of the trajectory [12].

Construction of a Three-Dimensional UAV Movement Planner

325

After the smoothing stage, the global algorithm takes into account the kinematic limitations of the movement of the device along the linear velocity vector and pitch angle.

4 Algorithm for Thinning a Piecewise Polyline and Reducing its Total Length The points obtained after the global planning stage on a two-dimensional map are passed to this function. Next, the segments connecting the corresponding pairs of such points are iteratively considered, and a piecewise polyline is sought that would contain some nodes of the original one, would be as short as possible and would not intersect any of the obstacles. The algorithm divides the piecewise polyline into two parts at the current iteration each time. The input of the algorithm is the coordinates of all points between the first A1 and the last AN point of the polyline. The points A1 and AN remain unchanged. After that, the algorithm finds a point Ak (1 < k < N) that is furthest from the segment connecting these two extreme points. If such a point is located at a distance less than some positive, sufficiently small value ε > 0, then the algorithm outputs a piecewise polyline A1 Ak AN , which smoothes the original curve with the accuracy of ε. If the distance is greater than ε, then the algorithm recursively calls itself on the set from the initial A1 to the given Ak and from the given Ak to the end point AN (in this case, this point Ak will be marked for preservation). At the end of all recursive calls, the output polyline is constructed only from those points that were marked for saving. At each step, a check is made for the intersection of the current piecewise polyline with two-dimensional obstacles. If an intersection is detected, the simplification of the function is canceled on its corresponding linear section. The considered algorithm is a modification of the existing Ramer-Douglas-Pecker algorithm.

Fig. 2. Modified Ramer-Douglas-Pecker Method.

From the comparison of the initial and final piecewise polylines shown in Fig. 2, it can be seen that as a result of this algorithm, it was possible to reduce the number of nodes from eight to five with a noticeable decrease in the length of the original curve, but without losing its qualitative behavior.

326

V. Kostyukov et al.

5 Modified Dubins Method After constructing a spatial piecewise polyline trajectory, it is necessary to smooth it, which can be effectively performed using the modified Dubins method, which also allows to give smooth parametric time representations for both the target curve and its corresponding trajectory velocity. In [12], such a method is considered for two measurements. This algorithm is a modification of the well-known Dubins method, which reduces to the conjugation of adjacent rectilinear segments of a piecewise polyline by arcs of circles [20]. The problem of obtaining smooth trajectories often arises when planning methods involving obtaining spatial trajectories in the form of individual points. Various techniques are used to smooth out such trajectories. For example, the D* algorithms are characterized by smoothing of trajectories due to continuous linear interpolation of weights at the boundary of each cell, which, in general, leads to a significant increase in their operation time [18]. When planning the movement of the device, it is necessary to smooth out to exclude discontinuities of the derivatives of each of the coordinates in their parametric time representation. In this regard, the task arises of smoothing the initial piecewise linear trajectory in such a way as to exclude or minimize the stops of the device in the vicinity of the specified nodal points of the trajectory, which are impossible for some types of devices. To solve this problem, the use of the Bezier curve [21] is not suitable since it does not make it possible to simulate motion with deceleration in sections of the trajectory with increased curvature. Below we consider the modified Dubins method in two dimensions, and then generalize it to the case of three. Let be given: nodes of a piecewise linear trajectory {Ai }, i = 1, 2, ..,N; ; the target speed of the vehicle on linear sections of the trajectory (cruising) speed is V k , the degree of decline of η velocity at the conditional maximum rotation of the trajectory (by π rad); the minimum radius of curvature of the arc at the vertex of the original piecewise polyline Rmin determined by the kinematic constraints of the apparatus. At the output, it is required to obtain the nodes {Bk }, j = 1, 2, ..,K (K 0 is performed vver,k > vver,l,max , then we limit vver,k and simultaneously make a primary correction of the pitch angle: ∼ v ver,k

= vver,l,max ,

 ∼ ∼ ϑ k = arctg v ver,k /vgor,k . ∼

(7) (8)

  If vver,k  < vver,l,max , then the pitch angle at this stage is taken as the initial:

ϑ k = ϑ k. Similarly, the vertical  component of the speed is adjusted when the conditions are met: vver,k < 0, vver,k  > vver,s,max : ∼ v ver,k

  = sign vver,k vver,s,max .

(9)

Here and below in this paragraph, the “tilde” sign will mean the adjusted value. After that, we make a secondary adjustment of the pitch angle, directly based on the inequality (3). The pitch angle correlates with the horizontal and vertical components of

Construction of a Three-Dimensional UAV Movement Planner

329

the velocity. Therefore, if inequality (3) is not met, then the pitch angle can be adjusted either by changing the horizontal velocity or by changing the vertical one. First, to satisfy the inequality (6), we will try to change the horizontal velocity, ∼

leaving the vertical, calculated earlier in (10), the same. In this case, when ϑ k = ±ϑ pr the required horizontal velocity must be equal to: ∼      vgor0,k =  v ver,k /tg ϑ pr , (10) where ϑ pr = ϑ pr,l is for the case of lifting, and ϑ pr = ϑ pr,s for the case of lowering the height. However, in this case, the inequality for the horizontal velocity (8) may not be fulfilled, which will force us to switch to the option of changing the vertical velocity to ensure the desired pitch angle, namely, its reduction compared to the value in (7). In the latter case, the vertical velocity adjusted at this stage is equal to:

 ∼ ∼ (11) v ver,k = vgor,k tg ϑ k , ∼

with simultaneous execution ϑ k = ±ϑ pr . After the adjustments, the vertical component of the velocity and the pitch angle are within the acceptable ranges. In general, there remains the possibility that the horizontal velocity does not belong to a given range in accordance with inequalities (3). If inequality (3) is not fulfilled, we adjust the horizontal and vertical components of the velocity as follows: ⎧ ∼ ⎨ v gor,k = vgor,min if vgor,k ≤  vgor,min , ∼ (12) ∼ ∼ ∼ ⎩ v gor,k = vgor,max , v ver,k = v gor,k tg ϑ k if vgor,k ≥ vgor,max . It is obvious that the last correction does not lead to the exit of the vertical component of the velocity from its permissible range given by the inequality (3). Thus, the linear velocity vector and the pitch angle are corrected. ∼

Then it remains to determine the adjusted position of the second point Bk : ⎡    ⎤  ∼ ∼ ∼   ⎢ Bk−1 (1) + v gor,k (t k − t k−1 )(Bk (1) − Bk−1 (1))/ Bk − Bk−1 ⎥ ⎢ ⎥  gor  ⎥ ⎢  ∼  ⎥ ⎢  ∼ ∼ Bk = ⎢  ⎥. (13)  ∼ ⎢ Bk−1 (2) + v gor,k (t k − t k−1 )(Bk (2) − Bk−1 (2))/ Bk − Bk−1 ⎥ ⎢ ⎥  gor  ⎦ ⎣ ∼ ∼ Bk−1 (3) + sign(ϑ k ) v ver,k (t k − t k−1 ) Let’s consider the algorithm for accounting for kinematic constraints in its entirety. We organize a search of the original set of nodes {Bk }, k = 1, 2, .., K, set at points in time {t k }, k = 1, 2, .., K. We will not adjust the position of the first node. First, we

330

V. Kostyukov et al. ∼

find the corrected position of the second node by the first node B2 , using the previously considered methodology for taking into account speed limits and pitch angle restrictions (3)–(8) and equality (13). Next, we find the adjusted position of the third node in the ∼

same way B3 . Next, we proceed to the analysis of the three nodes with numbers k = 2, 3, 4, and speeds in them. We adjust the speed in the 4th node according to criteria (3)–(6), and ∼





Eq. (13). Repeat these steps for all three nodes Bk−2 , Bk−1 , Bk by k = 3, 4, …, K.

7 Simulation Results Let’s consider an example of building a target trajectory using the global scheduler algorithm described in the previous paragraphs. Calculations will be made in the Matlab application. The initial data for the test example is as follows. The positions of the start and target points: A = [0, 185534, 200], m; B = [389621, 779243, 50], m; recommended absolute altitude of the UAV - Hrec = 150, m; maximum absolute altitude of the UAV flight Hmax = 400, m; required height adjustment from the terrain - Hrec = 50, m; kinematic constraints: [ϑ pr,l , ϑ pr,s , vver,l,max , vver,s,max , vgor,min , vgor,max ] = [1.0472, rad; 1.0472, rad; 5, m/s; 5, m/s; 10, m/s; 40, m/s]. Figures 4, 5, and 6 show in the corresponding scenes the trajectories obtained at the considered five stages of this algorithm. Figure 4a below shows a piecewise polyline curve (in purple) obtained after applying the modified Ramer-Douglas-Pecker method. This curve has a length that is 6.8% less than the length of the original curve (shown in blue) entering the algorithm input. The results of the reconstruction of the heights of the preliminary spatial piecewise polyline trajectory according to the specified two-dimensional projection are shown in Fig. 4b. Here the desired curve is represented by a blue solid curve. Figure 5 shows the resulting spatial curve obtained after working out the correction module taking into account kinematic constraints discussed earlier. Finally, Fig. 6 shows the resulting piecewise polyline curve obtained as a result of the operation of the entire global scheduler algorithm. The results of numerical modeling carried out according to the global planning algorithm discussed in the previous sections are adequate to physical representations of the behavior of the UAV when moving over a complex terrain in conditions of cylindrical obstacles with complex generators, taking into account restrictions on the limit values of the key kinematic characteristics of the device.

Construction of a Three-Dimensional UAV Movement Planner

331

Fig. 4. Work of the scheduler at the first stage of the algorithm. Curve after optimization of a two-dimensional piecewise polyline (a) and reconstruction of the heights of the preliminary spatial piecewise polyline trajectory (b)

Fig. 5. A section of a piecewise polyline adjusted after the kinematic constraints accounting module.

332

V. Kostyukov et al.

Fig. 6. The resulting piecewise polyline curve.

8 Conclusion In this paper, we consider an algorithm for planning the trajectory of a robotic platform, primarily a UAV, in a mapped environment described by an array of terrain heights and including additional cylindrical obstacles with arbitrary polygonal generators. This algorithm combines the use of the well-known Dijkstra algorithm, which allows you to build optimal routes with accuracy up to the error of grid quantization, and special algorithms for reconstructing a spatial piecewise polyline curve from its two-dimensional projection with multi-stage correction according to the criterion of the minimum length of the polyline, followed by taking into account the specified kinematic limitations of the apparatus. This multi-stage correction implies, firstly, thinning of the nodes of the original twodimensional piecewise polyline obtained after working out the Dijkstra algorithm, in order to initially reduce the length of the curve; secondly, smoothing of the intermediate piecewise polyline is performed using the modified Dubins method, which implies obtaining smooth to the first derivative of time parametric representations of the corrected curve. Finally, consideration of the kinematic limitations of the apparatus and the corresponding final correction of the curve is made at the final stage and includes consideration of the limit values of the horizontal and vertical components of the linear velocity and acceleration vectors, as well as the pitch angle. The effectiveness of the developed global scheduler algorithm is confirmed by the results of numerical modeling. Note that the developed algorithm can be generalized to the case of dynamic obstacles. To do this, it is enough to replace the Dijkstra algorithm used to generate a two-dimensional projection trajectory with D*. Acknowledgements. The study was supported by the Russian Science Foundation Grant No. 22-29-00370, https://rscf.ru/project/22-29-00370/.

References 1. Hart, P.E., Nilsson, N.J., Raphael, B.A.: Formal basis for the heuristic determination of minimum cost paths. IEEE Trans. Syst. Sci. Cybernet. 4(2), 100–107 (1968)

Construction of a Three-Dimensional UAV Movement Planner

333

2. Piskorsky, D.S., Abdullin, F.H., Nikolaeva, A.R.: Optimization of the A-star path planning algorithm. Bull. SUSU. Comput. Technol. Control, Radio Electron. 20(1), 154–160 (2020). (In Russ.) 3. Stentz, A.: Optimal and efficient path planning for partially known environments. In: Intelligent Unmanned Ground Vehicles. The Springer International Series in Engineering and Computer Science, vol. 388, pp. 203–220 (1997) 4. Wang, Q., Hao, Y., Chen, F.: Deepening the IDA* algorithm for knowledge graph reasoning through neural network architecture. Neurocomputing 429, 101–109 (2021) 5. Zhou, R., Hansen, E.A.: Memory-bounded {A*} graph search. In: The Florida AI Research Society Conference (FLAIRS), pp. 203–209 (2002) 6. Holte, R., Perez, M., Zimmer, R., MacDonald, A.: Hierarchical A*: searching abstraction hierarchies efficiently. In: AAAI/IAAI, vol. 1, pp. 530–535 (1996) 7. Liu, B., Xiao, X., Stone, P.: A lifelong learning approach to mobile robot navigation. IEEE Robot. Autom. Lett. 6(2), 1090–1096 (2021) 8. Chen, B.Y., Chen, X.W., Chen, H.P., Lam, W.H.: Efficient algorithm for finding k shortest paths based on re-optimization technique. Transp. Res. Part E: Logistics Transp. Rev. 133, 101819 (2020) 9. Khatib, O.: Real-time obstacle avoidance for manipulators and mobile robots. Int. J. Robot. Res. 5(1), 90–98 (1986) 10. Platonov, A.K., Karpov, I.I., Kirilchenko, A.A.: The method of potentials in the problem of laying a route. M.: Preprint of the Institute of Applied Mathematics of the USSR Academy of Sciences, p. 27 (1974). (In Russ) 11. Filimonov, A.B., Filimonov, N.B.: Issues of motion control of mobile robots based on the potential guidance method. Mechatron. Autom. Control 20(11), 677–685 (2019). (In Russ.) 12. Kostyukov, V.A., Medvedev, M., Pshikhopov, V.H.: Planning the movement of ground robots in an environment with obstacles: an algorithm for constructing smoothed individual trajectories. Mechatron. Autom. Control 24(1), 33–45 (2022). (In Russ.) 13. Pshikhopov, V.K.H., et al.: Path planning for vehicles operating in uncertain 2D environments. Elsevier, Butterworth-Heinemann, p. 312 (2017) 14. Pshikhopov, V., Medvedev, M.: Decentralized management of a group of homogeneous moving objects in a two-dimensional environment with obstacles. Mechatron. Autom. Control 17(5), 346–353 (2016). (In Russ.) 15. Pshikhopov, V., Medvedev, M.: Group control of the movement of mobile robots in an uncertain environment using unstable modes. Comput. Sci. Autom. 5(60), 39–63 (2018) 16. Gaiduk, A.R., Martyanov, O.V., Medvedev, M., Pshikhopov, V.H., Hamdan, N., Farhud, A.: Neural network control system for a group of robots in an uncertain two-dimensional environment. Mechatron. Autom. Control 21(8), 470–479 (2020). (In Russ.) 17. Nazarahari, M., Khanmirza, E., Doostie, S.: Multi-objective multi-robot path planning in continuous environment using an enhanced genetic algorithm. Expert Syst. Appl. 115, 106– 120 (2019) 18. Hoy, M., Matveev, A.S., Savkin, A.V.: Algorithms for collision-free navigation of mobile robots in complex cluttered environments: a survey. Robotica 33(3), 463–497 (2015) 19. Shlyakhov, N.E., Vatamaniuk, I.V., Ronzhin, A.L.: Review of the methods and algorithms of a robot swarm aggregation. Mechatron. Autom. Control 18(1), 22–29 (2017) 20. Sapronov, L., Lacaze, A.: Path planning for robotic vehicles using generalized Field D. In: Unmanned Systems Technology X, vol. 6962, pp. 447–458. SPIE (2008) 21. Grigor’ev, M.I., Malozemov, V.N., Sergeev, A.N.: Bernstein polynomials and composite Bézier curves. Comput. Math. Math. Phys. 46, 1872–1881 (2006)

Identification of the Quadcopter Rotational Dynamics for the Tilt Angle Vadim Alexandrov(B)

, Ilya Rezkov , Dmitrii Shatov , and Yury Morozov

V.A. Trapeznikov Institute of Control Sciences of Russian Academy of Sciences, 65, Profsoyuznaya Street, Moscow 117997, Russia [email protected]

Abstract. The pitch and roll angles of the quadcopter attitude are controlled by torque arising from the difference in thrust of different rotors. Rotational dynamics of the quadcopter as a rigid body is considered. The pitch and roll angles that form the quadcopter tilt angle have similar dynamics for the case of symmetric quadcopter frame, so the pitch angle is studied separately in the paper. The finite frequency identification approach is used to find the transfer function from the experimental flight data. The approach needs data, when sine wave test signals are fed to the system input. In the case of the unstable loop of the quadcopter tilt angle, the closed loop identification procedure is used where the test signal is added to the setpoint of the operating controller. As a result, a more complex model is identified than is commonly used. Moreover, the nonlinear model of the quadcopter pitch angle is considered, taking into account the translation velocity. Parameters of this model are estimated from the same experimental data using the nonlinear grey-box model parameters identification procedure. The transfer function founded from this model confirms the structure of the transfer function obtained via the finite frequency identification procedure. Keywords: System Identification · Robotic System · Quadcopter Dynamics · Transfer Function · Nonlinear Model · Closed Loop Identification

1 Introduction The identification of robotic systems is an important problem to obtain a system model from experimental data for simulation and control design purposes. So, some parameters of the differential equations system that describes the quadcopter flight dynamics cannot be obtained from its design, which leads to the need for experimental studies. Several control loops operate simultaneously to provide the quadcopter flight. The model identification for altitude control loop has been studied in [1, 2]. The transfer function from the horizontal projection of a full thrust to the horizontal velocity of the quadcopter was experimentally identified, and aerodynamic drag was detected in [3]. The horizontal part of thrust is determined by the quadcopter tilt angle, which is a combination of pitch and roll angles. The pitch and roll angles control is one of the main loops for quadcopter flight control that provides the direction and value of the horizontal velocity. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 334–344, 2023. https://doi.org/10.1007/978-3-031-43111-1_30

Identification of the Quadcopter Rotational Dynamics

335

The problem of quadcopter model identification was studied in several research using the CIFER software, where the special test signal called frequency sweep is applied to the real quadcopter, then the corresponding frequency response is found, and coefficients of the linearized quadcopter models are identified. In [4] the linearized quadcopter model in form of the several transfer functions was identified, in particular, for pitch and roll angular velocities there were obtained the third order systems with pure differentiation plus time delay and the poles were stable, one was real, and the rest were complex conjugate. The oscillating second order transfer function was identified in [5] for roll and pitch loops, when looking for a low order model. The authors of [6] identified quadcopter model in state-space form basing on the results of transfer function identification. In [7] the rotors dynamics as the first order with delay is added into the state-space quadcopter model. The article [8] develops a method of linear model identification in state-space form for a multirotor vehicle based on the equation-error maximum likelihood estimator which reduces to least-square approach applied to experiment data in time domain. The real-time identification method for a quadcopter model was proposed in [9]. Here, the recursive Fourier transform coupled with least-square estimation was used to find the state-space model coefficients estimates. The proposed method uses multisine excitation signal added to the full motor command, which is similar to our identification approach. A survey of the quadcopter model identification is in [10]. Thus, finding a practically acceptable linear model for the nonlinear dynamics of the quadcopter is a non-trivial problem. Identification of the quadcopter rotational dynamics model from the control signals, which provide the torques related to the body frame axes, to the Euler angles of pitch and roll is the purpose of the paper. Only the pitch angle loop is considered because the roll angle loop dynamics is similar. The yaw dynamics, which is the third Euler angle, is not considered in the paper. Section 2 describes the application of the finite-frequency identification approach [11] for finding transfer functions of the linearized models from control signal to the pitch angle. The structure of the nonlinear model and its parameters evaluation are considered in Sect. 3. Conclusions are presented in Sect. 4.

2 Finite Frequency Identification Case The experimental quadcopter [1] is the frame S500, 10-inch propellers, the onboard flight controller Pixhawk 4 and the standard ArduPilot software suite. Control signals for each motor for symmetric X-frame type quadcopter are formed as u1 u2 u3 u4

= uF = uF = uF = uF

− 0.5uroll + 0.5uroll + 0.5uroll − 0.5uroll

− 0.5upitch + uyaw , + 0.5upitch + uyaw , − 0.5upitch − uyaw , + 0.5upitch − uyaw ,

(1)

where uF is the control value for thrust providing the altitude hold, uroll , upitch , and uyaw are controls for torques related to body frame axes x, y and z. It is assumed that u1,..,4 ∈ [0, 1],

336

V. Alexandrov et al.

and these values are then scaled for transmission to the motors controllers called electronic speed controllers (ESC). Our experimental quadcopter is almost symmetrical (see Fig. 1). Only the position of the battery along the x axis forms some difference between the x and y axes. Therefore, the rotational dynamics around y axis only will be considered below. The roll loop dynamics is similar to the pitch’s one, and yaw loop is not considered in the paper. Note, that the Euler’s angles values are not measured directly, but obtained from the extended Kalman filter implemented in ArduPilot flight software for onboard control system. Thus, the identification problem is to find a transfer function from control upitch to the pitch angle θ . The finite-frequency identification approach is based on analysis of the experimental data when the sine wave test signals on several different frequencies from wide enough range are fed to the system input. The quadcopter cannot be tested in open loop mode, since even in manual flight mode the attitude controllers must be active to provide horizontal stabilization. In the case of closed loop identification [1–3] the sine wave test signals are added to the controller setpoint. The preinstalled AltHold flight mode of ArduPilot onboard control system is used for the test. In this mode the throttle is automatically controlled to maintain the current altitude [12], and the attitude controller provides the desired values of pitch and roll angles and yaw angular velocity. The required values of the pitch, roll and yaw angles θ ∗ , ϕ ∗ , ψ ∗ for attitude controllers in our identification experiment of the pitch angle loop should be the follows one: θ ∗ = ηi sin(ωi (t − t0 )), ϕ ∗ = 0, ψ ∗ = 0,

y pitch

roll

(2)

x

yaw

z

Fig. 1. Body frame axes for X-frame type quadcopter.

where ηi and ωi are amplitude and frequency of the test signal (i = 0, .., nω , where nω is a number of the test frequencies from a given test set that are fed consequently to the system input), t is time and t0 is the start time of the experiment on the current frequency. Then, upitch , uroll and uyaw are the outputs of the corresponding controllers that stabilize the roll and yaw angles values and provide the sinusoidal variation of the pitch angle. Authors’ procedures are added to the open-source onboard software to form the test signals (2) and collect the experimental data. The finite-frequency identification procedure consists of two steps: 1) obtaining estimates of frequency response for several test frequencies ωi and 2) finding estimates

Identification of the Quadcopter Rotational Dynamics

337

of coefficients of the identified transfer function using the frequency response estimates. The frequency response estimates are obtained as P(jωi ) = αi + jβi =

αy i + jβy i , i = 1, .., nω , αui + jβu i

(3)

where αi and βi are called the frequency parameters, which are found using output αy i , βy i and input αui , βu i values estimated during experiments by Fourier filter formulae [1, 2, 11]:  t0 +Ti 2 θ (t) sin(ωi (t − t0 ))dt, αˆ yi = ηi Ti t0  t0 +Ti 2 θ (t) cos(ωi (t − t0 ))dt, βˆyi = ηi Ti t0  t0 +Ti (4) 2 upitch (t) sin(ωi (t − t0 ))dt, αˆ ui = ηi Ti t0  t0 +Ti 2 βˆui = upitch (t) cos(ωi (t − t0 ))dt, ηi Ti t0 i = 1, . . . nω , where Ti = 2π q/ωi is the filtration time (experiment duration), where q ∈ N is the number of periods of the test frequency. Examples of the test data for the pitch loop are shown in Figs. 2, 3. The frequency parameters values obtained from experimental flights are presented in Table 1. Table 1. Experimentally obtained frequency parameters estimates. 



i

Frequency, ωi , rad/s

Real part,α i

Imaginary part,βi

1

20.11

−22.07

16.41

2

16.01

−41.46

25.09

3

12.03

−80.72

36.01

4

8.00

−204.60

55.70

5

5.33

−390.45

122.72

6

4.00

−451.59

392.35

7

3.50

−399.43

331.84

8

3.00

−398.90

420.68

9

2.00

−117.65

418.49

10

1.00

−5.24

194.18

338

V. Alexandrov et al.

Fig. 2. Sine wave test with frequency 16 rad/s.

Fig. 3. Sine wave test with frequency 2 rad/s.

The second part of the finite-frequency identification procedure is to find a transfer function from obtained frequency parameters. The order of the identified transfer function should be less than the number of experimentally obtained frequency response estimates in Table 1. Then, the solution of the overdetermined system of algebraic equations provides estimates of the transfer function coefficients by the least squares criterion. If the transfer function order is unknown, then one can start with the most possible transfer function order and then decrease the order until there are no obviously cancellable poles and zeros in the identified transfer function left.

Identification of the Quadcopter Rotational Dynamics

339

The result for the pitch data from Table 1 is: P1 (s) =

2828 · (s − 0.156) .  (0.046s + 1)(0.18s + 1) s2 − 3.79s + 12.9

(5)

Fig. 4. Frequency response of identified pitch model (7) and points from experiments.

This transfer function determines the model structure and serves as an initial point for the optimization procedure that tries to find a better result for criterion [2]:      nω Pid (jωi ) − α i + jβi    , (6) min   i=1 χ α i + jβi  







where Pid (jωi ) are the frequency parameters values calculated using the identified transfer function, α i and βi denote the experimentally determined according to (3), (4) estimates of the frequency parameters, χ is the variables vector. The identified according to criterion (6) transfer function is: 

P2 (s) =



2654 · (s + 0.045) .  (0.048s + 1)(0.16s + 1) s2 − 3.86s + 12.4

(7)

It is shown in Fig. 4 that frequency response of the obtained function (7) matches the experimental data from Table 1. The structure of the identified transfer function (7) is unexpected. Therefore, a more detailed analysis of the quadcopter rotational dynamics is carried out below.

3 Nonlinear Model Identification The first principles consideration of the quadcopter rotational dynamics as a rigid body [10] for the separate pitch loop under assumption that the roll and yaw angles ϕ, ψ and the corresponding angular velocities ωx , ωz are equal zeros (i.e. interactions are not taken

340

V. Alexandrov et al.

into account) provides the differential equations: θ˙ = ωy , ω˙ y =

1 τpitch , Iy

(8) (9)

where θ is the pitch angle, ωy is the angular velocity around y axis, Iy is the body rotational inertia around the y axis, and τpitch is the torque from a difference of the rotors angular velocities 1,..,4 :  (10) τpitch = Kτ 2 2 + 4 2 − 1 2 − 3 2 , where Kτ is a proportional gain. Electronic speed controllers (ESC) of type BLHeli32 are installed on the experimental quadcopter, so the rotors angular velocities values i are measured. The transfer function from the controls ui , i = 1, .., 4 that are described by (1) to the rotor angular velocities i for (10) can be identified from the data of the Sect. 2 experiments via the finite-frequency identification procedure. The first order function with time delay is obtained: P (s) =

967.4 e−0.0042s 0.044s + 1

(11)

The Nyquist plot of (11) with the experimental points is shown in Fig. 5.

Fig. 5. Frequency response of identified the rotor angular velocity model (11) and points for each rotor from experiments.

This result is quite in line with expectations. The small delay is the sum of the cycle time of the flight controller 0.0025 s and certain delay in ESC. The gain of (11) corresponds to the rotor angular velocity in rad/s and ui ∈ [0, 1].

Identification of the Quadcopter Rotational Dynamics

341

Further, the static tests were made using weighing-machine to find relations between the control uF , average rotor angular velocity and thrust. These tests have confirmed that the ESC controller provides the rotor angular velocity proportional to the control value with the gain approximately equal to the gain in (11). Thrust is close to quadratic dependence on the rotor velocity with the gain KF = 0.0000115. It is known that the gain Kτ for torque is the product of this value and the lever of the rotors, which is measured as 0.17 m. Thus, Kτ = 1.955 · 10−6

(12)

and τpitch can be simulated by (10), where i are integrated from (11), while ui are determined by (1), uF corresponds to hovering and uroll , uyaw assumed to be zero. Thus, the system of Eqs. (1), (8) – (11) is a simple model for the pitch loop dynamics, where the only parameter Iy is undefined. The function nlgreyest from Matlab System Identification Toolbox with fmincon search method is used to estimate nonlinear greybox model parameters. The experimental data are obtained in the closed loop system, which model structure should be defined for the estimation procedure. The ArduPilot cascaded control system with proportional controller for the angle and PID controller with a filter for the angular velocity is added to the estimated system. The parameters of P and PID controllers are known. The input experimental data for the estimated system is the setpoint for the pitch angle. The outputs, which are used to compare results of experiment and simulation, are the pitch angular velocity ωy and control upitch . It is important for the case of closed loop system to use the experimental and simulated control values upitch for comparison, since the controller will provide a match of the controlled output for a wide range of parameters and model structures. The data set of 64000-time instances of 160 s with test sine waves from 8 to 1 rad/s is used. Results of identification procedure nlgreyest are shown in Fig. 6, where it is seen that for obtained value Iy = 0.0116 the simulated values of control upitch does not correspond at all to the experimental data. Therefore, the Eq. (9) does not represent all acting torques, and more complex dynamics should be considered. The translational velocity is taken into account in the model identified in [6–8]. Moreover, one can include all other state variables in the Eq. (9) so as not to restrict the possible identified model. Thus, instead of (9), the differential equation for the pitch angular velocity, considered separately, has the form: ω˙ y = p1 τpitch − p2 ωy − p3 θ − p4 vx ,

(13)

where vx is the translational velocity, keeping in mind that for simplicity ϕ = 0, ψ = 0 are assumed, i.e., vx is only determined by the pitch angle. Note that it is assumed that a positive value of the translational velocity corresponds to a positive value of the pitch angle. The dynamics of additional state variable vx is defined by the horizontal force, which is the product of the Earth gravity and the tilt angle tangent under condition of zero vertical velocity, and the drag approximation [3]: v˙ x = 9.81 · tanθ − p5 · |vx |vx ,

(14)

where p5 is the aerodynamic parameter. It is shown in [3] that this parameter is not a constant, and evaluation is made for operating points with pitch angle values of 5

342

V. Alexandrov et al.

Fig. 6. Comparison of simulated (nlgr) and experimental (data) values of control upitch and pitch angular velocity ωy for identified parameter of (9).

and 8 degrees. The estimate for the operating point corresponding to the identification experiment should be made here. Thus, the parameters p1..5 of the system (1), (8), (10), (11), (13), (14) should be identified. The function nlgreyest for the closed loop system with the same as above input and outputs data finds the parameters values: p1..5 = [87.44, 2.87, −2.98, 13.87, 0.21].

(15)

A good match of the simulated system and the experimental data can be seen in Fig. 7. This nonlinear model can be linearized, when tanθ ≈ θ is assumed for small angles. The transfer function for linearization of (14) in the operating point θ = 1◦ is: Pvx (s) =

13.73 . s + 0.267

(16)

Torque τpitch is related to control upitch through (11) multiplied by 4K τ 0 as a linearization of (10), where 0 ≈ 586 rad/s is an average rotor angular velocity for hovering. Then, the transfer function from upitch to the pitch angle θ in degrees for the operating point θ = 1◦ is: P3 (s) = e−0.0042s ·

3102 · (s + 0.267) .  + 1)(0.14s + 1) s2 − 4.02s + 26.52 (0.044s

(17)

The structures of the transfer functions (17) and (7) differ only in a small delay, revealed during the identification of the transfer function (11) for the rotor angular velocities. It should be noted that some parameters of the obtained transfer functions (17) and (7) differ significantly. Nevertheless, the identified transfer function structure of the tilt angle dynamics model is confirmed.

Identification of the Quadcopter Rotational Dynamics

343

Fig. 7. Comparison of simulated (nlgr) and experimental (data) values of control upitch and pitch angular velocity ωy for identified parameters of (13).

4 Conclusion The finite-frequency identification approach allows finding the transfer function under the assumption that the system is linear, or the control test signals, and measured output values vary in a small neighborhood of the desired operating point, so the system can be linearized. It is important that this does not require prior knowledge of the mathematical model of the plant. In the presented study, the identified structure is confirmed via consideration of the nonlinear model of the rotational dynamics of the quadcopter tilt angle based on the first principles. The study will be continued to refine the model parameters and obtain their values for various conditions. Acknowledgements. Research partially supported by RSF, project No. 23-29-00588, https://rscf. ru/en/project/23-29-00588/.

References 1. Alexandrov, V.A., Rezkov, I.G., Shatov, D.V.: Identification of the quadcopter vertical translation dynamics. In: 28th Mediterranean Conference on Control and Automation (MED), pp. 363–368. IEEE, Saint-Raphael, France (2020) 2. Alexandrov, V., Rezkov, I., Shatov, D.: Linearized model identification for quadcopter vertical translation dynamics. In: 25th International Conference on System Theory, Control and Computing (ICSTCC), pp. 278–283. IEEE, Iasi, Romania (2021) 3. Alexandrov, V., Rezkov, I., Shatov, D.: Finite-frequency identification of the quadcopter translation dynamics. In: 24th International Conference on System Theory, Control and Computing (ICSTCC), pp. 471–476. IEEE, Sinaia, Romania (2020)

344

V. Alexandrov et al.

4. Wei, W., Tischler, M.B., Schwartz, N., Cohen, K.: Frequency-domain system identification and simulation of a quadrotor controller. In: AIAA Modeling and Simulation Technologies Conference, AIAA, National Harbor, Maryland (2014) 5. Sakulthong, S., Tantrairatn, S., Saengphet, W.: Frequency response system identification and flight controller tuning for quadcopter UAV. In: Third International Conference on Engineering Science and Innovative Technology (ESIT), IEEE, North Bangkok, Thailand (2018) 6. Niermeyer, P., Raffler, T., Holzapfel, F.: Open-loop quadrotor flight dynamics identification in frequency domain via closed-loop flight testing. In: AIAA Guidance, Navigation, and Control Conference, AIAA, Kissimmee, Florida (2015) 7. Cho, S.H., Bhandari, S., Sanders, F.C., Tischler, M., Cheung, K.K.: System identification and controller optimization of coaxial quadrotor UAV in Hover. In: AIAA Scitech 2019 Forum, AIAA, San Diego, California (2019) 8. Cunningham, M.A., Hubbard, J.E., Jr.: Open-loop linear model identification of a multirotor vehicle with active feedback control. J. Aircraft 57(6), 1044–1061 (2020) 9. Alabsi, M.I., Fields, T.D.: Real-time closed-loop system identification of a quadcopter. J. Aircraft 56(1), 324–335 (2019) 10. Zhang, X., Li, X., Wang, K., Lu, Y.: A survey of modelling and identification of quadrotor robot. Abstract Appl. Anal. Article ID 320526 (2014) 11. Alexandrov, A.G., Orlov, Y.: Comparison of the two methods of identification under unknownbut-bounded disturbances. Autom. Remote Control 66(10), 1647–1665 (2005) 12. ArduPilot Altitude hold mode. https://ardupilot.org/copter/docs/altholdmode.html. Accessed 21 June 2023

Development of a Firmware for Multirotor UAV Flight Controller Implemented on MCU MDR 32 Daniyar Wolf , Vadim Alexandrov , Dmitrii Shatov(B) , Ilya Rezkov , Peter Trefilov , and Roman Meshcheryakov V.A. Trapeznikov Institute of Control Sciences of RAS, 65, Profsoyuznaya St., Moscow 117997, Russia [email protected]

Abstract. The paper is devoted to study unmanned aerial vehicle (UAV) flight operational features and UAV’s automatic control system operating in manual mode (it processes pilot’s control commands). The analysis of operation in special cases and conditions is carried out for the aircraft equipped with a radio-operating human-controlled system and automatic flight control system. The main goal of the research is to develop fully functional flight controller the flight controller based on the domestic made microcontroller unit (MCU) MDR 32. The main novelty is that board LDM-BB-K1986BE92QI with the MDR32F9Q2I microcontroller core is used as a hardware base for the flight controller. The development of hardware and software are described separately including some features that was used to overcome several limitations of the selected MCU. The base flight controller firmware modules, functions and algorithms are described. A sample quadrotor based on the S550 frame was assembled using the developed flight controller to carry out testing flights. Two test experiments are presented: the first one is a checking of the flight controller base functionality (parameters configuration and calibration) and the second one is a test flight performed by a pilot. Keywords: Aerial Vehicle · Quadrotor · Flight Controller · Euler’s Angles · Attitude Control · Manual Control Mode

1 Introduction Developing of manned/unmanned vehicles is accompanied by wide using of automation tools, the main of which is a flight controller that is a software and hardware complex implementing the vehicle control system. Known flight controllers use different hardware and differ also in firmware parts, which implement corresponding software logic and have various application areas. There are dozens of onboard firmware for UAV’s. All of them have their own pros and cons depending on the specific flight task, so it is impossible to select “the best flight controller firmware”. The existing firmware can be divided into three groups: the first ones are intended to acrobatic flights (drones races), the second group is for © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 345–356, 2023. https://doi.org/10.1007/978-3-031-43111-1_31

346

D. Wolf et al.

freestyle flights (manual control) and the rest ones are for autonomous flights using GPS navigation (automatic control). The following firmware can be considered as the most popular for autonomous flights: 1. Ardupilot [1] – perhaps this is the most popular open-source firmware for autonomous navigation with GPS. The autopilot has been actively developed for many years and is now considered as one the most reliable solutions for both flight enthusiasts and professionals when it comes to automatic flights. 2. Pixhawk [2] – this is professional open-source autopilot developed by high qualified engineers from industry together with scientists. The firmware is supported by active community and provides control of all kinds of vehicles (ground, aerial and undewater) in several mentioned modes (race, freestyle and automatic). The corresponding flight controller is part of the PX4 ecosystem that contains open hardware standard (pixhawk), mavlink and uavcan/cyphal communication protocols, software for mission planning (groundcontrol) and other useful issues. 3. iNav [3] – this autopilot is intended to automatic navigation and autonomous tasks like flight through given points and return to home (RTH). The firmware is branch from the project Cleanflight, so is Betaflight, that is why they have a lot of similar functions, even user interface of configurator looks similar. However, besides multirotors iNav also supports fixed-wing aircrafts and radio-controlled cars. The iNav project uses all the GPS capabilities for mission planning (waypoints), it is constantly evolving and improving. Yet its development lags the latest flight characteristics achieved by Betaflight, when it comes to freestyle and racing drones. iNav supports several, but not all Betaflight flight controllers. In general, a flight controller consists of a processor for computing (MCU), a gyroscope and an accelerometer installed on a printed circuit board. Processors in flight controllers are called microcontrollers. They store and execute autopilot firmware. Most modern firmware is implemented on 32-bit microcontrollers of STM32x family, due to their extensive computing capabilities and a lot of useful free software. Companies, that develop flight controllers, use actively such microcontrollers because they have significant computing resources and a numerous set of functions required to effective control of UAV. In present, there are five main series of STM32 processors: F1, F3, F4, F7 and H7, which differ by size of memory and computing power. The Atmel controller series such as ATMEGA2560 and ATMEGA32U-2 are also used by minor part of developers for implementation of flight controllers firmware. However, software support of these controllers is limited and underdeveloped in comparison to the STM32x series. This means that flight controllers developers on Atmel microcontrollers face limitations in available functions and tools which complicates the development of complex and advanced unmanned systems. As a result, STM32 controllers, supported by extended functionality and turnkey solutions in forms of libraries and frameworks, are chosen preferably by developers for their flight controllers projects. For software developers this allows to focus on implementation of applied functions instead of low-level programming of standard functions.

Development of a Firmware for Multirotor UAV Flight Controller

347

Necessity of flight controller onboard firmware development on base of a domestic made MCU is caused by political and economic situation. Current conditions in the world motivate researchers groups to create their own scientific and engineering projects in this area. Without a doubt, such separate technological base allows to strengthen strategic independence. An autopilot firmware development is bounded by the technical capabilities of the microelectronic base, so such a high-tech software development at the level of aerial vehicle control systems will allow to build feedback with manufactures of domestic microelectronics. Novelty of the paper results on one side lies in the use of the domestic MCU as a hardware base for the flight controller. Such choice requires development a lot of software, since the known autopilots are not compatible with this selected microcontroller. Thus, the other part of novel results consists of engineering and programming features developed and implemented in the flight controller firmware. The paper goal is to present development of the whole quadrotor construction and hardware-software complex “aircraft – onboard computing system – autopilot – operator – environment – disturbances”. This complex allows to carry out advanced and complementary research on design of future and existing UAV’s and their automatic control systems with adjustment to domestic microprocessor technology.

2 Problem Statement The practical application of the flight automatic control methods is impossible without information about the model of aircraft characteristics (in general, non-stationary) and model of the pilot’s actions (commands). In this relation, it is necessary to use theoretical and practical approaches to determine mathematical models accounting aerodynamic characteristics and to estimate their affect on the flight dynamics. The general considered problem is to develop cyberphysical system operating in conditions of an aerial environment. In particular, an aerodynamic system of quadrotor UAV is considered. The certain problem is to design automatic attitude control system that stabilizes UAV at Euler’s angles – roll, pitch, and yaw.

3 Model of the Flight Controller At first, the flight controller development was reduced to modeling the pilot behavior at the aircraft control problem in several certain situations. Such a model was designed for 4 control channels: throttle, roll, pitch, and yaw. The automatic control system of aerial vehicle should simultaneously reject disturbances caused by several sources: asymmetry of structural elements of the aircraft, nonideal work of the propulsion system and its motors, effect of wind action and so on. Design of such control system is the main problem of aerodynamic system control. The considered problem is solved using classical feedback approach with PID control law, that allows to obtain desired time performance (system reaction time on disturbances and pilot commands). This quality index is crucially important for aircrafts, having very fast dynamics and usually unstable in open loop state, and should be selected at the stage of initial design research.

348

D. Wolf et al.

In the present time, the several approaches are widely used to model pilot’s control actions: structural, optimal, neural network and fuzzy-logic based [4–6]. Design of universal model fit for flight study of different situations with further formalization implies great difficulties and requires additional research, so at the stage of initial R&D, the simple approximation base model was used. To create a specific model for the rather special situations, a structural model of pilot’s actions [7] was chosen. The structure of the pilot’s actions model significantly depends on the character of the special situation caused by a disturbing factor. From the considered special situations (for roll, pitch and yaw), the following pilot actions were selected: • a time delay with control intervention (a pilot intervenes in the control by tilting the joystick sticks intended to control the axes X for roll, Y for pitch and Z for yaw after the time t spent since the situational factor occurs). • rejection of disturbances in angular velocities (ωx , ωy , ωz ) and angles (roll ϕ, pitch θ and yaw γ ). After tilting the corresponding control stick the pilot holds it in fixed position until the value the angle (or angular velocity) returns to the initial one. • stabilization of motion parameters (corrective tilting of the control stick is intended to stabilize zero (or desired) values of the roll, pitch and yaw angles and the corresponding angular velocities). The dynamics of the roll ϕ and pitch θ angles are practically the same, so further the control of the pitch angle is only considered. The control loop for the roll angle for a practically symmetrical quadrotor will have the same structure and controller parameters as the pitch one. According to the Newton-Euler’s formalism the rotational dynamics on pitch angle is described as follows [8]:  1 x ω˙ y = ωx ωz Iz −I Iy + Iy τpitch , (1) θ˙ = ωy cos ϕ − ωz sin ϕ, where ωx , ωy , ωz are the angular rotational velocities relative the quadrotor mass center, Ix , Iy , Iz are the moments of inertia, τpitch is the torque relative to the Y axis. It is assumed that in (1) the angular velocities ωx , ωy and the angle ϕ are close to zero. Then, the reduced dynamics, that relates torque τpitch with control signal upitch , is determined by the first order time lag [9, 10], so the linearized transfer function from control upitch to the corresponding angle θ is the following: P(s) =

Kθ , s2 (T s + 1)

(2)

where s is the Laplace variable, Kθ is the proportional gain, T is the time constant. One can note that the value of T in (2) is rather small: its value, obtained from the experiments, is estimated as less than 0.1 s. However, this lag cannot be discarded, since it is required to design the fast closed loop system to provide stability of the system with two integrators, in the that case the value T cannot be considered as small and omitted. The control loop for the pitch angle is designed in the well-known form of a backstepping system, that consists of two loops: the inner on the angular velocity and the

Development of a Firmware for Multirotor UAV Flight Controller

349

external on the pitch angle. The block diagram of the control loop is presented in Fig. 1. The inner loop implements a PID controller of the angular velocity that generates the control u according to the following equation: u = C(s)eωy , C(s) =

Kp + Ksi + Kd s , Tf s + 1

where C(s) is the PID controller transfer function with coefficients Kp , Ki , Kd for the corresponding proportional, integral and differential parts, Tf is the filter time constant, eωy = ωy ∗ − ωy is the tracking error, where the desired setpoint value ωy ∗ is generated by an output of the proportional (P) controller on the pitch angle:   ωy ∗ = Kp θ eθ = Kp θ θ∗ − θ , where θ ∗ is the setpoint value of the pitch angle and θ is its current value, Kp θ is the coefficient of the P controller. The diagram also shows PWM’s that produce signals for motors, P(s) that simplified denotes the real control plant and gyroscope that measures the angular velocities.

Fig. 1. The block diagram of the pitch angle control system.

In the case θ ∗ = 0 (if pilot does not set the desired pitch angle), the control system operates in a stabilization mode that rejects the external disturbances and parametric uncertainties which could disturb the horizontal attitude of the quadrotor. The control system for the roll angle ϕ is designed in a similar way. The yaw angle γ control operates according to a simplified variant only with a PI controller for the corresponding angular velocity ωz (without design of a backstepping control system), which is due to the control plant slower dynamics for this Euler’s angles component.

4 Design of the Flight Controller Hardware It follows from the mathematical model of the flight controller that UAV is a complex of complicated interconnected systems, which are hard to describe and implement during real design process even in simplified approximate degree. The UAV operating performance is mainly determined by the separation structure of control functions, information transmission from an operator to computing system of the flight controller and an instant processing of the inner and external disturbances effects.

350

D. Wolf et al.

A pilot located onboard of an aircraft can solve the disturbance rejection problem using his own acceleration senses, that is described with previously introduced terms «the reaction time» and «the special situation». The first one is determined as the elapsed time from special situation initialization till the beginning of the pilot actions intended to reject disturbing factor called the special situation. It is known that the average reaction time on such situations varies in range 100–300 ms. If the pilot control the aircraft remotely, he does not receive any acceleration senses and cannot react on special situations, so the main problem of design is reduced to replace the pilot actions (with the same or higher reaction time) with an automatic system based on a cyberphysical mathematical model, which is then implemented in available domestic made MCU providing modern technology level. The functional diagram of the designed model and corresponding control system is shown in Fig. 2, where r(t) denotes the setpoint signal of the considered loop, y(t) is respectively the measured output and control signal u(t) that is fed through PWM Generator to electric speed unites (ESC) and then to motors, ax , ay , az are the projections of the accelerometer output. The MCU implements system, that allows to replace the pilot and stabilize the quadrotor horizontal attitude [11, 12]. The angular velocities for pitch and roll are separately controlled by two PID controllers (Fig. 2 presents a generic structure of the quadrotor hardware and its operation scheme), which generate control signals that are added by mixer to the whole motors control signal (it includes throttle, pitch, roll and yaw controls). The feedback uses the measured signal y(t) which is generated by a sensor. Modern equipment for gyroscopes and accelerometers operates with high rate of information transmission, for instance, sensor MPU6050 provides data with rate 400 kHz, so the MCU should processes data with the same rate. Another two important characteristics of inertial measurements unit (IMU) are the maximum sampling frequency and signal-to-noise ratio (SNR) regardless of mechanical vibrations and electromagnetic interference. The wide usage among aircraft modelists has got microchip MPU6000 that supports sampling frequency up to 8 kHz and provides good noise immunity. To a lesser extent, the IMU’s are used MPU6500, MPU9250 and ICM20689 providing higher transmitting rates at the cost of worse SNR. The abroad autopilots usually operate using reaction time in range 2–5 ms. Resuming, the minimum clock rate of the MCU processor should be selected not lower than 50 MHz. Another requirement on quadrotor MCU is the present of PWM channels. Both described conditions hold for microchip of MDR32F9Q21 series that provides three timers with 4 channels supporting PWM operating mode. However, in real application it occurs, that only the third timer can be used for this purpose. A hardware part of the flight controller is shown in Fig. 2. The actuators consist of motors with the ESC’s. UAV control signals from a pilot are received via a radio channel by a communication device (receiver), then fed to the microcontroller (MC) inputs by a bus (the information channel can be any – PWM, PPM, IBUS, SBUS, etc.). After that the MC transforms them either directly into pulse-width modulation (PWM) signals fed to ESC 1–4 which set the corresponding rotation speed of the electric motors M1–4 or uses in control logic as setpoints for the angles control loops.

Development of a Firmware for Multirotor UAV Flight Controller

351

Fig. 2. The flight controller connections scheme including MCU, ESC’s, motors, sensors and a receiver.

An experimental sample of the flight controller is a microcontroller device consisting of a circuit board with electronic elements placed on it according to Fig. 2. The microcontroller itself is placed on the debug board LDM-BB-K1986BE92QI manufactured by “PKK Milandr”. The core of the MDR32F9Q2I microcontroller is built using ARM Cortex-M3 technology with the LQFP-64 form factor and 128 KB FLASH memory and 32 KB SRAM memory. Operating clock rate is 80 MHz (the autopilot operates on the frequency 72 MHz). Operating temperature range of the MK is −40…+85 °C. Also, a module for a 3-axis gyroscope and accelerometer GY-521 MPU-6050 is placed on the board (see Fig. 3). The board and pins protection is provided by a 3D printed bath of the appropriate sizes.

352

D. Wolf et al.

Fig. 3. Appearance of the experimental sample of the flight controller based on a domestic made MCU: A) in a protective cover; B) on the UAV board.

5 Software Implementation of the Flight Controller The flight controller firmware architecture consists of three main systems: propulsion control, sensor data processing and automatic attitude control. The propulsion control system contains functions performing operations with motors. The sensors system collects data about angular velocities and accelerations from gyroscope and accelerometer, processes it to calculate the Euler’s angles, stores all data in logs. The attitude control system consists of the functions, that implement pitch, roll and yaw control loops, and performs stabilization in a given by a pilot attitude. In turn, each system consists of subsystems (modules): 1. 2. 3. 4. 5. 6.

information collection modules (Sensors); attitude calculation module (with IMU); automatic attitude control modules: PID, Mixer, Motors controller; external control commands receiving module: IBUS, SBUS, etc.; low-level drivers that provide communication exchange (UART, I2C); the core of the system with a resource manager.

Structure and interconnections of the flight controller software modules and blocks are presented in Fig. 4. The software implementation of the flight controller allows to select the reaction time in the range 2–4 ms, which corresponds to the similar characteristics of the well-known firmware such as Ardupilot and Pixhawk. The subsystem for planning input and output data processing is implemented in the special class, called task scheduler. The subsystems of the UART and I2C drivers level implement the necessary communication between the receiver, the MPU6050 sensor with the MPU and IBUS subsystems. The IMU subsystem is a module that collects data from the MPU6050 sensor using the I2C protocol. This module solves the following tasks: 1) using the I2C driver module, it accesses the memory registers containing the sensor measurments; 2) it converts the received data to the required measure units (m/s2 for an accelerometer, rad/s for a gyroscope); 3) calculates the Euler’s angles values (in degrees). Sensor measurements are read at a frequency of 250 to 500 Hz, depending on the selected operating mode of the system core. The separate control loops are implemented with the PID control in backstepping structure for angles pitch and roll, and with the

Development of a Firmware for Multirotor UAV Flight Controller

353

Fig. 4. Software architecture of the developed flight controller.

simple PI control for yaw angular velocity. The setpoints in these loops are the desired angles values for pitch and roll and the angular velocity for yaw. The information transmission rate between the Motors controller module and ESC’s meets the condition of dShot600 standard. The software that implements the autopilot functions is developed in C++ using object-oriented programming paradigm for the above architecture of the flight controller device. The program satisfies the requirements of the RTOS level system for domestic made microcontrollers and has the following characteristics [13–15]: • the ensured response time to the inner and external situations (noise from the equipment, the peripheral device, the data processing subsystems and etc.); • the presence of a subsystem for planning the input information processing and forming the control signals; • increased requirement for reaction time to external and internal events (receiving external measurements and control signals time, the input signals processing time and time for generating control signals); • the presence of a special subsystem for generating motors control signals, considering the hardware capabilities of the used MCU. At this stage of development, there is no user interface to confige the flight controller settings.

6 Experimental Testing The main quadrotor load-bearing frame was assembled from the S550 frame. The frame is made up of four beams connected cruciformly. Electric motors of the 2212 series, providing 920 rpm, were chosen as propulsors. Propellers with 9.4 blade in 5 increments have a total thrust of about 4 kg with 4S Li-Po battery. This load capacity is enough to provide power to drive the assembled test platform weighing 1560 g, including the battery. To control the motors, ESC’s with a power of up to 30 A and the SimonK firmware preinstalled were used. To receive signals and control the quadrotor, the FSiA6B receiver was chosen, which works with the control panel using the IBUS protocol.

354

D. Wolf et al.

The FLYSKY FS-i6 remote control joystick was used to control the quadrotor. Thus, this test setup provides an opportunity to carry out experiments with the developed flight controller using quadrotor that is made of available and commonly used components that can be purchased on open markets (see Fig. 5).

Fig. 5. The experimental quadrotor setup: A) Appearance UAV on the ground; B) UAV in flight controlled by the developed flight controller.

The flight controller was tested for automatic flight stabilization in manual control mode, the experiments were performed in the following sequence: 1. Configuration of the flight controller parameters and calibration actions: tuning the PID controller coefficients and determining the bounds of stabilization control, calibration of the accelerometer, gyroscope and ESC’s. 2. Testing flight in manual control mode: the angles setpoints for the desired attitude are set by a pilot using joystick, then the stabilization control quality (in terms of control errors and reaction time) is estimated using logs data. Tests from the first category allows to prepare the quadrotor for takeoff. Without their implementation, it is impossible to proceed to flights, since the quadrotor is unstable in the open loop state and overturns without stabilization in the horizontal attitude (on the pitch and roll angles). The PID controllers parameters were found using a combination of the heuristic and analytical approaches: the controllers coefficients were designed using the Ziegler–Nichols method, the range bounds on the permissible automatic control values were determined empirically based on the results of the intermediate tests, the frequency of control discreteness was determined as the maximum available, taking into account the flight controller hardware capabilities. Calibration of sensors and ERS is performed automatically when the flight controller is initialized. The second experiments category is the manually controlled flight: the pilot uses a joystick to control the throttle control and Euler angles channels, determining the desired flight trajectory. Test flights carried out at the ICS RAS testing area showed good controllability of the UAV in this mode on all channels. Figure 6 shows the experimental data plot during the flight: the change in pitch angle θ in blue, the desired pitch angle value of the pitch angle θ ∗ in red, and the corresponding values of the angular and desired angular velocities ωy in green and ωy ∗ in blue color respectively.

Development of a Firmware for Multirotor UAV Flight Controller

355

Fig. 6. The pitch angle and the corresponding angular velocity, and the corresponding desired values of these signals during the test flight.

7 Conclusion The paper presents results of the flight controller development basing in domestic made MCU MDR32 series, which is a new engineering result. A fully functionalautopilot firmware has been developed that implements the standard function of free flight under the control of a pilot, which operates similarly to well-known common flight controllers firmware. The article describes new hardware and software solutions that have been developed and implemented to take into account the specifics of the selected processor module. The experimental study of the flight controller shows that it provides good attitude automatic stabilization and therefore comfort takeoff, flighting, and landing characteristics in manual control mode. The testing setup was designed from the commonly used components, and it can be considered as a standard one, so the developed flight controller might be used in practice as a firmware for wide range of UAV. Currently, the functionality of the developed hardware and software is being expanded. The hardware improvement of the flight controller is aimed to increasing the degree of UAV’s autonomy by means of placing on the circuit board and integration in the flight controller a barometer, magnetometer, and GPS sensors, lidar, radar, etc. The final purpose of this research is to create a fully functional quadrotor flight controller intended for all kinds of applications, including usage in advanced group control systems [16].

References 1. Ardupilot Homepage. https://ardupilot.org. Accessed 23 May 2023 2. Pixhawk Homepage. https://docs.px4.io. Accessed 23 May 2023 3. iNav Homepage. https://github.com/iNavFlight/inav. Accessed 23 May 2023

356

D. Wolf et al.

4. van der El, K., Pool, D.M., van Paassen, M.M., Mulder, M.: Effects of preview on human control behavior in tracking tasks with various controlled elements. IEEE Trans. Cybern. 48(4), 1242–1252 (2018) 5. Efremov, A.V., Tyaglik, M.S., Tyaglik, A.S., Irgaleev, I.K.: Developing the mathematical model of a pilot in a control manual preview tracking task. Russ. Aeronaut. 62, 394–400 (2019) 6. Verezschikov, D.V., Voloschin, V.A., Ivaschkov, S.S., Vasiliev, D.V.: Application of fuzzy logic to create a simulation model of the pilot’s control actions. In: MAI Proceedings (99). (in Russian). http://trudymai.ru/published.php?ID=91926. Accessed 23 May 2023 7. Natalijn, V.M.: Modelling of the Pilot Control Actions in the Special Situations. Sci. Bull. Moscow State Tech. Univ. Civ. Aviat. 138, 205–209 (2009). (in Russian) 8. Alexandrov, V.A., Rezkov, I.G., Shatov, D.V. Identification of the quadcopter vertical translation dynamics. In: Proceedings of the 28th Mediterranean Conference on Control and Automation, pp. 363–368. IEEE, Saint-Raphaël (2020) 9. Alexandrov, V.A., Rezkov, I.G., Shatov, D.V.: Linearized model identification for quadcopter vertical translation dynamics. In: Proceedings of the 25th International Conference on System Theory, Control and Computing, pp. 278–283. IEEE, Ias, i (2021) 10. Mikrin, E.A., Zubov, N.E., Lapin, A.V., Ryabchenko, V.N.: Analytical formula of calculating a controller for linear SIMO-system. Differ. Eqn. Control Process. 1, 1–11 (2020). (in Russian) 11. Wolf, D.A., Alexandrov, V.A., Rezkov, I.G.: Automation of the UAVs pilot’s behavior using a native microcontroller. Industr. Autom. Control Syst. Controllers 3, 9–16 (2023). (in Russian) 12. Lipovii, D.A., Maltsev, A.C.: Development of a flight controller architecture for a quadrocopter based on a single-board computer raspberry pi. Bull. Novosibirsk State Univ. Ser.: Inf. Technol. 18(3), 19–33 (2020). (in Russian) 13. Wolf, D.A.: Software implementation of group control of collector motors. In: Proceedings of the 33rd International Scientific and Technical Conference “Extreme Robotics”, pp. 206–212. RTC, St. Petersburg (2022). (In Russian) 14. Mamchenko, M.V., Romanova, M.A. Trefilov, P.M.: Algorithm for sensor data merging using analytical module for priority sensor selection. In: Proceedings of the 2020 International Conference on Industrial Engineering, Applications and Manufacturing. IEEE, Sochi (2020). https://ieeexplore.ieee.org/document/9111978 15. Mamchenko, M.V., Romanova, M.A. Trefilov, P.M.: An algorithm for evaluating the measured values of dynamic objects under the influence of external factors. In: Proceedings of the 2020 International Russian Automation Conference, pp. 1069–1073. IEEE: Sochi (2020) 16. Kutakhov, V.P., Meshcheryakov, R.V.: Group control of unmanned aerial vehicles: a generalized problem statement of applying artificial intelligence technologies. Control Sci. 1, 55–60 (2022)

Autonomous Landing Algorithm for UAV on a Mobile Robotic Platform with a Fractal Marker Dmitry Anikin, Artem Ryabinov(B)

, Anton Saveliev , and Alexander Semenov

St. Petersburg Federal Research Center of the Russian Academy of Sciences (SPC RAS), 39, 14th Line, St. Petersburg 199178, Russia [email protected]

Abstract. This article describes experiments related to the simulation of automatic landing of UAV on a mobile robotic platform using computer vision and a control system based on PID and polynomial regulators within the Gazebo environment. Algorithm has been developed to generate control inputs for maintaining the velocities of the UAV based on the data from the computer vision system and feedback from onboard sensor devices. A series of experiments were conducted at altitudes ranging from 5 to 20 m, which allowed for identifying limitations that affect the successful landing of the UAV. Measurements were taken of the landing time and landing error, calculated as the distance between the platform center and the UAV’s center of mass upon completion of the landing. The average landing time and error ranged from 19.64 s and 0.16 m at an initial altitude of 5 m to 121.01 s and 0.27 m at an initial altitude of 20 m. Analysis of the obtained results revealed that both the average error and landing time increase with the initial altitude, which is associated with the accuracy of marker recognition at altitudes above 15 m. The obtained results can be valuable for further improvement of systems for automatic landing of UAVs on mobile platforms. Keywords: Unmanned Aerial Vehicle · Quadcopter · Marker Detection · Automated Landing · Computer Vision · Navigation

1 Introduction In the modern world, autonomous control systems and unmanned aerial vehicles (UAVs) and specifically quadcopters are gaining increasing popularity and are being used in various fields of activity, including industry [1], transportation [2], and the agricultural sector. The applications of quadcopters include monitoring and surveillance of territories, such as detecting forest fires, environmental pollution zones, and more. One of the most challenging tasks faced by developers of multirotor UAVs is the creation of a reliable automatic landing system to a predefined point. Most modern commercial and open-source autopilot systems allow for landing based on GPS. However, in certain situations, such as landing in urban areas or near tall buildings, the GPS signal may be unavailable or inaccurate, leading to landing errors. Furthermore, often, the © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 357–368, 2023. https://doi.org/10.1007/978-3-031-43111-1_32

358

D. Anikin et al.

landing needs to be performed not on a stationary target but on a moving object (e.g., another robotic platform). In light of these challenges, the use of computer vision technologies for landing purposes becomes particularly promising. The goal of this research is to develop and test algorithms that generate control inputs for UAVs based on computer vision for the automatic landing of UAVs onto a predefined ground robotic platform (GRP) at various flight altitudes. To achieve this goal, experiments were conducted to simulate the automated landing of a quadcopter model [3], equipped with an autopilot system and a video camera, onto a GRP equipped with a fractal ArUco marker [4]. Corresponding high-level autopilot algorithms were developed and tested, integrated into low-level software representing the Pixhawk PX4 autopilot [5]. The simulation was conducted using the Gazebo environment [6], with initial heights of 5, 10, 15, and 20 m. The research includes quantitative measurements of spatial errors and time associated with the landing process. The results of the performance analysis of the system under various conditions can be valuable for the developers of control and landing systems for UAVs in various practical applications [7, 8]. The structure of this article is organized as follows: Sect. 1 provides a literature review, including an analysis of related research studies. Section 2 describes the methods, algorithms, and environments used for conducting the experiments. Section 3 presents the description and results of the experiments. Section 4 analyzes the obtained experimental data and draws conclusions. Finally, Sect. 5 provides an overall conclusion, and suggests avenues for further development of the study.

2 Analysis of Existing Research The authors of the study [9] propose an approach based on computer vision that enables an unmanned aerial vehicle (UAV) to detect, track, and land on a mobile ground robotic platform using ArUcO markers. The authors did not provide the landing time results and spatial error during landing. The UAV reached a maximum height of 4 m. The authors of the study [10] present a fully autonomous vision-based system that addresses limitations posed by wind disturbances by tightly integrating localization, planning, and control. The approach proposed in this work employs an extended Kalman filter with simulated GPS measurements for platform position, orientation, and velocity estimation at a large distance between the platform and the UAV. In close proximity to the platform, fiducial AprilTags are utilized for estimation. The flight height was set at 0.5 m, and the landing duration was 5.7 s for a static platform and 11.5 s for a moving platform. The authors did not specify the spatial error of the landing. The study [11] presents a method for autonomous landing that can be implemented on a micro-UAV. The approach requires broadband feedback control loops to ensure safe landing in the presence of various uncertainties and wind disturbances. The authors demonstrate the system architecture, including dynamic modeling of the UAV, implementation of a Kalman filter for optimal mobile platform localization, and the development of model predictive control for UAV guidance. Simulations were conducted with and without wind conditions. The flight height was set at 3 m, and the wind speed was 5 m/s. In the simulation without wind, the spatial error was 0.21 m, and the landing time

Autonomous Landing Algorithm for UAV

359

was 44.25 s. In the simulation with wind, the error was 0.37 m, and the landing time was 105.24 s. The authors of the study [12] propose an autonomous landing system using a composite landmark. In this system, a circular landmark with notches and a two-dimensional landmark are combined to provide visual localization over a wide range. Real-world experiments were also performed in an open area. The maximum flight height was 5 m. The proposed method achieved continuous localization results within a range of 0.5 m to 6 m, with a localization error of 0.1–0.6 m. The maximum average positional error was 0.32 m, with an 80% probability of successful landing. The landing time was 19.5 s. In the study [13], a custom visual serving controller is employed to guide the UAV, generating control commands for linear velocities based on relative position data obtained by processing frames from an onboard camera mounted on a stabilizer. The authors do not provide information about the flight height, consider only the reflection of sunlight as an external disturbance, and do not specify the error. The study mentions a landing speed of 6.5 s after detecting the marker. In the simulation, out of 19 landing attempts, 17 were successful, and in real-world conditions, out of 22 attempts, 19 were successful. Unsuccessful attempts were attributed to sunlight reflection, which hindered marker recognition. The authors of the paper [14] propose a system that enables tracking and landing of a UAV on a ground vehicle. The system utilizes vision-based input without external localization, where the UAV observes the motion of the UGV, predicts its movement, and generates a smooth approach trajectory to the predicted position. Experiments were conducted in the Gazebo simulation environment using the Rotors simulator. The flight altitude was set at 3 m, and the landing time took 13.5 s for linear motion and 13 s for a curved trajectory. The authors do not specify the error. In the paper [15], a solution for high-precision landing based on marker usage is presented. The flight altitude was set at 20 m, with an average error of 11 cm, and the landing time took 162 s with a wind speed of 2.7 m/s. The authors of the paper [16] propose a developed system for autonomous landing of a multi-rotor aerial vehicle (MAV) on a moving platform based on a vision-based tracking device. The system combines GPS positions of the MAV and the platform for MAV navigation when the marker is outside the camera’s detection range or in the camera’s blind spot. The flight altitude of the MAV was set at 4 m, and the landing time took 35 s with an error of 30 cm. After analyzing the proposed solutions, it was concluded that many papers do not specify the spatial landing error. Additionally, almost all works focus on landing at low altitudes (up to 6 m). Therefore, this study will present a comprehensive testing of the developed algorithms, considering both the landing time and the spatial landing error using statistical analysis methods. This will allow conclusions to be drawn not only regarding the performance but also the accuracy of the developed solutions.

360

D. Anikin et al.

3 Materials and Methods 3.1 Used Models The experiments were conducted in the Gazebo simulation environment. The modified model of the Clearpath Husky UGV [17] was used as the Unmanned Ground Vehicle (UGV), on which a fractal ArUco marker with a size of 1 × 1 m and five levels of nesting was placed (see Fig. 1a, b).

Fig. 1. a) Fractal ArUco marker, b) Modified model of the Clearpath Husky UGV.

This feature allows for efficient marker detection at various distances compared to a standard marker. The platform moves at a constant speed of 1 m/s along a fixed curved trajectory. The 3DR Iris quadcopter model [3], provided by the PX4 SITL simulation is used in this work. Additionally, a lower RGB camera with a resolution of 640 × 480 and a frame rate of 30 frames/second is mounted on the model. The technical specifications of the UAV are: • Mass: 1.535 kg; • Moments of inertia around the X, Y, and Z axes: 0.019125 kg m2 , 0.29125 kg m2 , 0.055255 kg m2 , respectively; • Maximum rotor speed: 1100 RPM; • Dimensions: 47 cm × 47 cm × 11 cm. 3.2 The Landing Control Algorithm The landing control algorithm involves guiding the UAV to the center of the platform by setting linear velocities along the X, Y, and Z axes using the landing controller, which communicates with the autopilot (refer to Fig. 2). The landing controller consists of two adaptive height-proportional-integralderivative (PID) controllers that generate linear velocities along the X and Y axes, and a logarithmic polynomial (LP) controller that generates linear velocities along the Z axis based on the relative coordinates r = [x, y, z]T . The vector of output linear velocities Vsp is passed to the input of the cascaded autopilot controller system. The velocity PID controller generates the necessary accelerations asp to achieve the desired velocities. The accelerations are transformed into a quaternion representing the UAV’s position in the air, denoted as qsp , , which is then passed to the proportional (P) controller for attitude control. The output of the attitude PID controller is a vector of required

Autonomous Landing Algorithm for UAV

361

Fig. 2. UAV Controller System.

angular velocities sp . The angular velocities are used to generate virtual commands, which are normalized values of the throttle δRsp , aileron δAsp , and elevator δE sp . These commands are processed by a mixer, which generates the propeller torques [18]. The proportional-integral-derivative (PID) controller is described by the following expression [19]:  t de(t) , (1) u(t) = Kp e(t) + Ki e(t)dt + Kd dt 0 where: e(t) is the error signal; K p is the proportional gain, K i is the integral gain, K d is the derivative gain, t is time. To discretize the continuous PID controller, the integral term is approximated using the trapezoidal rule, and the derivative term is approximated using finite differences:



ei−1 − ei de(t) ≈ , dt t t

e(t)dt =

0

(2)

1 n (ei−1 + ei )t, i=1 2

(3)

where: Δt is the discretization step, i is the current step number, n is the last step number. To reduce derivative noise and smooth the control signal, a low-pass filter with an infinite impulse response is applied to the derivative part: Ei = (1 − r)Ei−1 + r

ei−1 − ei , t

(4)

where: E i is the filtered derivative component, r is the smoothing coefficient. The tuning of the PID controller was performed empirically by observing the response of the UAV to individual control signals. As a result, the following coefficients were determined (see Table 1). Table 1. Coefficients of the tuned adaptive height PID controller. Height, meters

Kp

Ki

Kd

20

0.305

0.0325

0.0025

5

0.5

0.0705

0.0005

362

D. Anikin et al.

For adaptive self-regulation of height based on the obtained coefficients, an approximation was constructed using curves. Exponential, logarithmic, and linear curves were tested, and the exponential approximation showed the best results. Kp (z) = exp(−0.098z − 0.2),

(5)

Ki (z) = exp(−0.154z − 1.878),

(6)

Kd (z) = exp(0.321z − 9.21),

(7)

where z is the current altitude of the UAV in meters. The final control law takes the following form: Vx,y = Kp (z)ei + Ki (z) ·

1 n (ei−1 + ei )t + Kd (z)Ei , i=1 2

(8)

where ei is value of the error signal; z is current altitude of the UAV in meters; E i is value of the filtered derivative component of the controller. The generation of linear velocity along the Z-axis is based on the logarithm of a polynomial function of normalized distance in the vertical plane [20].   (9) Vz = Vz0 logn 1 + ax + bx2 , where x = RRz0z ; V z0 is initial descent velocity; n is logarithm base; a, b is coefficients determining the velocity profile; Rz0 is initial distance along the Z-axis; Rz is current distance along the Z-axis. According to this law, the approach speed logarithmically decreases to zero as the distance to the landing site approaches zero (see Fig. 3). The maximum descent speed is calculated using the following formula:   a2 , (10) Vzmax = Vz0 · logn 1 − 4b where V z0 is initial descent velocity; n is logarithm base; a, b are coefficients determining the velocity profile. The landing algorithm is schematically represented as follows (see Fig. 4). After the marker enters the camera’s field of view, the PID controllers generate control signals for following the visual target. The altitude of the UAV remains constant during this movement (the descent rate is zero). When the UAV reaches a distance of 45 cm from the center of the platform, the descent process begins. The process ends when the descent altitude becomes less than 10 cm, at which point the motors are turned off, and the copter lands completely. In case the platform goes out of the camera’s field of view, after three seconds, the UAV ascends at a rate of 5 m/s in order to reacquire the platform. If the platform is not detected again within two seconds, an emergency return to the initial GPS position is initiated to perform a new maneuver.

Autonomous Landing Algorithm for UAV

Fig. 3. Descent velocity dependence on altitude.

Fig. 4. Block diagram of the landing algorithm.

363

364

D. Anikin et al.

4 Experiments Testing of landing was conducted at various heights: 5 m, 10 m, 15 m, and 20 m, with a total of 40 attempts. During each attempt, the time τ required for landing was measured, as well as the error ε, which is calculated as the distance between the center of the platform and the center of mass of the UAV. The graphical representation of the obtained results is presented in Fig. 5 and Fig. 6.

Fig. 5. Violin plots of landing time as a function of height with an approximation line.

Fig. 6. Violin plots of landing error as a function of height with an approximation line.

The provided graphical representations allow us to observe an increase in both landing time and error with an increase in initial height. For each height, the mean value and standard deviation of the landing time and error were calculated. Additionally, for

Autonomous Landing Algorithm for UAV

365

each height, the increments in landing time and error, τ and ε, were calculated as the difference between the corresponding mean values at the current and previous heights. The results of the experiments are presented in Table 2. Table 2. Coefficients of the tuned adaptive height PID controller. Height, m

Success rate

τ, sec

ε, m

τ, sec

ε, m

5

100%

19.64 ± 2.108

0.16 ± 0.078

0

0

10

100%

24.08 ± 2.019

0.22 ± 0.047

4.44

0.06

15

100%

48.51 ± 4.526

0.25 ± 0.066

24.43

0.03

20

82.50%

121.01 ± 14.188

0.27 ± 0.074

72.5

0.02

It can be observed that the height of 15 m is the highest at which stable landing on the platform was achieved in 100% of cases. As the initial height increases, the time required for landing significantly increases. The error also increases, but not to the same extent.

5 Analysis of Obtained Results An empirical assessment of the graphical representation of the obtained results allows for clear conclusions regarding the nonlinear relationship between the landing time (τ) and the initial height of the UAV. This can also be observed from the non-uniform increase in the values of τ. As for the landing error (ε), although the average error value increases with the initial height, this increase is not significant. To determine the impact of landing height on the error, statistical tests will be conducted. First, it is necessary to determine whether the error value samples obtained during the experiments follow a normal distribution. For this purpose, the classical Shapiro-Wilk test will be used [21]. The Shapiro-Wilk test returns a p-value, which indicates the probability of obtaining such or more extreme results, given that the null hypothesis of data normality is true. If the p-value is less than the chosen significance level (we consider the most common significance level of 0.05), the null hypothesis is rejected, and it is concluded that the data does not follow a normal distribution. Table 3 presents the p-values of the Shapiro-Wilk test for the landing error samples at different heights. Table 3. Results of the Shapiro-Wilk test for landing error samples at different heights. Height, meters

5

10

15

20

p-value

0.46

0.0006

0.029

0.49

366

D. Anikin et al.

The obtained p-values for heights of 10 and 15 m indicate that the data obtained in the corresponding experiments does not follow a normal distribution. In such cases, nonparametric methods are necessary for statistical analysis of the impact of landing height on the error. In this study, we will perform pairwise tests using the Mann-Whitney U test [22]. The Mann-Whitney test allows us to assess whether there are statistically significant differences between two groups of data, without assuming a normal distribution. During the analysis, pairwise comparisons of the ε samples were conducted for each landing height of 5, 10, 15, and 20 m. The null hypothesis assumed no statistically significant differences between these samples. As a result, p_ε values were obtained, which indicate the level of statistical significance of the differences between the samples. In this case, the null hypothesis is rejected if the obtained p-value is less than or equal to 0.05. Table 4 presents the results of the test. Table 4. Results of pairwise Mann-Whitney U test for landing error at different heights. Height, meters

5

10

15

20

5



2.09E−05

5.08E−07

6.04E−08



0.003

0.0004



0.091

10 15

The obtained results allow us to conclude that there is a statistically significant increase in landing error with increasing height, except for the increase from 15 to 20 m. The most significant differences are observed when comparing the error at the initial height of 5 m. Thus, the conducted research leads to the following conclusions: 1. The developed algorithm enables automated landing using a marker at different heights. The maximum height for marker recognition is 20 m, but periodic failures in marker recognition occurred, and any fluctuations of the UAV could affect the detection performance. In the worst cases, this resulted in the inability to perform the landing and the need for a repeated maneuver. Stable marker recognition and landing were achieved at heights of 15 m and below. 2. Increasing the initial landing height leads to a significant increase in landing time and a slight increase in error. If high precision and speed of landing are required, the algorithm should be used at heights not exceeding 5 m. 3. Increasing the height to 15 m allows for stable landing, but the landing time increases by almost 2.5 times compared to the landing from a height of 5 m.

6 Conclusion This article presents the examination and testing of an algorithm for automated landing of UAVs on a moving platform in the Gazebo simulation environment, measuring landing time and spatial error on different initial altitudes. The average landing time and error ranged from 19.64 s and 0.16 m at an initial altitude of 5 m to 121.01 s and 0.27 m at an initial altitude of 20 m. The obtained results demonstrate the effectiveness of the

Autonomous Landing Algorithm for UAV

367

algorithm and its stable performance in landings up to 15 m. To improve performance at higher altitudes, potential enhancements can be considered, such as improving detection algorithms, using more accurate and sensitive cameras, and employing additional sensors. Furthermore, improvements in UAV stabilization systems can be explored to reduce oscillations and enhance landing precision. Future work aims to expand this research by conducting simulation modeling with the influence of external factors on the UAV that can impact the algorithm’s effectiveness, such as wind, lighting conditions, and precipitation. The investigations conducted in this study open up new possibilities for autonomous UAV landings in constrained areas using computer vision. The experimental results can be valuable for the design and development of automatic landing systems for UAVs at various heights and under heterogeneous conditions.

References 1. Saveliev, A., Lebedev, I.: Method of autonomous survey of power lines using a multi-rotor UAV. In: Ronzhin, A., Pshikhopov, V. (eds.) Frontiers in Robotics and Electromechanics. Smart Innovation, Systems and Technologies, vol. 329, pp. 359–376. Springer, Singapore (2023). https://doi.org/10.1007/978-981-19-7685-8_23 2. Saveliev, A., Lebedeva, V., Lebedev, I., Uzdiaev, M.: An approach to the automatic construction of a road accident scheme using UAV and deep learning methods. Sensors 22(13), 4728 (2022) 3. PX4 User Guide. Gazebo Vehicles. https://docs.px4.io/v1.12/en/simulation/gazebo_vehicles. html#quadrotor. Accessed 14 Mar 2023 4. Romero-Ramire, F.J., Munoz-Salinas, R., Medina-Carnicer, R.: Fractal markers: a new approach for long-range marker pose estimation under occlusion. IEEE Access 7, 169908–169919 (2019) 5. Software Overview PX4 Autopilot. https://px4.io/software/software-overview. Accessed 14 Mar 2023 6. About Gazebo. Gazebo Official Website. https://gazebosim.org/about. Accessed 14 Mar 2023 7. Basan, E., Peskova, O., Silin, O., Basan, A., Abramov, E.: Data generation for modeling attacks on UAVs for the purpose of testing intrusion detection systems. Inform. Autom. 21(6), 1290–1327 (2022). https://doi.org/10.15622/ia.21.6.8 8. Sevostyanova, N., Lebedev, I., Lebedeva, V., Vatamaniuk, I.: An innovative approach to automated photo-activation of crop acreage using UAVs to stimulate crop growth. Inform. Autom. 20(6), 1395–1417 (2021). https://doi.org/10.15622/ia.20.6.8 9. Morales, J., Castelo, I., Serra, R., Lima, P.U., Basiri, M.: Vision-based autonomous following of a moving platform and landing for an unmanned aerial vehicle. Sensors 23(2), 829 (2023) 10. Paris, A., Lopez, B.T., How, J.P.: Dynamic landing of an autonomous quadrotor on a moving platform in turbulent wind conditions. In: 2020 IEEE International Conference on Robotics and Automation (ICRA), pp. 9577–9583. IEEE (2020) 11. Feng, Y., Zhang, C., Baek, S., Rawashdeh, S., Mohammadi, A.: Autonomous landing of a UAV on a moving platform using model predictive control. Drones 2(4), 34 (2018) 12. Xing, B.Y., Pan, F., Feng, X.X., Li, W.X., Gao, Q.: Autonomous landing of a micro aerial vehicle on a moving platform using a composite landmark. Int. J. Aerosp. Eng. 2019, 1–15 (2019) 13. Keipour, A., et al.: Visual servoing approach to autonomous UAV landing on a moving vehicle. Sensors 22(17), 6549 (2022)

368

D. Anikin et al.

14. Acuna, R., Zhang, D., Willert, V.: Vision-based UAV landing on a moving platform in gps denied environments using motion prediction. In: 2018 Latin American Robotic Symposium, 2018 Brazilian Symposium on Robotics (SBR) and 2018 Workshop on Robotics in Education (WRE), pp. 515–521. IEEE (2018) 15. Wubben, J., et al.: A vision-based system for autonomous vertical landing of unmanned aerial vehicles. In: 2019 IEEE/ACM 23rd International Symposium on Distributed Simulation and Real Time Applications (DS-RT), pp. 1–7. IEEE (2019) 16. Zhao, Z., et al.: Vision-based autonomous landing control of a multi-rotor aerial vehicle on a moving platform with experimental validations. IFAC-PapersOnLine 55(3), 1–6 (2022) 17. Husky UGV - Outdoor Field Research Robot by Clearpath. https://clearpathrobotics.com/ husky-unmanned-ground-vehicle-robot/. Accessed 17 Mar 2023 18. Controller Diagrams - PX4 User Guilde. https://docs.px4.io/main/en/flight_stack/contro ller_diagrams.html. Accessed 17 Mar 2023 19. Denisenko, V.V.: PID controllers: principles of construction and modification. Mod. Autom. Technol. 4, 66–75 (In Russ.) 20. Gautam, A., Ratnoo, A., Sujit, P.B.: Log polynomial velocity profile for vertical landing. J. Guid. Control. Dyn. 41(7), 1617–1623 (2018) 21. Shapiro, S.S., Wilk, M.B.: An analysis of variance test for normality (complete samples). Biometrika 52(3/4), 591–611 (1965) 22. Mann, H., Whitney, D.: Controlling the false discovery rate: a practical and powerful approach to multiple testing. Ann. Math. Stat. 18(1), 50–60 (1947)

Curl-Free Vector Field for Collision Avoidance in a Swarm of Autonomous Drones Tagir Muslimov(B) Ufa University of Science and Technology, 12, K. Marx St., Ufa 450008, Russia [email protected]

Abstract. To perform complex tasks, drones must have the ability to move autonomously. Ensuring collision avoidance is very important for the safe movement of autonomous drones indoors. FIRAS function-based potential field method is the standard for collision avoidance as implemented in isolated drones. However, its use in an autonomous swarm can be problematic. Its complex interconnected structure causes one of the known issues when there are multiple simultaneously active control objectives. They are intra-swarm collision avoidance and reaching the target relative distance (normally referred to as formation control). This paper shows that with collision avoidance active, the standard potential field method will cause a local minima-like effect in an autonomous swarm. To prevent it, this paper proposes a modified curl-free vector field-based algorithm. This modification enables extended lateral circular motion to prevent swarm members from getting stuck in a local minimum. Stability theory methods are invoked to show that the formation remains stable when running the proposed algorithm. Comparative numerical experiments were run on a drone swarm model in MATLAB/Simulink to illustrate the functioning of this algorithm. To prove the proposed method effective, the paper presents simulation results for standard vs modified potential field. Keywords: Potential Field Method · Indoor Drones · Swarm Control · Collision Avoidance · Multi-UAV System

1 Introduction Research into controlling swarms of autonomous robots is commonplace and of great interest [1]. A variety of research problems have drawn attention, including distribution of tasks between swarm members [2] and novel types of robots for better swarm efficiency [3]. Consensus-based formation control, including that of autonomous unmanned aerial vehicles (drones) is a closely related topic [4, 5]. Yet, controlling a swarm or a decentralized formation of autonomous drones is a complex problem if it involves avoidance of collision with a diverse set of obstacles [6]. That is critical where drones must be moving safely in a confined space. In such cases, avoiding drone-to-drone collision avoidance becomes especially relevant due to limited flight space.

© The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 369–379, 2023. https://doi.org/10.1007/978-3-031-43111-1_33

370

T. Muslimov

Isolated drone movement amidst obstacles is a well-researched problem with a variety of solutions, e.g., those based on rapidly exploring random trees (RRT) [7]. Potential field is another classical method; it was originally proposed to enable isolated robots to navigate safely through stationary obstacles. However, this method has recently found uses in controlling formations of mobile ground robots or drones where they should avoid collisions within the team [8, 9]. One possible implementation of the potential field method consists in splitting the flight space for each drone into individual regions with a version of the potential function for each [10, 11]. This approach has been proven effective when controlling a formation in a stationary-obstacle environment. However, if the swarm has to move cohesively with respect to intra-formation target points, then a conflict of control objectives might arise. This conflict causes drones to become ‘stuck’, an effect similar to that of local minima, which is typical of the classical potential field method. Thus, the goal hereof was to address this issue. To do so, the paper proposes curlfree vector field-based modification. A similar approach has been proposed in [12, 13]. However, paper [12] demonstrated using it to avoid collision of two drones facing each other in flight on a collision course. Paper [13] considers a multi-obstacle flight zone but the obstacles are static. In [14], a curl-free vector field was used to create a circular formation that preserves its arrangement around the target while tracking the trajectory. This paper is about a modification for cases where the swarm has to reach preset relative distances whilst prioritizing formation control. We further demonstrate the theoretical foundations and the numerical simulation to prove this approach effective.

2 Curl-Free Vector Field-Based Approach to Drone Swarm Control 2.1 Drone Swarm System Model The proposed model is based on an assumption that each micro-quadcopter in the system is receiving data respective of the adjacent aircraft and of the global coordinate system. Assume that the internal control loop (the autopilot) stabilizes the UAV. This can be verified by means of a test flight that performs hovering at the same spot as well as manual control by entering movement commands in four directions and commands to increase or decrease the altitude. The external control loop tracks the predetermined trajectory using special-purpose software. Control is proved by a central computer; however, quadcopters can only access local data on the adjacent aircraft, which makes the formation decentralized. Thus, the high-level control dynamics can be utilized to develop and implement swarm control. The swarm control model has a fairly standard description, see e.g. [15] for a general overview. Consider a swarm of N micro-quadcopters, each referred to as an ‘agent’. For high-level control, assume that each quadcopter is a system in n-dimensional space that has the following single-integrator dynamics: p˙ i = ui , i = 1, 2, . . . , N , where pi ∈ Rn is the aircraft’s attitude, ui ∈ Rn is the ith quadcopter’s control.

Curl-Free Vector Field for Collision Avoidance

371

Let us introduce the notation: Δij ∈ Rn is the desired relative distance between the ith and the jth agent. Apparently, for the swarm to operate correctly, the condition Δij = −Δji must be satisfied for each ith and jth agent pair except when i = j. For any i, the condition Δii = 0 must be satisfied. As a problem, swarm formation control can be worded as follows. Control Objective. Consider a swarm of N agents (quadcopters). For each ith agent, formation synthesize such control ui (t) that for t ≥ 0, the attitude satisfies the condition:   lim pj (t) − pi (t) = Δij ∀i, j, i = j. t→∞

Classical papers on the topic usually show how this could be achieved by using a control algorithm that relies on relative attitude mismatch: formation

ui

= kf

N   pj − pi − Δji , j=1

(1)

where kf is the positive tunable coefficient. We can analyze unidimensional (n = 1) formation control for a single axis. That analysis could be easily reapplied to a multidimensional case (n ≥ 1). Swarm system dynamics can be written as a matrix using the Laplacian matrix L as follows: p˙ = −kf Lp + b,  T T   where p = p1 , p2 , . . . , pN , b = b1 , b2 , . . . , bN , bi = −kf N j=1 Δji , L ∈ N ×N R is the Laplacian matrix that represents a graph of agent-agent interaction in the swarm. By default, the matrix L is written as L = D − A, where D ∈ RN ×N is a matrix where the diagonal element i corresponds to the in-degree of the ith vertex (the number of edges coming into that vertex); A ∈ RN ×N is the adjacency matrix of the interaction graph. For example, for a fully connected interaction structure the elements of the matrix L would be written as  −1, i = j lij = . N − 1, i = j Such Laplacian matrix only has one zero eigenvalue; other eigenvalues are positive and identical. This UAV swarm control model has been modified using an artificial potential field. We named the complete potential field-based swarm control algorithm APFfSwarm (Artificial Potential Field for Swarm). 2.2 Control Algorithm for Collision Avoidance Consider a case of swarm movement with formation control and collision avoidance. APFfSwarm for the swarm control algorithm is proposedly The artificial potential field Ur written as the sum of the formation control field and curl-free vector field as follows:    APFfSwarm CAfM  CAfS qi , qj + Ur (qi , dobst ), = UrForm qi , qj + Ur Ur

372

T. Muslimov

  where UrForm qi , qj is the component that controls making the formation;  CAfM  Ur qi , qj is the component that controls collision avoidance for moving obstacles  CAfS  (i.e., copter-to-copter collision avoidance); Ur qi , qj is the component that controls

T  y T are the collision avoidance for stationary obstacles; qi  pix piy and qj  pjx pj coordinates in the inertial coordinate system (ICS).   is defined The component UrForm qi , qj

  y T 1 N 2 x Form qi , qj  kf 2 j=1 qj − qi − Δji 2 given (1), where Δji  Δji Δji as Ur specifies the desired formation in the horizontal plane.  CAfM  CAfS The components Ur qi , qj and Ur (qi , dobst ) are defined as follows: ⎧  2   1  ⎨ 1 km − d1o , if d qi , qj ≤ do CAfM  r 2 i d q ,q ( i j) qi , qj = , Ur   ⎩ 0, if d q , q > d i

 CAfS

Ur

(qi , dobst ) =

1 s 2 kri



1 d (qi ,dobst )

0,



1 do

2

j

o

, if d (qi , dobst ) ≤ do if d (qi , dobst ) > do

,

where krmi and krsi are the positive tunable coefficients; do is the radius of the area around CAfS

is defined similarly the drone where the repulsion components begin to take effect; Ur CAfM to Ur but uses dobst (the distance to the nearest point of a stationary obstacle) instead of qj . For the ith UAV agent, define Control Algorithm 1 based on the control law APFfSwarm fi  consists of several components:  , which If d qi , qj ≤ do , then     = f escape qi , qj + f CAfM qi , qj + f CAfS (qi , dobst )    CAfM  qi , qj Rςij qi − qj = −∇qi −qj Ur CAfS −∇qi −dobst Ur (qi , dobst )Rςobst (qi − dobst ) +f escape qi , qj

APFfSwarm

fi

else APFfSwarm

fi

  = ∇UrForm qi , qj

Curl-Free Vector Field for Collision Avoidance

373

end The matrices Rςij and Rςobst are defined as follows:  ⎧ ⎪ ⎪ 0 1 if ςij = −1 ⎨   −1 0  ; Rςij qi − qj =  ⎪ 0 −1 ⎪ ⎩ if ςij = 1 1 0  ⎧ 0 1 ⎪ ⎪ if ςobst = −1 ⎨ −1 0  Rςobst (qi − dobst ) =  . ⎪ 0 −1 ⎪ ⎩ if ςobst = 1 1 0 Here, ςij and ςobst are the parameters that, in the program code, set the    copter rotation direction for collision avoidance maneuvers; f escape qi , qj    2 μi j=i qj − qi /qj − qi − Δji 2 is the vector that enables both UAVs to leave the safety zone and continue swarming,   μi is a tunable positive coefficient. Assume that μi = μj . The vector f escape qi , qj is based on the final positions of the swarm formation; this allows to make the formation faster. Let us demonstrate how this algorithm guarantees Lyapunov stability. For space considerations, consider that drones only encounter moving obstacles. Adjustment for stationary obstacles would only require a slight modification. Choose the following Lyapunov function: ⎧    CAfM  1 N  ⎨ if d qi , qj ≤ do 2 κ i=1 j=i Ur  qi , qj ,  , V = 1 N T   1 ⎩ qi − kf1N bi , if d qi , qj > do i=1 qi − kf N bi 2 where κ ∈ R>0 is a positive coefficient. A derivative of the function V can be written as follows: ⎧       ˙ CAfM qi , qj if d qi , qj ≤ do ⎨ 1κ N i=1 j=i Ur 2 T ˙ = . V   N  1 ⎩ q˙ i , if d qi , qj > do i=1 qi − kf N bi   Consider the case d qi , qj ≤ do using Control Algorithm 1. The derivative   CAfM qi , qj can be written as U˙ r    CAfM  CAfM T qi , qj = ∇pij Ur p˙ ij , U˙ r CAfM

CAfM

where pij  qi − qj . Let us introduce the notation Ur (qk , ql )  Ur,kl . If   d qi , qj ≤ do , then, in view of the control algorithm, the Lyapunov function’s derivative is written as follows:      N N      f escape qi , qj 1   ˙ CAfM  CAfM T   ∇pij Ur,ij κ qi , qj = κ Ur . CAfM −∇pij Ur,ij Rςij qi − qj 2 i=1 j=i

i=1 j=i

374

T. Muslimov

This equation transforms as ⎡ ⎧ 2 ⎤ ⎫ ⎪ ⎪ x x ⎪ ⎪ ⎪ ⎢ pj − pi + ⎥ ⎪ ⎪ ⎪ ⎪ ⎪ −μ K   ⎣ ⎦ ⎪ ⎪ i escape 2 ⎪ ⎪ y y ⎪ ⎪ ⎪ ⎪ p − p N N ⎨ ⎬ j i       1    CAfM ˙ y y κ qi , qj = κ Ur x x K pi − pj pi − pj ⎪ 2 ⎪    ⎪ i=1 j=i i=1 j=i ⎪ ⎪ ⎪ + ⎪ ⎪ y y ⎪ ⎪ x x ⎪ ⎪ −K p p − p − p ⎪ ⎪ i j i j ⎪ ⎪ ⎪ ⎪  ⎩  ⎭ =0 ⎧ ⎡ 2 ⎤⎫ x − px + ⎪ N ⎪ ⎨ ⎬  p i ⎢ j ⎥ =κ −μi Kescape ⎣  2 ⎦ . y y ⎪ ⎪ ⎭ + pj − pi i=1 j=i ⎩ The following notation is used:   1 1 1 − ≥ 0, Kescape  μi    3 do d qi , qj qi − qj − ji 22 d qi , qj 2   2  1  1 1 − ≥ 0. since d qi , qj ≤ do ; and K  krmi do d (qi ,qj ) d 6 (qi ,qj ) In the case of the other criterion of the collision   danger area,  the following can be  y y y x x x noted. For example, given that pj − pi − Δji ∧ pj − pi − Δji < 0, as the copters are in the ‘risk-of-collision zone’, as aircraft-to-aircraft distances are below the algorithm’s threshold  they are in the zone.  whilst If d qi , qj > do , then the reasoning is partly similar to the one in [15]; however, it is modified to account for the copter’s position outside the danger zone and for the system model’s greater dimensionality. Note that the vector b is an eigenvector for the Laplacian matrix L with the corresponding eigenvalue N: Lb = N b. The Lyapunov ˙ is then written as: function’s derivative V %T N  $     1 ˙ qi − bi −kf qi − qj + bi V= i=1 j=i kf N $ %T $ % N    1 1 qi − qi − qj − bi bi = −kf kf N kf i=1 j=i $ $ %T % 1 1 b (L ⊗ I2 ) q − b ≤ 0, = −kf q − kf N kf N where I2 is a second-order unit matrix. We used the following notation: q



y T

p1x p1

  x y T T  T T T and b  bx1 by1 , . . . , bxN byN , , . . . , pN pN bxi  −kf

N j=1

y

Δxji and bi  −kf

N j=1

y

Δji .

Curl-Free Vector Field for Collision Avoidance

375

We thus conclude that the derivative along the trajectories of the system for the Lyapunov function is negative semidefinite. Thus, given that the function V is positive definite, Lyapunov stability is ensured. Therefore, drones will not collide in motion because  the Lyapunov function is limited for limited states. A collision event implies that d qi , qj is zero, in which case the Lyapunov function’s value would tend to infinity.

3 Numerical Simulation Results and Discussion To visualize the effectiveness of the curl-free vector field for collision avoidance between swarming UAVs (i.e., moving obstacles), we simulated a specific case. The swarm was reduced to 4 UAVs; the final geometry of the formation was preset. Initial positions were picked to force UAV#4 to overtake UAV#3 in order to make the for-mation. Formationmaking and trajectory-following coefficients were picked to have such a ratio as to cause such overtaking to result in the collision of the involved UAVs if collision avoidance were not engaged. Numerical simulation shows that the standard FIRAS function-based potential field fails to prevent such collision. The distance between UAV#3 and UAV#4 begins to shorten drastically upon entering the ‘risk-of-collision zone’; it goes below 0.1 as a result, which creates a high risk of collision. Such UAV behavior is attributable to the fact that the term whose purpose is to make the formation actually causes UAVs to come too close to each other. The reason is that the formation-making action actually involves closing on, and overtaking another aircraft, which is necessary to shape the final geometry. At this point, UAVs #3 and #4 are virtually on the same line of movement; the standard FIRAS function-based potential field cannot generate a sufficient control action to initiate a lateral evasive maneuver. The situation is somewhat similar to local minima problem. It is a completely different outcome with the proposed curl-free vector field-based algorithm. This approach prevents UAVs from ‘getting stuck’ in the longitudinal axis and generates a control action that enables a lateral evasive maneuver, allowing the UAV to keep a safe distance. Figure 1 shows the swarm’s trajectories (Fig. 1a) and the distance between UAV#3 and UAV#4 (Fig. 1b), no collision avoidance algorithm. It also shows the distance from each UAV to a small obstacle near point (6; 6), which is used as reference (Fig. 1c). This distance was for testing whether the formation’s final geometry matched the preset. The collision event (zero distance between UAVs #3 and #4) is clearly seen. Figure 2 compares two algorithms: the FIRAS function-based one to the left and the proposed curl-free vector field-based one to the right. Figures 2a and 2b show the swarm trajectories. Figures 2c and 2d show the distances between UAV#3 and UAV#4. Figures 2e and 2f show each drone’s distance to the reference. As repulsion forces cause a substantial increase in speed, the distance dot plot is discontinuous in the ‘risk-ofcollision’ zone. However, note that at some point, the distance between UAVs #3 and #4 goes below 0.1 when using the standard potential field. The alternative, curl-free vector field prevents the value from ever going below 0.7. The comparative graphs also visualize the difference in movement trajectories as produced by the two algorithms. The distance-to-reference graph demonstrates that the final geometry was also attained.

376

T. Muslimov 7 6 5 4

UAV No.3 trajectory UAV No.4 trajectory

3 2 1 0 -1

a)

-2

-2

0

2

4

6

Distance, [m]

1.5

1

0.5

0

b)

0

5

10

15

20

25

Time, [s]

12

Distances, [m]

10 8 6 4 2 0

c)

0

5

10

15

20

25

Time, [s]

Fig. 1. Swarm trajectories, distance between UAV#3 and UAV#4, distances to the reference point.

Curl-Free Vector Field for Collision Avoidance

a)

b)

c)

d)

e)

377

f) Fig. 2. Collision avoidance algorithm comparison.

4 Conclusion This paper investigates a drone swarm system model that enables avoidance of collision with various types of obstacles. To that end, we implemented a curl-free vector field-based modification. Focus was made on drone-to-drone collision avoidance, representing moving obstacles. The paper proves the algorithm to be Lyapunov-stable. A MATLAB/Simulink simulation was run and proved the proposed method effective. Its results demonstrate that the method addresses the potential field’s weakness in swarm control. This weakness lies in causing a local minima-like effect of ‘getting stuck’. The

378

T. Muslimov

model and approach proposed herein are sufficiently versatile and could be modified to handle more complex swarming scenarios. Acknowledgements. The study was funded by a grant from the Russian Science Foundation (RSF) (project № 22-79-00168), https://rscf.ru/en/project/22-79-00168/.

References 1. Zakiev, A., Tsoy, T., Magid, E.: Swarm robotics: remarks on terminology and classification. In: Ronzhin, A., Rigoll, G., Meshcheryakov, R. (eds.) ICR 2018. LNCS (LNAI and LNB), vol. 11097, pp. 291–300. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-995823_30 2. Petrenko, V., Tebueva, F., Antonov, V., Ryabtsev, S., Sakolchik, A., Satybaldina, D.: Evaluation of the iterative method of task distribution in a swarm of unmanned aerial vehicles in a clustered field of targets. J. King Saud Univ.-Comput. Inf. Sci. 35, 283–291 (2023). https:// doi.org/10.1016/j.jksuci.2023.02.022 3. Darush, Z., Martynov, M., Fedoseev, A., Shcherbak, A., Tsetserukou, D.: SwarmGear: heterogeneous swarm of drones with reconfigurable leader drone and virtual impedance links for multi-robot inspection (2023). http://arxiv.org/abs/2304.02956 4. Popov, A.M., Kostrygin, D.G., Shevchik, A.A., Andrievsky, B.: Speed-gradient adaptive control for parametrically uncertain UAVs in formation. Electron. 11, 4187 (2022). https://doi. org/10.3390/ELECTRONICS11244187 5. Muslimov, T.Z., Munasypov, R.A.: Multi-UAV cooperative target tracking via consensusbased guidance vector fields and fuzzy MRAC. Aircr. Eng. Aerosp. Technol. 93, 1204–1212 (2021). https://doi.org/10.1108/AEAT-02-2021-0058 6. Izhboldina, V., Lebedev, I.: Group movement of UAVs in environment with dynamic obstacles: a survey. Int. J. Intell. Unmanned Syst. 11, 256–284 (2022). https://doi.org/10.1108/IJIUS06-2021-0038 7. Pshikhopov, V., Medvedev, M., Kostjukov, V., Houssein, F., Kadhim, A.: Trajectory planning algorithms in two-dimensional environment with obstacles. Inform. Autom. 21, 459–492 (2022). https://doi.org/10.15622/ia.21.3.1 8. de Angelis, E.L., Giulietti, F., Rossetti, G.: Multirotor aircraft formation flight control with collision avoidance capability. Aerosp. Sci. Technol. 77, 733–741 (2018). https://doi.org/10. 1016/J.AST.2018.04.002 9. Karkoub, M., Atınç, G., Stipanovic, D., Voulgaris, P., Hwang, A.: Trajectory tracking control of unicycle robots with collision avoidance and connectivity maintenance. J. Intell. Robot. Syst. 96, 331–343 (2019). https://doi.org/10.1007/s10846-019-00987-2 10. Qiao, Y., et al.: Formation tracking control for multi-agent systems with collision avoidance and connectivity maintenance. In: Drones 2022, vol. 6, p. 419 (2022). https://doi.org/10.3390/ DRONES6120419 11. Wang, N., Dai, J., Ying, J.: UAV formation obstacle avoidance control algorithm based on improved artificial potential field and consensus. Int. J. Aeronaut. Sp. Sci. 22, 1413–1427 (2021). https://doi.org/10.1007/S42405-021-00407-6/ 12. Choi, D., Lee, K., Kim, D.: Enhanced potential field-based collision avoidance for unmanned aerial vehicles in a dynamic environment. In: AIAA Scitech 2020 Forum. American Institute of Aeronautics and Astronautics, Reston, Virginia (2020). https://doi.org/10.2514/6.20200487

Curl-Free Vector Field for Collision Avoidance

379

13. Choi, D., Kim, D., Lee, K.: Enhanced potential field-based collision avoidance in cluttered three-dimensional urban environments. Appl. Sci. 11, 11003 (2021). https://doi.org/10.3390/ APP112211003 14. Dang, A.D., La, H.M., Nguyen, T., Horn, J.: Formation control for autonomous robots with collision and obstacle avoidance using a rotational and repulsive force–based approach. Int. J. Adv. Robot. Syst. 16, 172988141984789 (2019). https://doi.org/10.1177/1729881419847897 15. Toksoz, M.A., Oguz, S., Gazi, V.: Decentralized formation control of a swarm of quadrotor helicopters. In: 2019 IEEE 15th International Conference on Control and Automation (ICCA), pp. 1006–1013. IEEE (2019). https://doi.org/10.1109/ICCA.2019.8899628

Author Index

A Abbyasov, Bulat 116 Aida-Zade, Kamil 197 Akimenko, Tatiana 289 Akimova, Elizaveta 221 Alexandrov, Vadim 334, 345 Andrievsky, Boris 221 Anikin, Dmitry 357 Apurin, Artem 116 Asgarov, Ali 70

H Hamdan, Nizar 137 Hashimov, Vugar 197 Huang, Congyu 13

B Barinov, Arseniy 279 Bayramov, Azad 175 Blekanov, Ivan 185 Bogomolov, Alexey 289 Brosalin, Dimitry 137 Bzhikhatlov, Kantemir 59

K Kabalan, Ali 208 Kiselev, Gleb 81 Klimenko, Anna 279, 311 Knyazev, Andrey 93 Konstantinov, Andrei 233 Kosenko, Evgeny 208 Kostyukov, Vladimir 322 Kuznetsov, Nikolay 221

C Chebotareva, Elvira 23 Contreras, Luis 36 D Dobrynin, Dmitry

127

E Evdokimov, Igor 322 F Fedorov, Andrey

93

G Gaiduk, Anatoly 208 Gissov, Vladislav 208, 322 Gorbunova, Yaroslava 81 Gorodetskiy, Andrey 258 Guliyev, Samir 246 Gurenko, Boris 137

I Iskhakova, Anastasia 103 J Jatsun, Sergey

93

L Larkin, Eugene 289 Li, Hongbing 23 Li, Jie 13 Liu, Xiaofeng 13 Lukyanov, Vadim 47 M Magid, Evgeni 23, 116 Mambetov, Idar 59 Martínez-García, Edgar A. 116 Medvedev, Mikhail 137, 208 Melnik, Eduard 300 Meshcheryakov, Roman 103, 345 Mironov, Konstantin 258 Mitrofanov, Evgenii 185 Mitrofanova, Olga 185 Morozov, Yury 334

© The Editor(s) (if applicable) and The Author(s), under exclusive license to Springer Nature Switzerland AG 2023 A. Ronzhin et al. (Eds.): ICR 2023, LNAI 14214, pp. 381–382, 2023. https://doi.org/10.1007/978-3-031-43111-1

382

Author Index

Muliukha, Vladimir 233 Muslimov, Tagir 369 Mustafin, Maksim 23

Serebrenny, Vladimir 1 Sevostyanov, Danila 185 Sharov, Vadim 289 Shatov, Dmitrii 334, 345 Shen, Xin 1, 47 Suleymanov, Samir 175

N Nagoeva, Olga 59 Negrete, Marco 36 O Okada, Hiroyuki 36 Ortuno-Chanelo, Stephany

T Trefilov, Peter 345 Turovsky, Yaroslav 103 36

P Panov, Aleksandr 258 Parsayan, Ali 70 Podtikhov, Artur 164 Pshenokova, Inna 59 Pshikhopov, Viacheslav 137, 208 R Rezkov, Ilya 334, 345 Ryabinov, Artem 357 S Safarov, Jamil 93 Safronenkova, Irina 300 Sakamaki, Arata 36 Savage, Jesus 36 Saveliev, Anton 164, 357 Savinov, Vladislav 151 Semenov, Alexander 357 Semenov, Viktor 267

U Unagasov, Alim 59 Utkin, Lev 233 V Vasileva, Maria 137 W Wang, Ziyang 13 Wolf, Daniyar 103, 345 Wu, Guo 1, 47 Y Yakovlev, Konstantin

151

Z Zaborovsky, Vladimir 233 Zaitceva, Iuliia 221 Zhang, Jia 185 Zhu, Haoran 13