Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

A task level fusion autonomous switching mechanism

  • Bingyu Lv,

    Roles Data curation, Formal analysis, Software, Writing – original draft, Writing – review & editing

    Affiliations College of Computer Science and Technology, Jilin University, Changchun, Jilin, China, Key Laboratory of Symbolic Computation and Knowledge Engineering, Jilin University, Changchun, Jilin, China

  • Xianchang Wang,

    Roles Software, Validation

    Affiliations College of Computer Science and Technology, Jilin University, Changchun, Jilin, China, Key Laboratory of Symbolic Computation and Knowledge Engineering, Jilin University, Changchun, Jilin, China, Chengdu Kestrel Artificial Intelligence Institute, Sichuan, China

  • Rui Zhang

    Roles Conceptualization, Software, Supervision

    rui@jlu.edu.cn

    Affiliations College of Computer Science and Technology, Jilin University, Changchun, Jilin, China, Key Laboratory of Symbolic Computation and Knowledge Engineering, Jilin University, Changchun, Jilin, China

Abstract

Positioning technology is an important component of environmental perception. It is also the basis for autonomous decision-making and motion control of firefighting robots. However, some issues such as positioning in indoor scenarios still remain inherent challenges. The positioning accuracy of the fire emergency reaction dispatching (FERD) system is far from adequate to support some applications for firefighting and rescue in indoor scenarios with multiple obstacles. To solve this problem, this paper proposes a fusion module based on the Blackboard architecture. This module aims to improve the positioning accuracy of a single sensor of the unmanned vehicles within the FERD system. To reduce the risk of autonomous decision-making of the unmanned vehicles, this module uses a comprehensive manner of multiple channels to complement or correct the positioning of the firefighting robots. Specifically, this module has been developed to fusion a variety of relevant processes for precise positioning. This process mainly includes six strategies. These strategies are the denoising, spatial alignment, confidence degree update, observation filtering, data fusion, and fusion decision. These strategies merge with the current scenarios-related parameter data, empirical data on sensor errors, and information to form a series of norms. This paper then proceeds to gain experience data with the confidence degree, error of different sensors, and timeliness of this module by training in an indoor scenario with multiple obstacles. This process is from data of multiple sensors (bottom-level) to control decisions knowledge-based (up-level). This process can obtain globally optimal positioning results. Finally, this paper evaluates the performance of this fusion module for the FERD system. The experimental results show that this fusion module can effectively improve positioning accuracy in an indoor scenario with multiple obstacles. Code is available at https://github.com/lvbingyu-zeze/gopath/tree/master.

Introduction

With the rapid development of robots, they are used in various fields [1]. Researchers have focused on firefighting robots that have reduced casualties among rescuers. When a fire occurs in an indoor environment, the firefighting robot can adjust to the dynamic indoor environment in order to achieve a safe rescue or accurate firefighting [2, 3]. Accurate positioning is very important for firefighting robots in order to be able to do rescue work accurately [4]. In indoor scenarios where firefighting robots perform tasks, GPS signals are often unavailable [5]. The current indoor positioning methods have many methods, including ultra wide band (UWB), inertial measurement unit (IMU), infrared depth sensor (IDS), camera [4, 611]. UWB has high temporal resolution but is susceptible to indoor environments [4, 6]. The IMU is highly accurate, but there is a cumulative error [7]. IDS is easily detected at close range but has limited detection distance, high noise, and low accuracy [8, 9]. Camera accuracy is great, but the real-time performance is poor and vulnerable to light conditions [10, 11]. These sensors can be used to their advantage in different usage scenarios and are difficult to replace. Therefore, the problem of indoor positioning is still inherent.

The single sensor has limited information and is vulnerable to environmental influences. As a result, firefighting robots usually carry a large number of different types of sensors. The information acquired by multiple sensors may complement or contradict each other [12, 13]. The information acquired by each sensor has to be processed separately, increasing the workload. Therefore, multi-sensor fusion is necessary to process the information. Information fusion can improve information reliability, robustness, and utilization [4, 5, 7]. The existing research on multi-sensor fusion framework mainly focuses on the response to specific situations [7, 14, 15]. Sara et al. [14] proposed a three-level positioning framework based on additional visual sensors. The framework obtains the absolute position of the vehicle by tracking landmarks and also improves the GNSS measurement method. Li et al. [7] designed an indoor positioning method based on multiple sensors. They also constructed an extended Kalman filtering (EKF) algorithm framework. Dasgupta et al. [15] proposed sensor fusion software for autonomous vehicles and other autonomous robots to achieve precise positioning of autonomous robots. However, multi-sensor fusion is susceptible to environmental interference, leading to measurement errors and uncertainties. Therefore, facing a large amount of uncertain data, multiple domain experts are needed to process this data. The blackboard architecture has proven to be a very useful and flexible mechanism [16, 17]. In this paper, a fusion module in FERD system is proposed based on the Blackboard Architecture. The overall goal is achieved by quickly and autonomously switching different strategies for multiple unmanned vehicles. The position of unmanned vehicles is supplemented or modified by multiple channels. These strategies can reduce the impact of inaccurate positioning in an open environment, also reduce the risk of autonomous decision-making of each unmanned vehicle. Then, this fusion module in FERD system is realized in the task of firefighting. Finally, the performance of the proposed fusion module is evaluated. The main contributions of this paper are summarized as follows.

  • In this paper, a fusion module is designed and added to fire emergency reaction dispatching (FERD) system. This fusion module is used for accurately positioning fire equipment by multiple sensors. All sensors send their measured position information as primary data to this module. The position information of the real physical position of each fire equipment is used as auxiliary data. This module has developed a variety of fusion-related processing for precise positioning. These processes mainly include six strategies. They are the denoising, spatial alignment, confidence degree update, observation filtering, data fusion, and fusion decision. Compared to a single sensor, this paper proposed module is effective and scalable.
  • Multiple unmanned vehicles perform firefighting tasks and collect data in an indoor scenario with multiple obstacles. This demonstrates the effectiveness of the module in locating sensors in the field. This module merges with the current scenarios-related parameter data, empirical data on sensor errors, and information to form a series of norms. This paper then proceeds to gain experience data with the confidence degree, error of different sensors, and timeliness of this module by training in an indoor scenario with multiple obstacles. This process is from data of multiple sensors (bottom-level) to control decisions knowledge-based (up-level). This process can obtain globally optimal positioning results.

The remainder of this paper is organized as follows. Section RELATED WORKS introduces the related to the work of this paper. Section PRELIMINARIES introduces a few concepts and reviews previous works. Then, a few basic terms are provided that will be used in later sections. Section PROPOSED MODULE, this section emphasizes proposed module. Section RESULTS analyzes the errors of different sensors in detail. Then, this section introduces how to implement the fusion module based on Blackboard Architecture. Finally, this section gives the experimental results and analysis. Section DISCUSSION discusses the advantages and disadvantages of the proposed fusion module. Section CONCLUSION summarizes the significance of the fusion module and discusses future work.

Related works

This paper presents a novel and comprehensive sensor fusion scheme. The focus of this work is on the fusion method and the switching of fusion methods.

Kalman filtering

Some existing works with us have related ideas to Kalman Filtering as one of the data fusion methods. For the indoor positioning problem of miniature tracked robots, Li et al. [7] designed an indoor positioning method. This method is based on Bluetooth, a gyroscope, an accelerometer, a magnetometer, and other sensors. This method takes an inertial navigation system (INS) as the core, Bluetooth AOA positioning base station as the position observation, and magnetometer as the heading angle observation. They also constructed an EKF algorithm framework. This EKF is based on error states. Kim and Lee [18]proposed the EKF algorithm. This algorithm is a combination of a Camera, GPS, and sensor of in-vehicle for the precise positioning of vehicles. This algorithm combines multiple decisions made by different sensors and finally makes a fusion decision. Luo et al. [19] proposed a joint Kalman filter to fuse the positioning information. This information is the combined localization subsystem. This method provides critical position information for achieving high accuracy and efficient missions. The work of this paper determines whether to fuse the data from multiple sensors through confidence degree and error size. Then this work of this paper uses the Kalman filter to fuse the data. In this way, the proposed approach provides critical position information during operation. The proposed approach improves the positioning accuracy and efficient tasks of FERD system.

Multi-sensor fusion based blackboard architecture

The proposed fusion module combines the norms of data fusion and error parameters of multiple sensors with the Blackboard Architecture. Benjamin et al. [20] proposed that the design of the transmission trigger and estimator should be closely combined. The intelligent trigger mechanism of sensors can predict future information about these sensors. The purpose of this method further reduces the communication demand. Girraj et al. [21] optimized and evaluated thresholds of fusion norms. They proposed an iterative algorithm to calculate the individual optimal threshold of fusion norms. Based on the necessary and sufficient conditions of the optimal sensor norm, Liao et al. [22] derived a set of norms for sensors. These norms are significantly superior to the traditional iterative search algorithm. These norms are applied to multi-sensor decision fusion problems with high-dimensional sensors. Yang et al. [23] proposes a sensor fusion algorithm. They create redundancy, by using multiple sensors to measure the same physical variables. This algorithm is used to provide robust estimation with correct sensor information. These controllers reduce estimation errors and communication channel noise. However, multi-sensor fusion is susceptible to environmental interference, leading to measurement errors and uncertainties. The difference between the method proposed in this paper and these methods is that the accuracy of multiple sensors is considered, and some norms about accuracy are designed for fusion. In different environments, our method can filter the measurement accuracy of multiple sensors and switch the corresponding strategies about fusion. Since norms of accuracy require a combination of parametric and empirical data related to the current environment, firefighting robots need to be trained to obtain these data. These training values can be used as empirical values for FERD system to perform tasks. In this way, the method proposed in this paper is more likely to improve the correctness of decisions and reactions during the operation of FERD system.

Preliminaries

In this section, the Blackboard System (BBS) and General Blackboard Open Source (GBBopen) are first introduced. Then, previous work and some concepts are introduced. Finally, the basic concepts covered in this paper are introduced in detail.

Blackboard system and generic blackboard open source

(1) Blackboard system. BBS is an artificial intelligence technique based on knowledge bases and collaborative processing, which is a distributed computing model. It is widely used in complex problem solving, for example, in the field of decision support. It enables effective collaborative processing of multiple specialized modules to achieve more intelligent and efficient problem solving. The operation of the BBS is driven by events. It comprises a knowledge source (KS), blackboard (BB), and control shell (CS). In the BBS, a KS is similar to an expert. Each KS has its own expertise and capabilities. These KSs work together through a central data structure (BB) to solve a complex problem. BB is a shared data structure. Specifically, the information on the BB can be the description of the problem, the solution of the problem, etc. All KSs can read and write the information on BB to achieve the goal of solving problems together. A KS can write the problem or the solution on BB. KS can also read the information on the BB and process and solve the problem according to its knowledge and abilities. CS is an important part of the BBS. It is responsible for monitoring the information on the BB as well as coordinating and controlling the interaction and decision-making process between the various KSs. When new information appears on the BB, the CS can decide which KSs can read it and how to deliver it to other KSs according to predefined norms and strategies.

(2) Generic blackboard open source. GBBopen is an open-source blackboard system. It can be used to build various types of intelligent systems and applications. GBBopen provides a range of tools and supports multiple programming languages and operating systems. It can run on different platforms and be extended and customized on demand. Developers are able to quickly build and customize their own blackboard systems, implementing various types of intelligent applications and systems. In this paper, GBBopen is developed and implemented in the common list processing (Common Lisp) language environment. An expert system based on the norm-base developed by Clips (https://www.clipsrules.net) is a data-driven program. Unlike Clips, GBBopen does not require a data abstraction design; therefore, the implementation of a function may be completed quickly during development.

In GBBopen, the unit class is the base class for all classes defined. The properties of KS are ‘trigger-events’, ‘precondition-function’ and ‘execution-function’. Each knowledge source activation (KSA) is an instance of a KS. The attributes of KS and KSA as follows: ksi = {event, pre, exe, rating}; ksai = {rating, exe}. The running process of the CS is as follows: First, enter a series of KS. When an event occurs, the CS determines whether it is consistent with the ksi.event. If they are consistent, the state reads the data of agents from BB and the function ksk.pre is executed. The variable value of ksi.pre is provided by state. If the return value of the function ksi.pre is true, then ksi is triggered and generates an instance ksai, which stores a list of ‘pending-ksas’. Then, ksak.exe in ‘pending-ksas’ with the highest level is executed. The data of the agents on BB are changed. When certain conditions are met, the GBBopen ends its operation.

Autonomous switching of task-level strategies

In this subsection, we have previously developed a framework for ASTS [24]. ASTS can automatically switch strategies according to the different scenarios to react to various emergencies.

Implementation for ASTS.

ASTS is described in the form of tuples, a set of norms N, state S, and event E. (1)

  • Norm N. N = {r1, r2, …} represents a set of norms, which are the mapping relationships between predefined conditions and agent behavior and are the key to the effective operation of ASTS. Norm ri is expressed as ri = {Trii, Acti, Expi}.
    • Trii is the trigger condition and is used to determine if the associated actions can be performed. Trii consists of a series of logical judgment, which can be expressed as .
    • Acti is the execution action. Acti consists of a series of methods or actions, which can be expressed as .
    • Expi is the expected result to verify the result of Acti. Expi consists of a series of expected conditions, which can be expressed as .
  • State S. The data are read by State S. State S represents a set of instantaneous states denoted by S = {s1, s2, …}, which is arranged in chronological order. The instantaneous state si consists of a series of expressions, which are expressed as .
  • Event E. E = {e1, e2, …, ei} represents a set of instantaneous events sorted in chronological order. The instantaneous event ei consists of a series of sub-events, which are expressed as . When an event ei occurs, the instantaneous state sj is read to determine the Trik of rk.

Operation process of ASTS.

Definition 1. Norm, Event, and State. Eq (1) describes the components of ASTS. For example, a logistics truck named Tom starts work at 1o o’clock every day. Tom must follow the norms listed in the Table 1. According to norms ra and rb, Tom maintains a safe distance W from the object in front of it (set at 0.5m). If W is greater than 0.5 m and the power is greater than 30% when e1 occurs, then Tom moves forward. To verify the Acta result of the norm ra, we expect W to be equal to 0.5 m. If W is less than 0.5 m and the power is less than 30% when e1 occurs, Tom stops. The instantaneous events and states are listed in Table 2.

Definition 2. Triggered norm. When ei occurs, the instantaneous state si is read. If all logical judgment expressions in Trix are satisfied, then Norm rx is denoted as the ‘triggered norm’ at instantaneous state si. For example, an instantaneous event e1 occurs at 9:00. Tria is satisfied at the instantaneous state s1; then Tom starts moving forward. Norm ra is denoted as the ‘triggered norm’.

Definition 3. State transition. When ei occurs, norm rx is triggered at si, and Actx is executed. This process causes the instantaneous state to change from si to sj, which can be expressed as: (2) Where the notation ‘→’ denotes an instantaneous state changed from si to sj as a result of triggering rx. If the instantaneous state changed from s1 to s2 as a result of triggering ra. The instantaneous state changes from s2 to s3 as a result of triggering ra. These processes are expressed in Formula (3). (3)

Definition 4. Strategy. Norms are the key to the effective operation of multiple agent systems and are a predefined mapping relationship between conditions and behaviors. A strategy is a plan of action based on the current environment and goals. It is used to guide the agent’s decisions and behaviors in a specific domain and helps the agent achieve its goals in that domain. A strategy consists of a set of action plans, and norms can be used as part of developing a strategy. For example, Norms ra and rb form a goal-following strategy in Table 1. Norm rc forms an obstacle-avoidance strategy.

Definition 5. Switch norms or strategies. When ei occurs, norm rx is triggered at si. Then, ej occurs and ry is triggered at sj. The norm switches from rx to ry. Multiple switching norms accumulate when one norm is switched to another. For example, when e1 occurs, Norm ra is triggered at s1. Then, e2 occurs, and the norm ra is triggered at s2. Finally, e3 occurs, and the norm rb is triggered at s3. The norm switches from ra to ra and eventually to rb.

Based on the above description, this study summarises the pseudocode of ASTS as follows:

  1. i. Input N, S, E, T;
  2. ii. Initialisation: Path = ϕ, ResAct = ϕ, ResExp = ϕ, i = 0, j = 0, k = 0.
  3. iii. If an instantaneous event eiE occurs, the ASTS data are read by the transient state sjS.
  4. iv. Judge Trik of Norm N.
  5. v. If Trik = True, Norm rk is triggered; perform Actk and judge Expk.
  6. vi. The results for Actk and Expk are restored as ResAct and ResExp, respectively.
  7. vii. si, ej, rk, ResAct, ResExp denote the restored paths.
  8. viii. j + +, i + +.
  9. ix. Until i > T, exit ASTS.

Fire emergency reaction dispatching system

We previously developed a FERD system applied to a fire scenario to demonstrate the autonomy and switchability of ASTS [24]. We designed five strategies and two control modes for the FERD system to enable multiple types of firefighting equipment to switch between multiple strategies and multiple control modes depending on the changing environment. The system is described in detail below.

Components of FERD system.

The FERD system system is based on the framework of ASTS. The components include a set of tasks M, a set of fire equipment A, a set of sensors Q, and a set of agent norms N′. FERD system can be expressed as: (4)

  • Task M. The FERD system is applied to handle fire extinguishing tasks, Task M = {m1, m2, …}. Each task mi comprises a series of attributes denoted as .
  • Agent A. A certain number of agents is needed to achieve the ultimate goal, which can be expressed as A = {a1, a2, …}. ai denotes the i-th fire equipment. ai = {ui1, ui2, …, uik} denotes the attribute of ai. uik denotes the k-th attribute of ai.
  • Sensor Q. Each agent carries sensors to sense the environment. Sensor Q = {q1(τ1), q2(τ2), …}, and qi denotes the i-th sensor. The sampling periods of multiple sensors are different. τi is the sampling period of sensor qi. The sensor qi can obtain real values and observed values of the attribute ujk of the fire equipment aj during each sampling period τi. It can be expressed as . denotes real values of sensor qi for measure attribute ujk of fire equipment aj during the first sampling period τi1. denotes observed values of sensor qi for measure attribute ujk of fire equipment aj during the first sampling period τi1. These sensors have different sampling periods. At each time step , observed value of each sensor is obtained in FERD system. This process is shown in Fig 1.
  • Agent Norm N. The norm set N′ for the agent, which is denoted by . Where norm is expressed as . The internal elements of , , and the internal elements of Trik, Actk, Expk are consistent.
thumbnail
Fig 1. Time alignment in FERD system.

The time alignment process of n sensors.

https://doi.org/10.1371/journal.pone.0287791.g001

Objectives of FERD system.

The objectives of FERD system are expressed as Eq (5). Minimize the maximum time for FERD system rescue to complete tasks. And all tasks in M are completed. (5)

Constituent of FERD system.

The FERD system operates based on the GBBopen platform. In a FERD system, a set of multiple KSs make up a strategy. The CS selects the appropriate KS for the multiple agents to switch. After the strategy is executed, the data on the BB are changed until the fire extinguishing tasks are completed. Certain properties of the agent are obtained from external sensors. The components of the FERD system are described in detail below.

  1. (1) Blackboard. In the FERD system, a task and a type of agent are defined as classes, respectively. Each agent is defined as an instance. The attribute value of each agent is either directly or indirectly defined.
  2. (2) Knowledge Source. KS is a core component of the FERD system. In GBBopen, the attributes of KS are in correspondence with the Agent Norm N′, and the properties of the KS are extended, ‘postcondition-function’. The relationship between N′ and KS can be formally described as follows: and ksk.pre, , . KSs are divided into two submodules: the Switch Module and Strategy Module.
    1. 1) Classification of Switch Module. The agent can switch between the strategies based on multiple switch modules. The control modes are divided into Automatic Mode and Command Mode. Multiple agents can switch strategies based on their own identification or environmental identification, or based on user requirements or external guidance.
    2. 2) Classification of Strategy Module. Different strategies are generated from the agent state and the collected environmental information. These mainly include the Following Strategy, Motor Strategy, Evolution Strategy, Obstacle Avoidance Strategy, and Mobilization Strategy.
      1. a. Movement Strategy. The movement strategies are divided into Forward, Stop, and Fallback.
      2. b. Following Strategy.Trajectory Following and Goal Following are developed for agents to arrive at fire scenarios rapidly.
      3. c. Mobilization Strategy. In different scenarios, multiple agents can select Formation, Leave Team, or Return To Team.
      4. d. Obstacle Avoidance Strategy. Multiple agents can select Virtual Circle, Virtual Right Triangle, and Virtual Isosceles Triangle.
      5. e. Evolutionary Strategy. The FERD system provides two evolutionary strategies for autonomously evolving norms: Crossover and Other Method. In Crossover, the expectation of one mutated norm enrich the trigger of another mutated norm.
  3. (3) Control Shell. In the FERD system, after ksak.exe is executed, ksak.post is executed. The CS triggers various norms based on event changes to achieve fast switching strategies.

Coordinate transform

Because the three-dimensional coordinate cannot directly transform the two-dimensional coordinate, it is necessary to obtain the mapping relationship of each coordinate. The transformation between the world coordinate and the camera coordinate is through rigid body transformation. The transformation between the camera coordinate and the image coordinate is through perspective projection. The transformation between the image coordinate and the pixel coordinate is through affine transformation. The transforming formula between a point in the world coordinate and a point in pixel coordinate can be expressed as: (6)

Among them, Zc indicates the distance between the camera and the point in the world coordinate. This distance can be obtained in other ways. K3×4 indicates internal reference matrix. RT4×4 represents external reference matrix. fx, fy indicate focal length. Generally, they are equal. R3×3 represents rotation matrix, which describes the direction between the coordinate axis of a world coordinate and the coordinate axis of the camera coordinate. t3×1 indicates translation matrix, which describes the position of the origin of a world coordinate in the camera coordinate.

Camera calibration

The calibration method of Zhengyou Zhang is to fix the world coordinate system on the chessboard. This world coordinate system is defined in advance. Capture an image of the chessboard through the camera. The pixel coordinates (u, v) of each corner point are obtained using the image detection algorithm. The corner point is the connection point of the contour line of the object. The coordinate of each corner point (Xw, Yw, Zw=0) is calculated in the world coordinate system. Through Eq 6, the mapping relationship between the world coordinate and pixel coordinate is calculated. The internal parameter matrix (e.g. K3×4) and external parameter matrix (e.g. R3×3, t3×1) are obtained.

Monocular measurement

The measurement methods based on monocular vision can be divided into multiple categories. One of the generalizations is the ratio-based approach. The principle is that the distance is inversely proportional to the image size. The calculation formula is: (7)

Among them, f represents Focal length; P represents Pixel width of an object, pixel; W expresses Actual width of an object, m; D′ represents that unknown distance from an object to a camera, m.

Kalman filter based confidence degree

Kalman Filter is a data fusion algorithm. This algorithm fuses the measured values of different sensors with the same object. The measurement noise satisfies a normal distribution. Finally, a more accurate value is obtained. To study the consistency of sensor data, the concept of confidence degree is introduced here.

Suppose there exist n independent sensors. There are q1, q2, …, qn. The fusion process of the Kalman Filter based on confidence degree is as follows. (8)

Among them, K indicates Kalman gain. The degree of change in uncertainty after each fusion of data. A lower Kalman gain helps to improve the accuracy of fusion and vice versa. Ri represents measurement error of the i-th sensor. R0 represents measurement error of all sensors. indicates observation data of the i-th sensor. indicates equation of state of the i-th sensor. indicates the optimal estimate of the i-th sensor. It is obtained by classical Kalman filtering. At time t, represents state transfer matrix of the i-th sensor. indicates control matrix of the i-th sensor. represents control vector of the i-th sensor. indicates standard deviation of the i-th sensor. indicates confidence degree of the i-th sensor. represents observations based confidence degree. indicates predicted values based confidence degree. indicates standard deviation based confidence degree. represents fusion results of multiple sensors. indicates standard deviation of multiple sensors.

As can be seen from Eq 8, the effective fusion of multiple sensors is achieved. This process adjusts the confidence degree of each sensor in real time. When the observed values of a sensor is not accurate, the algorithm has a certain fault-tolerant ability. This algorithm affects the overall performance of multi-sensor fusion. In this paper, the Kalman Filter based confidence degree has the ability to be adaptive.

Proposed module

In this paper, for the phenomenon of inaccurate observed values of multiple sensors, Fusion Module is added in FERD system. This section provides a detailed description of the supplement or correct of a single sensor.

Fusion module of FERD system

This study added the fusion module in FERD system. KS consists of three sub-modules. They are the control module, execution module, and fusion module.

KS of fusion module

This subsection provides a detailed description of Fusion Module. Experts have developed a series of fusion strategies for the fusion of multiple sensors. As shown in Fig 2.

thumbnail
Fig 2. Fusion module in FERD system.

The Fusion Module include denoising, spatial alignment, confidence degree update, observation filtering, data fusion, and fusion decision.

https://doi.org/10.1371/journal.pone.0287791.g002

Denoising.

Interference values exist for every sensor. Interference values are values that describes the scenario inaccurately. In this paper, interference values are collectively referred to as random errors. Then the observed values contains a large number of random errors. The random errors not only increase the computer memory and computation overhead but also increase the computation error. So observed values of each sensor need denoising. The purpose of denoising for sensor qi is to reduce the Standard Deviation σi.

Spatial alignment.

The observed value of each sensor in the same platform using different coordinate systems. The observed value of different sensors involve different coordinate systems. Therefore, observed value must be converted into data in the same coordinate system before fusion. In fact, this process is to calculate the transformation relationship between coordinate systems. The system converts the attribute ujk of the fire equipment aj into a uniform type of sampled data from sensor qi by Spatially Aligning KS. In this paper, the transformation matrix of sensor qi is set to transi. The observation for each sampling period after spatial alignment is . The real value of the sampling period after spatial alignment is .

Confidence degree update KS.

The contribution of each sensor to data fusion is also different. Therefore, this paper proposes the concept of confidence degree. The fusion contribution of each sensor is defined by the confidence degree. For example, sensor qm and qn are both sensors that make observations of ujk. In each sampling period, their confidence degrees are updated by this KS. Their confidence degrees are denoted by wmg and wng, respectively. Where .

Observation filtering.

The error of observed values from each sensor is different. Therefore, the concept of error threshold is proposed in this paper. It serves as a precondition for fusion. The error of observed values of sensor qi on ujk in each sampling period is expressed as follows. (9)

For example, sensor qm and qn measure for ujk. Then, this paper calculates the error between observed values and real values. Their errors are denoted as εmg and εng, respectively. Select their error thresholds by this KS. Their error thresholds are denoted as εm and εn, respectively. Finally, observed values of two sensors in each sampling period are compared with the corresponding error thresholds separately. The filtering operation is performed mainly for Data Fusion KS.

Data fusion.

This KS is a process of multi-level, multi-spatial information complement and optimal combination processing by multiple sensors. In this process, multiple sources of data are fully utilized for rationalization and use. This not only takes advantage of multiple sensors operating in concert with each other but also integrates data from other information sources. This improves the intelligence of the whole system, as well as the accuracy and comprehensiveness of the information, and reduces the uncertainty of the information. This KS combines observations from multiple sensors into one data. This operation makes the measured values more accurate.

This paper assumes that there are Γ different fusion algorithms in this module. The observed values of sensor qm and qn for the same attributes (ujk) are fused. At this point, the observed values from both sensors have been denoised and spatially aligned. During each sampling period , FERD system can obtain observed values of all sensors. ∃εmgεm. Where .

By the fusion algorithm γ ∈ Γ, the result of Data Fusion KS during each sampling period can be expressed as . Then, (10)

Among them, the symbol ‘⦾’ is a symbol for data fusion, which indicates that two observed values are fused. The real values of the same property for the same object is the same for different sensors. The real values of the same property of the same object in each sampling period can be expressed as .

Fusion decision.

For the results of different data fusion algorithms, the role of this KS is to select the appropriate fusion algorithm γ. This algorithm is a fusion algorithm that is the closest to the real values. That is: (11) Where, j = 1, 2, …, x; k = 1, 2, …, y′; γ ∈ Γ.

Results

This experiment of performing firefighting tasks in an indoor scenario with multiple obstacles is a specific application scenario of this fusion module proposed. FERD system can use multiple sensors to detect unmanned vehicles during firefighting tasks. To verify the effectiveness of Fusion Module, three sensors are used to assist in the firefighting tasks. This section first introduces experiment design. Then, the experimental steps are described in detail. Subsequently, the experimental results are given. Finally, the effectiveness is further verified by visual analysis.

Experiment design

In this subsection, the experiment design is first described. And the details are described. Then, the limitations are considered and analyzed. Finally, evaluation criteria are given to verify the effectiveness of this fusion module. Table 3 summarizes the symbols commonly used in this experiment.

Experiment scenario.

This experimental sets the experimental scenario as a rectangular environment, as shown in Fig 3. This scenario is a rectangular base station laid by four UWB anchors located at the same horizontal position. One of anchors is used as the origin of rectangular.

thumbnail
Fig 3. Experiment scenario.

The size of this experiment scenario is 15m × 2.4m. Multiple obstacles are also placed within this experiment scenario.

https://doi.org/10.1371/journal.pone.0287791.g003

Experiment platform and equipment.

The experiments in this paper are conducted By taking the Chengdu Kestrel Artificial Intelligence Institute’s Interface Description Language(IDL)-Mapping tool (KIS-CORBA) as a cross-language data communication protocol, acquiring real-time information with multiple sensors, and publishing or subscribing to the information through the serial port of the Robot Operating System (ROS) to finally implement GBBopen to control firefighting equipment. This IDL is the IDL of Object Management Group (OMG).

1) Experiment Platform. All experiments require a PC platform and an ARM-embedded platform. The ARM-embedded platform used in this experiment is the Raspberry Pi (RPI). The Python environment (3.7.0), KIS-CORBA, and ROS platforms are installed into the RPI. The Visual Studio environment (VS2017), KIS-CORBA, Lisp environment (Allegro Common Lisp 10.1 express), LinkTrack technology (NoopLoop for NAssistant applications), and GBBopen are installed on the PC.

2) Experiment Equipment. This experiment uses an unmanned vehicle (RoboMaster) as the firefighting equipment for the FERD system, as shown in Fig 4.

thumbnail
Fig 4. Firefighting equipment.

An unmanned vehicle (RoboMaster).

https://doi.org/10.1371/journal.pone.0287791.g004

i. Sensors. Each piece of unmanned vehicle carried five items, including an IDS, a camera, an UWB anchor, a module of an IMU, and a RPI, as shown in Fig 4.

ii. Application Programming Interface (API). DJI-Innovations (DJI) provides several API instances (https://robomaster-dev.readthedocs.io/zh_CN/latest/) written in Python. This experiment is conducted to realize communication between the unmanned vehicle and GBBopen and to control the wheels of the unmanned vehicle. Because unmanned vehicle is written in Python and GBBopen is written in Lisp, cross-language communication is required. These experiments focused on the communication between Python and Lisp and finally applied the FERD system.

iii. Attributes. Various attributes exist for each firefighting task and each unmanned vehicle. These attributes can be expressed as follows: (12) (13)

iv. Data transmission. The remote server is connected to control the unmanned vehicle through Secure Shell (SSH). The position of the unmanned vehicle, the three-axis attitude angle, and the distance from the front are obtained through UWB, IMU, and IDS, respectively. Among them, UWB achieves high-precision indoor positioning by sending ultra-short pulse signals that exploit multipath propagation and the Doppler effect in indoor environments. The unmanned vehicle obtains images through the camera’s API and embeds a coordinate transformation program to output coordinates. The ROS serial port then reads the obtained data and publishes it in the form of topics. Finally, they are transmitted to GBBopen through KIS-CORBA. The data of the unmanned vehicle is sent to GBBopen, including pose, velocity, distance, etc. GBBopen sends the data to the unmanned vehicle, such as target point, speed, etc.

v. Pose information. Positioning errors occur due to various factors (e.g., the deployment mode of the base station, obstracles and the weakening of the signal) in the indoor environment. Therefore, the experiment assumes that the information measured by the UWB and IMU is the actual location information of the unmanned vehicle. In Fig 3, the number of unmanned vehicles is 2. The position of two unmanned vehicles are (10.8, 1.8) and (15,0.6), respectively. And the position of the firefighting task is (4.2, 0.6).

Experiment contents.

In FERD system, the fusion norms are designed, which guide each unmanned vehicle to perform tasks and the fusion positioning of multiple sensors. Briefly introduce the 22 norms that already exist within FERD system. Norm r1 creates multiple instances of the unmanned vehicle. Norm r2 and r3 are rules for setting the switching ways and sorting the fire locations. The switching ways include Command Mode and Automatic Mode. Norm r4r8 are rules about the following strategy. Among them, Norm r4 and r5 are the ways to determine the following of the unmanned vehicle ai. Norms r7 and r8 are parameters to set the unmanned vehicle ai following mode. Norm r9r13 are rules about the Mobilization Strategy. These strategies include Formation, Departure, and Return To the Team. Norm r12 and r13 are rules for setting the role of the unmanned vehicle ai. Norm r14r16 are three rules about the Motion Strategy. Here it is recommended that the linear velocity of the unmanned vehicle is between 0.05 ∼ 0.5m/s. The angular velocity of the unmanned vehicle is between 0 ∼ 90rad/s. Norm r17r19 are the three rules of the Obstacle Avoidance Strategy. Norm r20r22 are rules about Evolutionary Strategy. Table 4 shows the added norms and revised norms within FERD system in this paper. Norm r14r16 are rules about three revisions of the Motion Strategy. Norm r23r30 are rules about the Following Strategy. Norm r31r38 are rules about updating confidence degree and filtering of observations. Norm r39r40 are rules about different algorithms about the fusion of multiple sensors. For example, the fusion algorithm of the Kalman Filter based confidence degree and the weighted average-based fusion algorithm are applied in this paper. Norm r41 is the rule for fusion decision-making. This norm is that the fusion results from different algorithms are compared with gtj. The optimal fusion result is selected.

In indoor scenarios with multiple obstacles, each unmanned vehicle chooses the appropriate norms based on its state and assists in firefighting tasks through the sensors it carries itself. This experiment sets a fixed route for two unmanned vehicles. They pass through (4.8,1.8), (4.8,0.6), and (4.2,0.6) in turn. Eventually reaching the fire location (4.2, 1.2), they perform the fire extinguishing task.

Experiment limitations.

The experiment design can verify the effectiveness of the proposed module, but this module has certain limitations (e.g., timeliness). In this experiment, the relative velocity between two unmanned vehicles is set to check the real-time performance. Specifically, the relative velocity of two unmanned vehicles is used as an independent variable. The fusion results of multiple sensors are used as the dependent variable. Finally, the fusion results of multiple sensors are compared with the real values. This experiment sets different relative velocities between two unmanned vehicles to perform the fire extinguishing task. The relative velocities in this experiment are 0.03m/s, 0.06m/s, 0.09m/s, 0.12m/s, 0.15m/s, and 0.2m/s, respectively.

Evaluation metrics.

In this experiment, the FER is proposed as the fusion index of this fusion module. GBBopen receives the number of localization coordinates of all sensors simultaneously, expressed by Z. Among them, the distance between the coordinates localized by UWB and the ground truth is large; while the distance between the coordinates localized by other sensors and the ground truth is small. The distance between the coordinates localized by other sensors and the ground truth is small. The number of coordinates of UWB localization is larger than the distance to the ground truth, denoted by z. The FER is expressed as × 100%. With effectively fused data, one evaluation index is the error range of ED and MSE between observed values and the ground truth of a single sensor. Another evaluation index is the error range of ED and MSE between the fusion results of multiple sensors and their ground truth.

Steps

First, this experiment analyzes the different sensors in different scenarios. Next, the important parameters inside the norms are trained in indoor scenario with multiple obstacles. Finally, the two unmanned vehicles follow a set fixed route and assist in firefighting tasks.

Observed values of different sensors in different scenarios.

This step analyzes the different sensors in different scenarios. And the observed values are compared with the real values.

1) UWB Positioning. A unmanned vehicle a1 carries a UWB anchor to position itself. a1 starts with the real position (7.8,1.8). a1 is moved 10 times from right to left at 600mm intervals in sequence. a1 ends at the real position (2.4,1.82). The real positions of a1 are (7.8,1.8), (7.2,1.8), (6.6,1.8), (6,1.8), (5.4,1.8), (5.4,1.8) (4.8,1.8), (4.2,1.8), (3.6,1.8), (3,1.78), (2.4,1.82). Finally, the values of the positioning are compared with the real values.

2) IDS and Camera Ranging. First, the mapping relationship between world coordinates and pixel coordinates is obtained by Formula 6. The internal and external reference matrices of the camera are obtained by camera calibration. Then, a1 carries Camera and IDS to range unmanned vehicle a2. a2 moves 10 times in sequence at 600mm intervals from 600∼6000mm. The distances between a1 and a2 are 600mm, 1200mm, 1800 mm, 2400mm, 3000mm, 3600mm, 4200mm, 4800mm, 5400mm, and 6000mm respectively. Finally, the values of the ranging are compared with the real values.

3) UWB, IDS and Camera Positioning. a1 carries Camera and IDS to range a2. Then, two ranging results are fused. Next, a1 was positioned using both UWB and Camera. a2 robot positioning starts with the real position (7.8,1.8). a2 is moved 10 times from right to left at 600mm intervals in sequence. a2 ends at the real position (2.4,1.82). The real positions of a2 are (7.8,1.8), (7.2,1.8), (6.6,1.8), (6,1.8), (5.4,1.8), (5.4,1.8) (4.8,1.8), (4.2,1.8), (3.6,1.8), (3,1.78), (2.4,1.82). Finally, the values of the positioning are compared with the real values.

Train error thresholds and confidence degrees.

The accuracy and error of the observed values of each sensor are different. Each sensor also contributes differently to the fusion of multiple sensors. Therefore, some parameters of some norms are trained. They include the confidence degree of each sensor and the error threshold of all sensors. In this experiment, the number of training is 10. In the initial stage of training, the confidence degree of each sensor is equal. And the error threshold of all sensors is set to null. The results at the end of the first training are the initial values for the second training. Until the end of the tenth training session, the trained results are the initial values of the unmanned vehicles performing tasks.

Performing tasks.

The two unmanned vehicles follow a set fixed route and assist in firefighting tasks through the parameters of the norms that have been trained and the sensors that each carries.

Results and analysis

The results of the experiment steps are given in this paper. Then, the experiment results are analyzed.

Observed values of different sensors in different scenarios.

The results of the observed values and true values are analyzed by different perspectives (quantitative and qualitative) in different scenarios.

1) UWB Positioning. Figs 5 and 6 show the results of the observed values and real values of unmanned vehicle a1 analyzed by qualitative. Horizontal axis indicates the length of the rectangle. Vertical axis indicates the width of the rectangle. The positioning of a1 are compared 10 times in order from right to left. The red line indicates the real values of a1. The purple line indicates the observed value of UWB. Table 5 shows the results of the observed values and true values of unmanned vehicle a1 analyzed by quantification. Vertical indicates the number of experiments. Horizontal indicates the data analysis from different scenarios.

thumbnail
Fig 5. In indoor scenario without no obstacle.

Comparison of UWB position and real values in different scenarios.

https://doi.org/10.1371/journal.pone.0287791.g005

thumbnail
Fig 6. In indoor scenario with multiple obstacles.

Comparison of UWB position and real values in different scenarios.

https://doi.org/10.1371/journal.pone.0287791.g006

thumbnail
Table 5. Data analysis of between observed values of UWB position and real values.

https://doi.org/10.1371/journal.pone.0287791.t005

In Fig 5, the red line is smoother, while the purple line has fluctuations. So, there is instability in the observed values of UWB. In Fig 6, the red line is smoother. the purple line fluctuates more. So, there is instability in the observed values of UWB. And it is less stable than in an indoor scenario without obstacles. In Table 5, the quantitative results are analyzed as follows. In an indoor scenario without obstacles, the error range on the x-axis, y-axis, and European distance is -0.34∼-0.08 m, -0.15∼0.53 m, 0.01∼0.29 m. The SD on the x-axis and y-axis ranges from 0.0∼0.03 m, and 0.03∼0.10 m, respectively. The range of MSE on the x-axis and y-axis is 0.01∼0.12m2 and 0.01∼0.28m2, respectively. In an indoor scenario with multiple obstacles, the error range on the x-axis, y-axis, and European distance is -0.49∼-0.12 m, -2.35∼0.71 m, 0.05∼5.62 m. The SD on the x-axis and y-axis ranges from 0.01∼0.13 m, and 0.06∼0.90 m, respectively. The range of MSE on the x-axis and y-axis is 0.01∼0.24m2 and 0.02∼6.34m2, respectively. This quantitative result illustrates that the observed values of UWB fluctuate more on the y-axis than on the x-axis. The UWB observations fluctuate more in an indoor scenario with multiple obstacles than without obstacles. This is due to the fact that the accuracy of the short edge (y-axis) is worse than the long edge (x-axis) in the unobstructed case. The impact on the ranging accuracy in an indoor scenario with multiple obstacles and the impact of the ranging algorithm can cause the short edge to be affected more than the long edge.

2) IDS and Camera Ranging. Figs 7 and 8 show the results of the observed values and real values analyzed by qualitative. The real values are distance between a1 and a2. The observed values are a1 ranging a2 through the carried IDS and the Camera. Horizontal axis indicates the number of experiments. Vertical axis indicates the observed values of ranging. The red line indicates real values. The purple line indicates that observed values are a1 ranging a2 through the carried IDS. The green line indicates that observed values are a1 ranging a2 through the carried the Camera. Table 6 shows the results of the observed values and real values analyzed by quantification. Vertical indicates the real values. Horizontal indicates the error between observed values and real values by different sensors in different scenarios.

thumbnail
Fig 7. In indoor scenario without no obstacle.

Comparison of observed values and real values of ranging by different sensors in different scenarios.

https://doi.org/10.1371/journal.pone.0287791.g007

thumbnail
Fig 8. In indoor scenario with multiple obstacles.

Comparison of observed values and real values of ranging by different sensors in different scenarios.

https://doi.org/10.1371/journal.pone.0287791.g008

In Figs 7 and 8, there is no data for ranging by IDS in 4800∼6000mm, while the data is available for ranging by Camera. The experimental results show that ranging by IDS does not work, while it works by Camera in 4800∼6000mm. The experimental results show that observed values of ranging by IDS are smaller than the real values. Table 6 is analyzed from the quantitative perspective as follows. In an indoor scenario without obstacles, the error range between ranging by IDS and real values is 134.86∼420.35mm. The SD of these errors is 111.70. The range of error between ranging by Camera and real values is -1334.713∼489.13mm. The SD of these errors is 489.13. The SD of the Camera is 4.38 times higher than the IDS. In an indoor scenario with multiple obstacles, the error range between ranging by IDS and real values is 115.74∼647.74. The SD of these errors is 202.34. The range of error between ranging Camera and real values is -1154.43∼268.55mm. The SD of these errors is 529.73. The SD of ranging by Camera is 2.62 times higher than ranging by IDS. These results show that ranging by IDS and Camera has almost no effect in different scenarios. However, ranging by IDS is more stable than ranging by Camera. This is due to the detection of the target by a Camera that has errors (i.e., errors in the pixel width), resulting in errors in the numerator of Formula (7).

3) UWB, IDS and Camera Positioning. Figs 9 and 10 show the results of the observed values and real values analyzed by qualitative. Horizontal axis indicates the length of the rectangle. Vertical axis indicates the width of the rectangle. The unmanned vehicle a2 is compared ten times in order from right to left. The red line indicates the real values of a2. The pink line indicates the observed values of a1 by UWB to position itself. The purple line indicates that a1 by IDS to range a2. At the same time, the observed values of a1 by Camera to position a2. The green line indicates the monocular range results of a1 by Camera to a2. Also, the observed values of a1 by Camera to position a2.

thumbnail
Fig 9. In indoor scenario without no obstacle.

Comparison of observed values and real values in different scenarios.

https://doi.org/10.1371/journal.pone.0287791.g009

thumbnail
Fig 10. In indoor scenario with multiple obstacles.

Comparison of observed values and real values in different scenarios.

https://doi.org/10.1371/journal.pone.0287791.g010

Tables 7 and 8 shows the data analysis between the observed values and real values by different sensors in different scenarios. In these table, vertical represents real values, horizontal indicates the error between observed values and real values by different sensors.

thumbnail
Table 7. Data analysis of observed values and real values by different sensors in indoor scenario without obstacles.

https://doi.org/10.1371/journal.pone.0287791.t007

thumbnail
Table 8. Data analysis of observed values and real values by different sensors in indoor scenario with multiple obstacles.

https://doi.org/10.1371/journal.pone.0287791.t008

In Figs 9 and 10, the trend of red line is flat, while the trends of pink, purple, and green lines are fluctuation. The trend of pink line are greater than purple and green lines. When the real coordinates of a2 (8.4,1.8) and a1 are (3.6,1.8), (3,1.78) and (2.4,1.82) in that order, there is no data for ranging by IDS. So no observed values are Camera position (IDS ranging). While the data is available for ranging by Camera, so observed values are Camera position (Camera ranging). In Tables 7 and 8, this experiment compares the observed values of a1 by different sensors with the real values. In indoor scenario without obstacles, the error ranges of a1 by UWB position itself on the x-axis, y-axis, and ED are -0.34∼-0.08 m, -0.15∼0.53 m, and 0.01∼0.29 m, respectively. The range error of a1 by Camera position (IDS ranging) position a2 on the x-axis, y-axis, and ED is -0.54∼-0.27 m, -0.21∼-0.03 m, 0.07∼0.33 m, respectively. The range error of a1 by Camera position (Camera ranging) position a2 on the x-axis, y-axis, and ED is -0.29∼1.17 m, -0.21∼-0.04 m, 0.02∼1.38 m, respectively. In indoor scenario with multiple obstacles, The range error a1 by UWB position itself on the x-axis, y-axis, and ED is -0.49∼-0.12 m, -2.35∼0.71 m, 0.05∼5.62 m, respectively. The range error of a1 by Camera position (IDS ranging) position a2 on the x-axis, y-axis, and ED is -0.68∼-0.12 m, 0.21∼0.45 m, 0.13∼0.53 m, respectively. The range error of a1 by Camera position (Camera ranging) position a2 on the x-axis, y-axis, and ED is -0.36∼1.15 m, 0.18∼0.45 m, 0.09∼1.35 m, respectively.

From the above qualitative and quantitative analysis results, it is clear that the stability of UWB observations varies from scenario to scenario. This experiment allows the IDS and Camera to complement or fuse each other for ranging the unmanned vehicle. When the Camera (IDS ranging) measurement fails, Camera position (Camera ranging) and UWB can complement each other.

Train error thresholds and confidence degrees.

Figs 11 and 12 show the change curve of training confidence degree and error threshold.

thumbnail
Fig 11. Training confidence.

Training variation curves of confidence degree. The different color lines show the change training trend of different sensors’ confidence degrees. The pink line shows the change training trend of UWB positioning. The purple line shows the change training trend of Camera position (IDS ranging). The green line shows the change training trend of Camera position (Camera ranging). The yellow line indicates the change training trend of Camera position (Camera and IDS fusion ranging).

https://doi.org/10.1371/journal.pone.0287791.g011

thumbnail
Fig 12. Training error threshold.

Training variation curves of error thresholds.

https://doi.org/10.1371/journal.pone.0287791.g012

In Fig 11, the four different channels have the same confidence degree of 0.25 in the initial stage of training. The confidence degrees of the four different channels gradually differ as the number of training sessions increases. The confidence degree of the Camera position (IDS ranging) shows a fluctuating trend of rapid rise. The confidence degree of UWB’s position shows a fluctuating trend of slow rise. The confidence degree of the Camera position (Camera ranging) shows a slow downward trend. The confidence degree of the Camera position (Camera and IDS fusion ranging) shows a rapid downward trend. After training, their confidence degrees are as follows. Camera position (IDS ranging) has the highest confidence degree of 0.52. This is followed by the confidence degree of UWB position with 0.36. The confidence degree of Camera position (Camera and IDS fusion ranging) was 0.12. The lowest confidence degree of Camera position (Camera ranging) was 0.04. Fig 12 indicates the training trend of the error threshold for all sensors. At the initial stage of training, the error threshold of all sensors is 0 m. In the first training, the training error threshold of all sensors fluctuates greatly. The maximum is 0.50m, and the minimum is 0.35m. With the increase in training times, the fluctuation amplitude of all sensor error thresholds tends to be stable. The stable value reaches 0.47m. After training, the error threshold of all sensors reaches 0.46m.

Timeliness.

Table 9 shows the timeliness results of the fusion module. The table indicates the relative speeds of the two unmanned vehicles in the horizontal direction. The longitudinal direction indicates whether the position after fusion is lagged and the lag value. It is obvious from the table that there is no lag in the fusion results of multiple sensors when the relative speeds between a1 and a2 are set at 0.03m/s, 0.06m/s, 0.09m/s, and 0.12m/s. While the relative speed between a1 and a2 is 0.15m/s and 0.2m/s, the fusion results of multiple sensors show lag. At a relative velocity of 0.15m/s, the fusion results differ from the real values by 0.3m. At a relative velocity of 0.2m/s, the fusion results differ from the real values by 2m. From the analysis in the table, it is clear that the relative velocity is within 0.12m/s to ensure the validity of the fusion results of multiple sensors.

Performing tasks.

Figs 1315 shows the visualization results of the FERD system performing task. a1 and a2 perform firefighting tasks in an indoor scenario with multiple obstacles. Horizontal axis indicates the length of the rectangle. Vertical axis indicates the width of the rectangle.

thumbnail
Fig 13. Real trajectory of a1 and a2.

This figure shows the real trajectory of a1 and a2 are perform a task.

https://doi.org/10.1371/journal.pone.0287791.g013

thumbnail
Fig 14. Positioning trajectory of a1 and a2 by UWB.

The red and blue lines indicate the positioning trajectory of a1 and a2, respectively. The red line is available at https://doi.org/10.5061/dryad.31zcrjdrt (DOI: 10.5061/dryad.31zcrjdrt).

https://doi.org/10.1371/journal.pone.0287791.g014

thumbnail
Fig 15. Comparison between UWB positioning and fusion positioning of a1.

https://doi.org/10.1371/journal.pone.0287791.g015

In Fig 13, a1 follow a fixed route. It starts at (10.8,1.8) and passes through J1, J2 and J3 in sequence. Finally, it reaches fire T (4.2,1.2) to extinguish the fire. In fact, it is unavoidable that a1’s wheels have a slight deviation when moving forward. As a1 moves forward, simultaneously and quickly adjust its deviation, so that a1 does not differ too much from this route. It is equivalent to a1 moving along a track. Therefore, the real trajectory of a1 coincides with the fixed route. The red line indicates the real trajectory of a1. a2 starts from (15,0.6). a2 passes through A′, B′, C′, and D′ in that order. a2 finally reaches (4.05,1.08). The blue line indicates the real trajectory of a2. The trajectories of a1 and a2 are from right to left. The yellow line indicates the a2 by multiple sensors to observe a1. The yellow line is also a fusion trajectory fused of a1.

However, the positioning trajectory obtained by UWB is not consistent with the real trajectory in Figs 13 and 14. This is due to the fact that the experiment is conducted in an indoor environment with many obstacles. The accuracy of UWB positioning is affected by signal interference, reflections, and other factors.

Fig 15 shows the valid coordinates of a1 localized by different sensors. The figure includes the visualization results of single-sensor and multi-sensor fusion. The red dots are the coordinates of a1 localized by the UWB carried by a1. Purple dots are the coordinates localized by the fusion module for a1. Specifically, a1 is fused through the positioning coordinates of a single sensor and a2 through multiple sensors observing the positioning coordinates of a1. In addition, the Fig 15 is somewhat different from the result of locating a1 by fusion in Fig 14. The trajectory of a1 by UWB localization in Fig 14 is a continuous line, as shown in the red line in Fig 14. While it is a discontinuous point, as shown in the red dots in Fig 15. This is due to the fact that the coordinates obtained by GBBopen are different from the frequency of the localization coordinates obtained by UWB. Among them, UWB obtains the positioning coordinates more frequently than the coordinates obtained by GBBopen. And the data obtained by GBBopen is also susceptible to the influence of net clusters.

In this paper, Fig 13 is analyzed from different perspectives. First, compare the distances of a1 and a2 to J1. It can be seen that the ED from a1 to J1 is shorter than a2 to J1. From the above analysis, a1 is the leader and a2 is the follower. From the figure, it can be seen that a1 gradually approaches J1. According to the Norm r4, then a2 uses the goal following approach toward a1. In this figure, a2 by multiple sensors can observe a1 when a1 has traveled to A. Meanwhile, a2 has traveled to A′. According to the Norm r25r27, so a2 takes a goal following towards a1. Until a2 by multiple sensors do not observe a1, at this point a1 travels to B. Meanwhile, a2 has traveled to B′. Since both a1 and a2 are in the execution task area and a2 by multiple sensors do not observe a1. According to Norm r5, then a2 takes a trajectory following towards a1. a2 travels from B′ to C′. Similarly, a2 by multiple sensors to observe a1 when a1 has traveled to C. Meanwhile, a2 has traveled to C′. According to the Norm r25r27, so a2 takes a goal following towards a1. Until a2 by multiple sensors do not observe a1, at which point a1 travels to D. a2 has traveled to D′. Since both a1 and a2 are in the execution task area and a2 by multiple sensors do not observe a1. According to Norm r5, then a2 takes a trajectory following towards a1. a2 starts traveling from D′. Until both a1 and a2 complete their firefighting tasks.

Evaluation.

This paper statistics the raw data acquired by GBBopen. GBBopen receives the number of localization coordinates of all sensors simultaneously is 4070. The number of coordinates of UWB localization that are larger than the distance to the ground truth is 3608. So FER is calculated as 11.35%.

Table 10 shows compare results of the range of error and MSE between single sensor and multi-sensor fusion. The range of error of ED between the single sensor observation and the true value is 0.80∼1.47 m. The error range of ED between the observed and true values after multi-sensor fusion is 0.59∼1.02 m. The Min of the error of the ED after multi-sensor fusion increased by 26.25% over the single-sensor observations. The Max of the error of the ED after multi-sensor fusion decreased by 30.61% over the single-sensor observations. The range of MSE between the single sensor and real values of the a1 after multi-sensor fusion is 1.06∼1.35m2. The MSE of the multi-sensor fusion results over the real value of a1 is 0.66∼0.86m2. The Min of the MSE after multi-sensor fusion is 37.74% higher than that of the single-sensor observations. The Max of the MSE after multi-sensor fusion is increased by 36.30% compared to that of the single-sensor observation.

thumbnail
Table 10. Compare results between single sensor and multi-sensor fusion.

https://doi.org/10.1371/journal.pone.0287791.t010

The analysis of the above experiment results shows that the fusion of multiple sensors is achieved in this paper. This fusion module ensures precise positioning between adjacent unmanned vehicles using multiple sensors. The unmanned vehicles are coordinated to accomplish the firefighting task. This fusion module reduces the impact of FERD system in open environments. This module also improves the environmental awareness of each unmanned vehicle.

Discussion

Norms are designed by experts in multiple fields. Each expert can describe different fields according to his experience. This fusion module involves multiple domains. Therefore, experts construct complex fusion strategies in a data processing module by combining multiple norms. When the measured data is unstable, this fusion module can reflect a certain degree of robustness. For example, if a sensor is not working or information is confusing, the FERD system can continue to perform firefighting tasks. The FERD system demonstrates its scalability when new needs arise. For instance, this paper adds new norms within this fusion module. These norms can meet the needs of subsequent environments, and long-lasting applications.

This fusion module proposed in this paper suffers from certain limitations. This paper discuss the uncertainties and limitations of fusion module in terms of the delay of the data, the suitability of the sensors, and the protection measures of the sensors.

  • Data latency. The data acquisition is performed in FERD system with ROS as the medium. The first consideration is how to ensure real-time data fusion. The involved delay includes transmission delay, propagation delay, processing delay, etc. So it is more demanding on the network. FERD system is an event-trigger mechanism. It is more important the specific events monitored change at any time. This mechanism has strong random characteristics. This mechanism cannot predict the target object. In some respects, this may limit the open system. However, in a dynamic environment, this is more generic for fusion methods. This generalization is the fusion of data by norms.
  • Sensor selection. Fire environments have special properties, such as high temperatures and dense smoke. These special properties can have an impact on the performance and reliability of the sensor. The range of applications for a sensor can also affect its performance and reliability in a fire environment. Namely, some sensors may only be suitable for specific types of fires and may not work properly for other types of fires. Therefore, the range of application of sensors needs to be considered when applying them to avoid limitations and misjudgments. The sensors covered in this paper are a range of DJI-Innovations RoboMaster (EP) products, including an infrared depth sensor (IDS) and a camera. IDS has an operating temperature range of -10 to 40 degrees Celsius. It is possible for the range accuracy to be reduced or even for the measurement to fail in special weather or environments, such as dense smoke, rain, fog, or direct sunlight. The camera needs to ensure that it is not disturbed and is unobstructed. There are video streams and images in the thick smoke. Its operating temperature range is -10 to 40 degrees Celsius. Extreme environments are not considered in this paper. For instance, the fire environment when studying evacuation strategies for large public buildings [25]. Therefore, the sensors involved cannot face complex high-temperature or dense smoke environments and are only suitable for mild fire situations. This paper also proposes some methods to improve the data collection capability of the FERD system and ultimately improve fusion efficiency. For example, sensors that can resist harsh environments could replace the above sensors to further improve their performance and stability.
  • Sensor protection. Protective measures for sensors in a fire environment are also very important. For example, packaging and encapsulation with fire-resistant materials are required to protect the sensors from high temperatures and smoke. Moreover, the protection level of the sensor and the match between the protection level and the application environment need to be considered to ensure the performance and reliability of the sensor.

At present, many possible applications are being considered to test the feasibility of this fusion module. To move from the current system to the actual application, it is also necessary to add some functions reflecting the real situation in these norms (e.g. the important parameters of some norms). In theory, these parameters use constant values. In application, these parameters use empirical values. These experience values are dynamically adjustable. Also, the parameters can be affected by random, system noise, due to larger environmental factors.

Conclusion

This paper proposes a fusion module in the FERD system based on Blackboard Architecture. The main goal of this fusion module is to overcome the inaccuracy of a single sensor in indoor scenarios with multiple obstacles and improve measurement accuracy. The module utilizes multiple sensors to complement or correct the positioning of each unmanned vehicle. Specifically, this module develops a variety of fusion-related processing techniques for precise positioning. This paper then proceeds to gain experience data on the confidence degree, error of different sensors, and timeliness of this module by training in an indoor scenario with multiple obstacles. Compared with a single sensor, the module proposed in this paper is switchable and scalable. Finally, the performance of the fusion module was evaluated to demonstrate the effectiveness of localization based on the field sensors. Some important applications are target identification and target tracking tasks for multiple agents. These areas mainly include identifying the motion intention of the target, local path planning, etc. For example, when a sensor has short-term abnormal data, the system determines the abnormal conditions based on the status of the abnormal data and automatically filters this abnormal data. Then this system can still work normally without affecting subsequent operations. In the future, we will apply the proposed fusion module to other applications. The application of autonomous driving systems, for example, can improve safety.

References

  1. 1. Pastor Francisco and Ruiz-Ruiz Francisco J. and Gómez de Gabriel Jesús M. and García-Cerezo Alfonso J. Autonomous Wristband Placement in a Moving Hand for Victims in Search and Rescue Scenarios With a Mobile Manipulator. IEEE Robotics and Automation Letters. 2022;7(4):11871–11878.
  2. 2. Rihab Lahouli and Muhammad Hafeez Chaudhary and Sanjoy Basak and Bart Scheers. Tracking of Rescue Workers in Harsh Indoor and Outdoor Environments. Luxembourg: International Conference on Ad-Hoc, Mobile and Wireless Networks:ADHOC-NOW.; 2019.
  3. 3. Quentin Vey and François Spies and Baptiste Pestourie and Denis Genon-Catalot and Adrien van den Bossche and Thierry Val et al. POUCET: A Multi-Technology Indoor Positioning Solution for Firefighters and Soldiers. Spain: International Conference on Indoor Positioning and Indoor Navigation, IPIN 2021, Lloret de Mar.; 2021.
  4. 4. Long Zhenhuan and Xiang Yang and Lei Xiangming and Li Yajun and Hu Zhengfang and Dai Xiufeng. Integrated Indoor Positioning System of Greenhouse Robot Based on UWB/IMU/ODOM/LIDAR. Sensors. 2022;22(13):4819. pmid:35808314
  5. 5. Shuliang Zhang and Xiangquan Tan and Qingwen Wu. Self-Positioning for Mobile Robot Indoor Navigation Based on Wheel Odometry, Inertia Measurement Unit and Ultra Wideband. Malaysia: International Conferences on Vision, Image and Signal Processing, ICVISP 2021, Kuala Lumpur.; 2021.
  6. 6. Zhu Xiaomin and Yi Jianjun and Cheng Junyi and He Liang. Adapted Error Map Based Mobile Robot UWB Indoor Positioning. IEEE Transactions on Instrumentation and Measurement. 2020;69(9):6336–6350.
  7. 7. Li Luo and Zhang Feng and Wang Hong-Wei and Cui Long. Crawler Robot Indoor Positioning Based on a Combination of Bluetooth and IMU. Xiamen: International Conference on Robotics, Control and Automation.;2022.
  8. 8. Andreas Meuleman and Hakyeong Kim and James Tompkin and Min H. Kim FloatingFusion: Depth from ToF and Image-Stabilized Stereo Cameras. Israel: European Conference on Computer Vision, ECCV 2022, Tel Aviv, Israel.; 2022.
  9. 9. Deng Yong and Xiao Jimin and Zhou Steven Zhiying. ToF and Stereo Data Fusion Using Dynamic Search Range Stereo Matching. IEEE Transactions on Multimedia. 2022;24:2739–2751.
  10. 10. Akamine Sota and Totoki Shingo and Itami Taku and Yoneyama Jun. Real-time obstacle detection in a darkroom using a monocular camera and a line laser. Artificial Life and Robotics. 2022;27(4):828–833.
  11. 11. Hussain Babar and Wang Yiru and Chen Runzhou and Yue C. Patrick. Camera Pose Estimation Using a VLC-Modulated Single Rectangular LED for Indoor Positioning. IEEE Transactions on Instrumentation and Measurement. 2022;71:1–11.
  12. 12. Maximilian Berndt and Dennis Krummacker and Christoph Fischer and Hans D. Schotten. Unified Multi-Modal Data Aggregation for Complementary Sensor Networks Applied for Localization. Finland: IEEE Vehicular Technology Conference.; 2022.
  13. 13. Cai, Xiaobo and Han, Ke and Li, Yan and Li, Xuefei and Zhang, Jiajin and Zhang, Yue. A sensor attack detection method based on fusion interval and historical measurement in CPS. Austin: IEEE International Performance, Computing, and Communications Conference.;2020.
  14. 14. Baldoni Sara and Battisti Federica and Brizzi Michele and Neri Alessandro. A Hybrid Position Estimation Framework based on GNSS and Visual Sensor Fusion. Portland: Position, Location and Navigation Symposium.; 2020.
  15. 15. Dasgupta Sagar and Rahman Mizanur and Islam Mhafuzul and Chowdhury Mashrur. Sensor Fusion-based GNSS Spoofing Attack Detection Framework for Autonomous Vehicles. Computing Research Repository. 2021:abs/2106.02982.
  16. 16. Gautam Shroff and Saurabh Sharma and Puneet Agarwal and Shefali Bhat. A blackboard architecture for data-intensive information fusion using locality-sensitive hashing. Chicago: International Conference on Information Fusion.;2011.
  17. 17. Stewart Ryan and Palmer Todd S and Bays Samuel. An agent-based blackboard system for multi-objective optimization. Journal of Computational Design and Engineering. 2022;9(2):480–506.
  18. 18. Kim H and Lee I. LOCALIZATION OF A CAR BASED ON MULTI-SENSOR FUSION. International Archives of the Photogrammetry, Remote Sensing & Spatial Information Sciences. 2018;42(1):247–250.
  19. 19. Luo Qinghua and Yan Xiaozhen and Wang Chenxu and Shao Yang and Zhou Zhiquan and Li Jianfeng et al. A SINS/DVL/USBL integrated navigation and positioning IoT system with multiple sources fusion and federated Kalman filter. Journal of Cloud Computing. 2022;11(1):1–17.
  20. 20. Benjamin Noack and Clemens Öhl and Uwe D. Hanebeck. Event-Based Kalman Filtering Exploiting Correlated Trigger Information. Sweden: International Conference on Information Fusion.; 2022.
  21. 21. Sharma Girraj and Sharma Ritu. Joint Optimization of Fusion Rule Threshold and Transmission Power for Energy Efficient CSS in Cognitive Wireless Sensor Networks. Wireless Personal Communications. 2022;123(3):2107–2125.
  22. 22. Liao Yiwei and Shen Xiaojing and Rao Hang. Analytic Sensor Rules for Optimal Distributed Decision Given K-Out-of-L Fusion Rule Under Monte Carlo Approximation. IEEE Transactions on Automatic Control. 2022;65(12):5488–5495.
  23. 23. Yang Tianci and Lv Chen. A Secure Sensor Fusion Framework for Connected and Automated Vehicles Under Sensor Attacks. IEEE Internet of Things Journal. 2022;9(22):22357–22365.
  24. 24. Wang Xianchang and Lv Bingyu and Wang Kaiyu and Zhang Rui. ASTS: Autonomous Switching of Task-level Strategies. Applied Mathematics and Computer Science. 2023; Forthcoming.
  25. 25. Wang Fuyu and Xu Xiao and Chen Mengkai and Nzige Juma and Chong Fawen. Simulation research on fire evacuation of large public buildings based on building information modeling. Complex System Modeling and Simulation. 2022;1(2):122–130.