• No results found

Multi-sensor data fusion techniques for RPAS detect, track and avoid

N/A
N/A
Protected

Academic year: 2021

Share "Multi-sensor data fusion techniques for RPAS detect, track and avoid"

Copied!
12
0
0

Loading.... (view fulltext now)

Full text

(1)

Thank

7KH50

RXWSXWV

50,75

Citatio

See th

Version

Copyri

Link to

you for do

0,75HVHDUF

VRI50,78

5HVHDUFK5H

on:

is record i

n:

ght Statem

o Published

wnloading

FK5HSRVLWR

QLYHUVLW\UHV

HSRVLWRU\KWWSUHVHDUFKEDQNUPLWHGXDX

in the RMI

ment: ©

d Version:

this docum

RU\LVDQRSH

VHDUFKHUV

IT Researc

ment from

HQDFFHVVG

ch Reposit

the RMIT R

GDWDEDVHVK

ory at:

Research R

KRZFDVLQJW

Repository

WKHUHVHDUF



FK

PLEASE DO NOT REMOVE THIS PAGE

Cappello, F, Sabatini, R and Ramasamy, S 2015, 'Multi-sensor data fusion techniques for

RPAS detect, track and avoid', in Proceedings of the SAE 2015 AeroTech Congress and

Exhibition, Warrendale, PA, United States, 22-24 September 2015, pp. 1-12.

https://researchbank.rmit.edu.au/view/rmit:33426

Accepted Manuscript

2015 SAE International

(2)

Page 1 of 11

2015-01-2475

Multi-Sensor Data Fusion Techniques for RPAS Detect, Track and Avoid

Francesco Cappello, Subramanian Ramasamy, Roberto Sabatini

RMIT University – SAMME, Melbourne, Australia

Copyright © 2015 SAE International

Abstract

Accurate and robust tracking of intruders and objects is of growing interest amongst the Remotely Piloted Aircraft System (RPAS) scientific community. The ability of Multi-Sensor Data Fusion (MSDF) algorithms to support effective Detect, Track and Avoid (DTA) mission- and safety-critical tasks, and to aid in accurately predicting future states of intruders is explored in detail. RPAS are currently not equipped to routinely access all classes of airspace primarily due to the incipient development of a certifiable on-board Detect-and-Avoid (DAA) system. By employing a unified approach to cooperative and non-cooperative DAA, as well as providing enhanced Communication, Navigation and Surveillance (CNS) services, a cohesive logical framework for the development of an airworthy DAA capability can be achieved. In order to perform effective detection of intruders, a number of high performance, reliable and accurate avionics sensors and/or systems are adopted including non-cooperative sensors (visual and thermal cameras, LIDAR, acoustic sensors, etc.) and cooperative systems (Automatic Dependent Surveillance-Broadcast (ADS-B), Traffic Collision Avoidance System (TCAS), etc.). In this paper MSDF techniques required for sensor and system information fusion are critically analysed and their algorithms are presented. An Unscented Kalman Filter (UKF) and a more advanced Particle Filter (PF) are adopted to estimate the state vector of the objects based on maneuvering and non-maneuvering tasks. Furthermore, an artificial neural network is conceptualised to exploit the use of statistical learning methods, which act on combined information obtained from UKF and PF.

Introduction

Communication, Navigation and Surveillance (CNS) technologies are by definition, systems employed to support a seamless global ATM system [1]. The current and future improvements in CNS technologies will benefit the overall performance of Air Traffic Management (ATM) systems for both civil and military sectors of the aerospace industry. The development of these systems will lead to improvements in handling and transfer of information, efficient surveillance services using Automatic Dependent Surveillance (ADS) and ADS-Broadcasting (ADS-B), and improve the accuracy, integrity, availability and continuity of the navigation system [2]. Consequentially, separation distances between aircraft and/or Remotely Piloted Aircraft System (RPAS) will be reduced leading to a better demand and capacity balance of the airspace. More specifically, development of innovative communication systems will lead to a more direct and efficient air-ground linkages, interoperability across all applications and, better data formats and bandwidth leading to a reduction in communication errors [3]. In the foreseeable future, navigation systems will provide high-integrity,

high-reliability and worldwide all-weather navigation services [1]. Four-dimensional navigation accuracy will be developed that will allow for better airport and runway utilization and greater cost savings from reduction of ground-based navigation aids. The progression towards area navigation (RNAV) and Global Navigation Satellite System (GNSS) based operations will result in worldwide navigation coverage supporting Non-Precision Approach/Precision Approach (NPA/PA) capabilities that are currently being supported by majority of the airports infrastructure around the globe [4]. These improvements will offer a reduction in workload to the on-board and remotely operating pilot and thus promoting a less stressful environment. The key expected benefits of the development CNS/ATM technologies are listed below:

 Reduction in human errors by on-board and remotely operating pilot

 Increased airspace capacity and efficiency  Reduced separation distances between aircraft

Future surveillance systems will continue to use traditional Secondary Surveillance Radar (SSR) and Mode-S transponders and will be gradually introduced in high-density continental airspace [1]. As the implementation of ADS expands across continents, it will provide a better surveillance picture. Aircraft position will be automatically transmitted to ATM by the Next Generation Flight Management System (FMS) along with other information such as altitude, speed and distance through communication satellites and data links. ADS and ADS-B provide seamless position reporting between communication and navigation technologies and are also capable of providing information during transoceanic operations. ADS-B is a cooperative aircraft position determination surveillance technology using satellite navigation and periodically broadcasting to enable aircraft to be tracked. Aircraft to aircraft ADS-B transmission will also operate to provide situational awareness and facilitate in creating adequate, yet safe, separation and conflict avoidance distances. Therefore ADS-B technology is a key enabler to carry out efficient Detect, Track and Avoid (DTA) tasks for all types of aircraft including RPAS.

This research builds upon the development of a highly robust and effective Detect-and-Avoid (DAA) system for RPAS based on both non-cooperative sensors and cooperative systems [5]. Currently the RPAS sector of the aerospace industry is experiencing unparalleled levels of growth across the civil and military domains. RPAS technology is being proposed as an alternative to manned aircraft in an increasing number of applications especially in dull, dirty, dangerous (D3) missions and safety critical tasks. Some application domains include surveillance, aerial mapping, border patrol and fire detection. In particular, small to medium-sized RPAS have the ability

(3)

Page 2 of 11

of performing tasks with higher maneuverability [6]. The evolution of RPAS in the near future presents a series of challenges to develop systems which provide key functionalities (i.e., key operational and technical capabilities) to the RPAS platform which will enable constant access into all classes of airspace [7]. Identifying the commonalities and differences between manned and Remotely Piloted Aircraft (RPA) is the first step toward developing a regulatory framework that will provide, at a minimum, an equivalent level of safety for RPAS integration into all classes of airspace and aerodromes [8].

Safety is represented for manned aircraft by a consolidation system of rules regarding design, manufacture, maintenance and operation of aircraft including Air Traffic Management (ATM) procedures, systems and services, personnel licensing, and aerodynamic operations [9]. These rules have to be applied to RPAS on the basis of equivalent level principle, referring to the need to maintain a safety standard at least equivalent to the one required for manned aircraft [9]. For manned aviation the pilot, together with supporting systems, contributes to safe flight by avoiding traffic, ground obstacles and dangerous weather [10], electronic sensors and system will be adopted to provide key functionalities to the RPAS to satisfactorily mimic the function of the pilot in order to DTA objects and obstacles on a collision path. A shortcoming of RPA is that the situational awareness of the pilot is restricted to the field of view (FOV) of the Human machine interface (HMI) display screen. DAA functions and techniques will provide increased autonomy to the platform and consequentially reduce the pilot workload, stress and in addition offer increased safety and situational awareness [11]. To facilitate civilian, military and academics in their research to develop key systems and software which RPAS require to be integrated into all classes of airspace, a roadmap has been developed by the International Civil Aviation Organization (ICAO) as part of the Aviation System Block Upgrades (ASBU) [12]. Specifically, RPAS integration entails the insertion into the air traffic management system, DAA (air and ground) and situational awareness), weather awareness [10]. The DAA capability will enable RPAS to perform equally or exceed the performance of the see-and-avoid ability of the pilot in manned aircraft systems [7]. Both cooperative sensors and non-cooperative systems for DAA tasks are being developed to enable RPAS to routinely access all classes of airspace [5, 7].

Both cooperative and non-cooperative systems/sensors are used for the DAA function measurements, the hardware devices include; passive and active Forward Looking Sensors (FLS), Traffic Collision Avoidance System (TCAS) and Automatic Dependent Surveillance – Broadcast (ADS-B) [5, 7]. Enhanced surveillance solutions based on Automatic Dependent Surveillance-Broadcast (ADS-B) system and Traffic Collision Avoidance System (TCAS) and required for addressing cooperative DAA functions [5, 7]. A number of ground/space based sensors/systems are used for surveillance of the surrounding environment of an Aircraft [7]. The progressive development of the GPS-based ADS-B systems shall gradually change the roll of radar use in civil surveillance applications [13]. New systems which are to be developed should be interoperable and compatible with current systems in use [13]. The airspace surveillance communication dynamics where ADS-B and other technologies such as Satellite Communications (SAT COM) are used, will all eventually be fused with onboard non-cooperative sensors on the aircraft. The emphasis of this paper is on fusing cooperative systems and non-cooperative sensor measurements for enhanced airborne surveillance (DTA) solutions illustrated in Figure 1.

- Cooperative Systems - Non-cooperative Sensors - Naturally Inspired Sensors

Track Avoid

Detect

Detect, Track and Avoid Loop

Figure 1. Airborne surveillance system and DAA functions.

In Intelligent Surveillance Systems (ISS) technologies such as computer vision, pattern recognition, and artificial intelligence are developed to identify abnormal behaviors in videos. As a result, fewer human observers can monitor more scenarios with high accuracy [14]. This means that surveillance operations are constantly evolving into an increasingly more autonomous nature where human interaction is needed less frequently. Multiple sensing is the ability to sense the environment with the concurrent use of several, possibly dissimilar, sensors [15]. The reason for multiple sensing is to be able to perform more tasks, more efficiently, in a reliably way [15]. These techniques are increasingly being applied to aerial robotic systems such as Vision Based Navigation (VBN) techniques for approach and landing tasks [16, 17]. However due to the inherent robotic nature of the RPAS, its engineering is focusing on an increasingly intelligent approach to design and integration. Therefore surveillance systems onboard RPAS are also being developed to reduce the workload of pilots and to increase autonomy to the vehicle. Surveillance systems use a number of sensors to obtain information about the surrounding environment. Visual and thermal sensors are two of the many sensors used. To fully exploit the fact that more than one sensor is being used, the information is fused using an approach called data fusion. Data fusion is implemented in the research areas such as signal processing, pattern and image recognition, artificial intelligence, and information theory. A common feature of these architectures is that they incorporate multiple levels of information processing within the data fusion process [18, 19]. More specifically sensor fusion is the combining of sensor data from unlike sources so the resulting information has less uncertainty than if these sources were used as an individual sensor. Current state-of-the-art airborne sensors and MSDF methods for surveillance detection and tracking tasks have employed non-cooperative and cooperative sensors/systems for DAA solutions [7]. Data fusion of surveillance sensors utilizes these sensors/systems in order to detect and track obstacles and intruders. A non-cooperative collision avoidance system for RPAS using pulsed Ka-band radar and optical sensors has been developed [20]. Acoustic-based DAA system for small RPAS by adopting a multi sensor platform with acoustic sensors that allows detection of obstacles and intruders in a 360º Field of View (FOV) and by performing quick-reaction manoeuvres for avoidance has been adopted [21]. This system allows all weather operations and offers advantages in terms of power consumption and cost. An approach to the definition of encounter models and their applications on the DAA strategies, presented in [7], takes into account both cooperative and non-cooperative scenarios. Ground-Based DAA (GBDAA) systems using electronic sensors are also currently being developed. A summary of

(4)

Page 3 of 11

airborne surveillance systems currently available is provided in Table 1.

Table 1. Airborne surveillance systems [adopted from 1].

Airspace Current system Future System

Oceanic continental

En-route airspace with high / low-density traffic

Primary radar/ SSRVoice position reports OMEGA/LORAN-CNDB ADS RNAV/RNPG NSS Continental airspace with high-density traffic Primary radar SSR Mode A/C SSR [Mode A/C or Mode S] ADS Terminal areas with high-density Primary radar SSR Mode a/c SSR [Mode A/C or Mode S] ADS

These ground based systems provide information on manoeuvre decisions for terminal-area operations [22]. In regards to image processing techniques, the target that is well camouflaged for visual detection will be hard (or even impossible) to detect in the visual bands, whereas it can still be detectable in the thermal bands [23]. A combination of visible and thermal imagery may then allow both the detection and the unambiguous localization of the target (represented in the thermal image) with respect to its background (represented in the visual image) [24]. Significant outcomes are expected from novel Communication, Navigation, Surveillance/Air Traffic Management (CNS/ATM) systems, in line with the large-scale and regional ATM modernisation programmes including Single European Sky ATM Research (SESAR) and Next Generation Air Transportation System (NextGen). The key enabling elements required for the evolution of CNS/ATM and Avionics (CNS+A) framework have been identified as part of these programmes [25-30]. In this paper a novel technique for data fusion is introduces and illustrated with a number of architectures based on visual/thermal sensor fusion with ADS-B/TCAS systems. This architecture will be employed for DAA applications for both manned and RPAS. The paper is arranged as follows, an overview of the methodology adopted for data fusion is presented followed by a list of algorithms adopted for the fusion of cooperative systems and non-cooperative sensors information for tracking. Furthermore the architectures adopted for ADS-B/TCAS and visual/thermal fusion are also described.

DTA Multi-Sensor Architecture and Data Fusion

MSDF is an effective way of optimizing large volumes of data and is implemented by combining information from multiple sensors to achieve inferences that are not feasible from a single sensor or source [31]. MSDF is an emerging technology applied to many areas in civilian and military domains, such as automated target recognition,

battlefield surveillance, and guidance and control of autonomous vehicles, monitoring of complex machinery, medical diagnosis, and smart buildings [32]. Techniques for MSDF are drawn from a wide range of areas including pattern recognition, artificial intelligence and statistical estimation [31]. There are many algorithms for data fusion which optimally and sub-optimally combined the information to derive a processed best estimate. Mathematically the type of system determines if the system can be optimally estimated, in general linear systems can be optimally estimated with the original Kalman Filter and non-linear systems are sub-optimally estimated with approximation techniques which linearize the non-linear system. Real life navigation and surveillance applications are almost always described by non-linear equations therefore approximation techniques are commonly applied, some techniques include; the Extended Kalman Filter (EKF), Unscented Kalman Filter (UKF), Square Root-UKF (SR-Root-UKF), Cubature Kalman Filter (CKF), Point Mass Filter (PMF) and the Particle Filter (PF) [16, 17, 33]. In the past three decades, the EKF has become the most widely used algorithm in numerous nonlinear estimation applications [34]. In comparison to the UKF, the EKF is difficult to manipulate (i.e. computationally intractable) due to derivation of the Jacobean matrices [16]. Furthermore, the accuracy of the propagated mean and covariance is limited to first order Taylor series expansion, which is caused by its linearization process [35]. The UKF overcomes the limitations of the EKF by providing derivative-free higher-order approximations by approximating a Gaussian distribution rather than approximating an arbitrary nonlinear function [36, 37]. The UKF is more accurate and robust in navigation applications by also providing much better convergence characteristics. The UKF uses sigma points and a process known as Unscented Transform (UT) to evaluate the statistics of a nonlinear transformed random variable [35]. Due to the advancements in modern computing technologies, algorithms such as the UKF and the PF can now be implemented effectively for real time systems. The aim of this research is to adopt MSDF techniques to provide the best estimate of track data obtained from cooperative systems and non-cooperative sensors. The UKF is used to processes the cooperative sensors and a PF is adopted to combine the non-cooperative sensor measurements. The PF or Sequential Monte Carlo (SMC) are a broad and well known category of Monte Carlo algorithms developed in the last two decades to provide approximate solutions to intractable inference problems [38]. The methodology is used for nonlinear filtering problems seen in Bayesian statistical inference and signal processing. Using the PF for Filtering or estimation problems, the observations are used to measure internal states in systems. The system dynamic process and sensors are also filled with random noise utilised in the filtering process. The objective is to compute the posterior distributions of states or hidden variable, from the noisy and partial observations. The main objective of a PF is estimating the posterior density of the hidden state variables using the observation variables if applicable. The PF is designed for a hidden Markov Model, where the system consists of hidden and observable variables. The dynamical system describing the evolution of the state variables is also known probabilistically obtained from the prediction and updating process of the PF. A set of particles is initialised in the PF to generate a sample distribution, where each particle is assigned a weight, which represents the probability of that particle to be sampled from the probability density function. Weight disparity leading to weight collapse is a common issue encountered in these filtering algorithms; however it can be mitigated by including a resampling step before the weights become too uneven [39]. ADS-B and TCAS are exploited in the MSDF architecture which provides position and velocity estimates for the detection and tracking solution. The MSDF technique used to fuse the cooperative sensors is the UKF which is MSDF technique N, the MSDF technique used to fuse the non-cooperative sensors is the PF

(5)

Page 4 of 11

which is MSDF technique NC and the MSDF technique used to fuse the MSDF N and MSDF NC sensors is based on a memory based learning. Future research will aim to use the non-maneuvering tracking algorithms to teach the NN. A typical architecture for MSDF is illustrated in Figure 2.

Cooperative Systems Non-Cooperative Sensors

Tracks of Intruders/ objects

C MSDF NC MSDF DTA MSDF ADS-B TCAS Visual Thermal MMW Radar Lidar Acoustic

Figure 2. Multi-Senor Data Fusion for DAA.

In order to estimate the nonlinear functions, approximation methods are used. On-board trajectory replanning with dynamically updated constraints based on the intruder and the host dynamics is at present used to generate obstacle avoidance trajectories [7, 40]. The states of the tracked obstacles are obtained by adopting UKF/PF or other MSDF techniques is used in order to predict the trajectory in a given time horizon. Coarse-resolution radar based DAA solution is implemented for mini RPAS and its information is fused with data from ADS-B system. In most realistic problems the real world may be described by nonlinear differential equations, and the measurements may not be a function of those states. In this case the measurements are a linear function of the states, but the model of the real world is nonlinear. Previous research included the use of the IMM filter to fuse the measurements of the cooperative non-cooperative unified approach [7]. In order to predict the track of the target object/aircraft MSDF techniques are used to obtain the best estimate and then a probabilistic model is used. What is being proposed is that the UKF is used to obtained a better estimate, therefore a better prediction will be obtained because the errors are essentially reduced. The detection information provided by several awareness sensors should be evaluated and prioritized taking into account the rest of the UAS information such as telemetry, flight plan and mission information [41]. A multi-sensor surveillance system includes function such as detection and tracking which utilize cooperative and non-cooperative information sources in a multi-sensor architecture. The output of an effective multi-multi-sensor tracker provides a unified surveillance picture with a relatively small number of confirmed tracks and improved target localization [19]. Global multi-sensor fusion and tracking that is capable of processing (potentially anomalous) AIS tracks and contact-level or track-level data from other sensors to produce a single, consolidated surveillance picture [19]. In this paper, we focus on the second of these four technology areas including; a high-performance multi-sensor fusion that provides the basis for anomaly-detection work [12]. The following filters have been used for tracking [37] of intruders.

Obstacle detection and tracking algorithms Tracking for single object non maneuvering target

1. The optimal Bayesian filter 2. The Kalman Filter

3. The Extended Kalman Filter 4. The Unscented Kalman Filter 5. The Point Mass Filter 6. The Particle Filter

Maneuvreing object tracking

1. The optimal Bayesian filter

2. Generalized pseudo-Bayes Filters (of order 1 and order 2) 3. Interactive Multiple Model

4. Bootstrap Filter (Auxuilary)

5. Extended Kalman Auxiillary Particle Filter

Single Object Tracking in Clutter

1. The optimal Bayesian filter 2. Nearest Neighbor filter 3. Particle filter

4. Extended Kalman auxiliary particle filter

Single- and Multiple-Object Tracking in Clutter: Object-Existence-Based Approach

1. Optimal Bayes’ recursion

2. Joint Probabilistic Data association (JPD)

3. Interacting Multiple Models – Joint Integrated Probabilistic Data Association (IMM-JIPDA)

Multiple-object tracking in clutter: random-set-based approach

Random Finite Set Formalism (RFS)

1. The optimal Bayesian multi-object tracking filter 2. The probabilistic hypothesis density approximation

Approximate filters

1. Gaussian mixture PHD filter 2. Particle PHD filter

3. Gaussian mixture CPHD algorithm

Object-existence-based tracking filter

1. The integrated probabilistic data association filter 2. Joint IPDA derivation from Gaussian mixture PHD filter Several well-known Bayesian algorithms for recursive state estimation. These algorithms, known as filters, result from the general Bayesian recursive equation [37]. These filtering algorithms are viewed as algorithms for the single, multiple, maneuvreing and non-maneuvreing object tracking problem. Maneuvering objects are those objects whose dynamical behavior changes overtime. An object that suddenly turns or accelerates displays a maneuvering behavior with regard to its tracked position [37]. While the definition of a maneuvering object extends beyond the tracking of position and

(6)

Page 5 of 11

speed, historically it is in this context that maneuvering object tracking theory developed [37]. In the future, this service has to incorporate 4D navigation in order to know at what moment the RPAS is going to arrive at the different flight plan waypoints. All flight phases of the RPAS go through to perform a mission are also provided (such as on-ground, departure, en-route, mission or arrival). This data is going to be important in selecting a suitable conflict reaction. As a result of this prioritization, the ADF service may choose between two possible outputs: collision avoidance maneuver required or self-separation maneuver required [41]. A simple false track discrimination scheme may be described by the track status propagation as illustrated in Figure 3.

Track Tentative Status

Track Threshold False Track Discrimination Scheme

Confirmation or Termination of Track based on New Measurements

Figure 3. Multi-Senor Data Fusion for DAA.

 Each new track has the tentative status, which may be changed by confirmation or termination using subsequent measurements  if the probability of object existence rises above a predetermined

track confirmation threshold, the track becomes confirmed, and stays confirmed until termination; and

 when the probability of object existence falls below a predetermined track termination threshold, the track is terminated and removed from memory

The Unscented Kalman Filter

The UKF is an alternative to the EKF which shares its computational simplicity while avoiding the need to derive and compute Jacobians and achieving greater accuracy [37]. The UKF is based upona the UT process, this method approximates the moments of a non-linearly transformed random variable. The UT can be used as the basis of a non-linear filtering approximation which has the same form as the EKF but is derived in a quite different manner. The posterior PDF of

xk using Bayes’ rule is given by:

( ) ( ) ( ) (1) The UKF approximates the joint density of the state and measurement conditional on the measurement history is approximated by a Gaussian density expressed as:

( ) ([ ] [ ̂ ̂ ] [ ]) (2)

This implies that:

( ) ( ̂ ) (3) In general, the moments appearing in these equations cannot be computed exactly and so must be approximated. The UT is used for this purpose. The state prediction is performed by predicting mean and covariance matrix of which represent the moments of transformation given by:

( ) (4) where the statistics of given . The predicted mean and covariance matrix are given by:

( ) ( ( ) ) ( ( ) ) (5) ( ) ( ( ) )

( ( ) ) ( ) (6) Approximations to the predicted mean and covariance matrix are to be obtained using the UT. As described in Section 2.4.1, sigma points and weights are selected to match the mean and the covariance matrix of ( ). With ( ) where , the UT approximations to the predicted mean and covariance matrix are [37]:

̂ ∑ (7) ∑ ( ̂ )( ̂ ) (8) The measurement prediction: The predicted statistics for the measurement yk are moments of the transformation are given by:

( ) (9) where the statistics of given are available from the prediction step. Hence we have:

( ) ( ( ) ) (10) ( ) ( ( ) ) ( ) (11) ( ) ( ( ) ) (12) Let and denote the sigma points and weights, respectively, selected to match the predicted mean and covariance matrix. The transformed sigma points are ( ), . The UT approximations to the moments are [37]:

̂ ∑ (13) ∑ ( ̂ )( ̂ ) (14) ∑ ( ̂ )( ̂ ) (15) The resulting expression for the posterior PDF is given by:

(7)

Page 6 of 11

( ) ( ̂ ) (16) where:

̂ ̂ ( ̂ ) (17) (18) As with the EKF, the UKFis strongly reminiscent of the KF with the main difference being the dependence of the posterior covariance matrix on the observed measurements. In UKF this occurs because the calculated sigma points generated in the UT process are moment approximations and are determined based on state estimates which are measurement dependent [37]. Generally the UKF performs to the same order of computational expense as the EKF, however it generally performs much better. This can be attributed to increased accuracy of the moment approximations in the UKF attributed to the UT process. In fact, the UKF achieves the same level of accuracy as the second-order EKF, which requires the derivation of Jacobian and Hessian matrices [37]. Before concluding, we note that the assumption of additive noise in the dynamic and measurement equations has been made only for the sake of convenience [37]. The UT is equally applicable for non-additive noise. For instance, if the dynamic equation is ( ), then the quality undergoing a non-linear transformation is the augmented variable [ ] , which has the statistics ( ) [ ̂ ] and ( ) ( ) [37]. The UT can be applied to the random variable transformed through the function toapproximate the predicted state statistics [37]. Since is of dimension a larger number of sigma points will be required than for the case where the dynamic noise is additive [37]. Similar comments hold for a measurement equation which is non-linear in the measurement noise [37]. A summary of the UKF is given by below.

Algorithm Review

1. Determine sigma points and weights to match a mean ̂ and covariances matrix . 2. Compute the transformed sigma points ( ),

.

3. Compute the predicted state statistics:

̂ ∑ (19) ∑ ( ̂ )( ̂ ) (20) 4. Determine the sigma points and weights to

match mean ̂ and covariance matrix .

5. Compute the transformed sigma points ( ), . 6. Compute the predicted measurement statistics:

̂ ∑ (21) ∑ ( ̂ )( ̂ ) (22) ∑ ( ̂ )( ̂ ) (23)

7. Compute the posterior mean and covariance matrix

̂ ̂ ( ̂ ) (24) (25)

The Particle Filter

There are different methods for target tracking, but particle filtering (PF) is one of the most important and widely used methods. The PF is implementation of Bayesian recursive filtering using sequential Monte Carlo sampling method [42]. The PF is used in statistics and signal processing, which is known as bootstrap filter and Monte Carlo filter respectively [43]. The target state is represented by where is time. is the set of all observations. The posterior density function is ( ) is represented by a set of weighted samples. Observations ( ) in the PF method can be non-gaussian. The main idea of the PF is the estimation of the Probability Distribution Function (PDF) of the target object by a set of weighted samples {( ( ) ( )) } such that ∑ ( )

[43]. Based

on the observations, likelihood of each particle is applied in weight assignment by the equation below in each step:

( ) ( ( )) (26)

The PF consists of two stages, prediction and update. In the prediction state, next step particles are calculated by using current step particles and target state transition equations. In the update stage, the weights are assigned to new particles by the following equation [43]:

( ) ( )

( ) (27)

where is the importance density function which is used in the sampling function. Posteriori density function is estimated by using the following equation:

( ) ∑ ( ) (28) The next step state vector is estimated by using the weighted average on current step particles with the following equation [43]:

̂ ∑ (29)

Sequential Importance Resampling Particle Filter

One of the basic challenges in implementation of PF is appropriate choice of proposal density function. One of the common suboptimal choices of proposal density function is transitional prior distribution [43].

( ) ( ) ( ) (30) By substituting ( ) into each particle’s weight, a new weighting can be calculated based on:

(8)

Page 7 of 11

In each step of the SIR algorithm, re-sampling is done after state estimation. After re-sampling, weights of all particles are equal too where is the number of particles. Hence the above equation is modified as follows:

( ) (32)

SIR Algorithm

Initialise: the particles using prior distribution of target. FOR i=1:N

a. Prediction: Generate particles using dynamic model ̂ ( ) (33) b. Weighting: Calculate weight for particle

( ) (34)

END FOR

Normalize the particle weights by

̂

(35)

Resampling: Resample the particles and obtain the new part

with same weight and repeat from stage 2.

{( )}

{( )} (36)

Statistical Learning Techniques

Statistical learning theory is a framework for machine learning drawing from the fields of statistics and functional analysis [44]. Statistical learning concerns itself with finding a predictive function for a task/problem. Computer vision, speech recognition and bioinformatics are areas which Statistical learning theory has been successfully applied. The goal of learning is prediction. Learning falls into many categories, including supervised learning, unsupervised learning, online learning, and reinforcement learning. From the perspective of statistical learning theory, supervised learning is best understood [45]. Supervised learning involves learning from a training set of data. Every point in the training is an input-output pair, where the input maps to an output. The learning problem consists of inferring the function that maps between the input and the output in a predictive fashion, such that the learned function can be used to predict output from future input. In machine learning problems, a major problem that arises is that of overfitting. Because learning is a prediction problem, the goal is not to find a function that most closely fits the (previously observed) data, but to find one that will most accurately predict output from future input [46]. Some examples of Statistical learning techniques include but are not limited to; Artificial Neural Networks, Instance Based Learning,

Learning with hidden variables and Kernel machines which are succinctly depicted in Figure 4 [47].

Statistical Learning Methods Statistical Learning Methods Learning with Hidden Variables Learning with Hidden Variables Instance-Based Learning Instance-Based

Learning Artificial Neural Networks

Artificial Neural Networks Kernel Machines Kernel Machines

Figure 4. Artificial intelligent techniques (Adapted from [33]).

The difficulty is increased by the fact that the environment is fraught with uncertainty [48]. In this case the RPA or agent can handle uncertainty by using the methods of probability and decision theory, but first they must learn their probabilistic theories of the world from experience. The state-of-the-art methods used for learning probability models are primarily based on Bayesian networks. Learning models are used to store and recall specific instances. In addition, neural networks learning and kernel machines are prevalent techniques used for information learning.

The detection of maneuver can rely on the properties of the innovation process of the filter that is matched to the non-maneuvering system. When the maneuver takes place the residual are no longer zero mean white noise and changes their statistical properties. The proposed detection algorithm is based on statistical testing theory. The decision problem can be formulated as statistical test of two hypothesis which checks probability density function of innovation process. Hypothesis H0 represents non maneuvering object and alterative hypothesis H1 represents a maneuvering object. The ability to make rapid and effective decisions is difficult to do, but Information is represented as real vectors in artificial neural networks. This uniformity allows for arbitrary hierarchical combinations of neural classifiers, because the features or outputs of one classifier can simply be regarded as inputs of another classifier, and outputs of several classifiers can be combined by concatenation of the real output vectors. The equations for an artificial neural network learning algorithm are presented below:

Function: Perceptron-Learning (example, network) returns a

perception hypothesis.

Inputs: examples, a set of examples, each with input

and output . Network, a perception with weights , , and activation function .

Repeat for each in examples do

(9)

Page 8 of 11

Err [ ] ( ) (38)

( ) [ ] (39)

Until some stopping criteria is satisfied Return: Neural-Net-Hypothesis (network).

The method which is adopted to fuse the measurements from cooperative and non-cooperative sensors is artificial neural networks. Neural networks seem to be particularly well suited for the combination of inputs from completely different sources. So we will be using them and the corresponding learning and training strategies for these problems. Artificial neural networks have become a standard paradigm for pattern recognition. They are also well-suited for the formulation of sensor-, information-, or decision-fusion scenarios [49]. The Heuristics based MSDF is shown in Figure 5.

Non Parametric Learning Techniques

Instance Based Learning

- Nearest-Neighbour Methods - Kernel Methods

Artificial Neural Networks

- Neural Structures - Single Layered NN - Multi-Layered NN - Leaning NN structures

Figure 5. Heuristics based MSDF (Adapted from [33]).

Typically, three-layer networks are used for classification: The first layer or input layer contains the data to be analyzed. Here, it is usually assumed that the data are given as real vectors of fixed dimensionality (perhaps after some appropriate pre-processing) [49]. The second layer or hidden layer contains non-linear combinations of the data, which can be regarded as features that are formed by neural-network learning or training [49]. The third layer or output layer represents the decision of the neural network. For a classification problem with k classes there are typically k output-neurons. Ideally, the network should respond with the jth unit vector to the jth class. In practice, the output value of the jth unit (typically between 0 and 1) is regarded as the strength of evidence for or belief in the object belonging to class j [49]. After proper training of the network, the output of the jth unit should actually be an estimate of the a posteriori probability p(cj..d) of class j. Figure 7 is an illustration of artificial intelligence techniques used to fuse cooperative and non-cooperative sensors techniques. In the proposed method the RBF or the HRBF neural networks (NN) are applied as probability density function (pdf) estimators [50]. The neural networks are trained for certain observation scenario. In this step innovation process of filter tracking non-maneuvering object is used as learning data. The RBF neural network consists of three layers: an input, a middle and an output layer [50]. The input layer corresponds to the input vector space. Each input neuron is fully connected to the middle layer neurons. The middle layer consists of m neurons. Each middle layer neuron computes an activation function which is usually the Gaussian function 1 [51].

( )

( )⁄ (

‖ ‖

) (40)

Where: – center of activation function, – width of activation function, – dimension of the estimated pdf. The output layer consists of neurons which correspond to the classification problem. The output layer is fully connected to the middle layer. Each output layer neuron computes a linear weighted sum of the outputs of the middle layers as shown in equation 2 [51].

( ) ∑ ( ), (41) Where is the weight value between the middle layer neuron and the output layer neuron. For multidimensional pdf, the activation function can be given by [51]:

( ) ( ) ( ) (

( ) ( )

) (42) where is the covariance matrix. Application of activation function of the form (3) allows reducing the number of neurons in the middle layer. The learning task of RBF network is split into two separate subtasks [5]. The first is determination of activation function parameters (centers and weights) and the second -the calculation of the output weights w. The centers are associated with the clusters of input data, represented by the average vector x. The self-organization of this clustering process is usually done by the competition among neurons, according to the strategy of "Winner Takes All" (WTA) or "Winner Takes Most" (WTM) [50]. After adaptation of the centers the adjustment of the width parameter o is done in a way to provide the continuous approximation of the data. Simultaneously, the weights of the output neurons are calculated. The sum of all weights is always equal to one [50].

Vision and Thermal Fusion

In this section we aim to provide a basic conceptual understanding of the fusion of the cooperative systems and non-cooperative sensors. Due to the significant details required to explain all processes involved in the fusion of all the sensors and system one example is taken for this paper [52]. In this paper we present the fusion concept of combining ADS-B and TCAS systems with a visual and thermal fused configuration. The medium/filter used to fuse these measurements is dome using artificial neural network which can be trained to perform better. The proposed training which will be used to teach the system is non-manoeuvring and manoeuvring tracking. The image fusion between the thermal and visual sensors is carried out. In recent years, much development has been undertaken by computer vision community in the area of image sensing technologies using a multi-sensor framework. Application of this technology has found it place in many civil and military areas such as medical imaging and remote sensing. Single sensor operations and technologies do not provide the information and advantages of multi-sensor systems. A benefit of multi-sensor systems is that information that is not normally considered useful can be exploited for example complementary information about the region surveyed so that image fusion provides an effective method to enable comparison and analysis of such data [53]. By definition image fusion is the process of combining information in two or more images of a scene to enhance viewing or understanding of the scene [54]. The fusion process must preserve all relevant information in the fused image [55]. The methodology of image fusion and the computational processes involved were developed by the MIT Lincln Laboratory where inspiration for the models developed were derived from biological models of colour vision for fusion or IR radiation and

(10)

Page 9 of 11

visible light [56]. The process involved in the biological make up of certain species snakes (i.e., rattlesnake and pythons) follows the following process. In the colour image fusion methodology the individual image inputs are first enhanced by filtering them with a feedforward center-surrounding shunting neural network. The operation serves to [57]:

1. Enhance special contrast in the individual visible and IR bands 2. Create both positive and negative polarity IR contrast images 3. Create two types of single-opponent colour contrast images The resulting single opponent colour contrast images represent greyscale fused images that are analogous to the IR-depressed visual and IR enhanced visual cells of the rattlesnake [24]. The observer evaluation of the fusion process is critical to the effectiveness of the system and for the operator ergonomics. Therefore the evaluation of the fusion process would be to seen an ideal fused image ‘ground truth’ against on to which all others can be compared, this would be the most natural approach, however this is rarely the case with real systems. In a multi-sensor system images from several input sources can only be fused into one output if the inputs contain spatially and spectrally independent information, this is also rarely the care in real fusion applications, so the generally accepted objective is summaries as follows, only the most visually important input information is used to generate the output [24]. Considering that the most common application, and in the case of our application, is for display purposes, the preservation of perceptually important visual information becomes an important and critical issue. Therefore, the ideal fused output image should therefore contain all the important visual information, as perceived by an observer viewing the input images. Also, a single-level fusion system must ensure that no distortions or other ‘false’ information is introduced in the fused image [58]. The visual and thermal image fusion process is illustrated in Figure 6.

Common Field of View Coefficient Map

Generation Coefficient Map

1. Fusion Decision Map Generation 2. Detection 3. Segmentation 4. Target Tracking

Stop Fusion Process Begin Fusion Process

Visual Image Thermal Image

Figure 6. Vision/thermal image fusion process.

The tracking data obtained from the visual and thermal image fusion process is combined with ADB-B measurements as illustrated in Figure 7.

Processing and Data Sorting ADS-B and TCAS

Data Fusion

Visual and Thermal Image Fusion

UKF PF

Artificial Neural Network

Multi-Sensor Track Update

Track Management

Begin Fusion Process

Stop Fusion Process

Fused Measurements

Figure 7. Vision/thermal image and ADS-B measurements fusion process. The typical performance criteria include [58]:

1. Identification and localisation of visual information in the input and fused images.

2. Evaluation of the perceptual importance.

3. Measuring (quantification) the accuracy with which input information is represented in the fused image.

4. Distinguishing true scene (input) information and fused artefacts in the output fused image.

Conclusions

This paper summarizes initial research activities conducted on surveillance MSDF methods for mini RPAS. In particular, Target detection and tracking methods for single and multiple target situations. Multi-sensor architectures in which information from the sensors is combined leads to significantly more robust on-board surveillance systems. Multi-sensor systems enrich the surveillance picture, and provide an effective basis for anomaly detection processing. Civil-military aircraft and RPAS can significantly benefit from this architecture, in that aircraft can operate in high density traffic with high performance. A conceptual analysis on the ground based and airborne sensors methods has been carried out. Cooperative and non-cooperative sensor/systems were utilised for DAA function measurements, passive and active Forward Looking Sensors (FLS), Traffic Collision Avoidance System (TCAS) and Automatic Dependent Surveillance – Broadcast (ADS-B). Enhanced surveillance solutions based on Automatic Dependent Surveillance-Broadcast (ADS-B) system and Traffic Collision Avoidance System (TCAS) and required for addressing cooperative DAA functions. The emphasis of this paper is on fusing cooperative systems and non-cooperative sensor measurements for enhanced airborne surveillance (TDA) solutions. This review discussed the application of MSDF techniques for non-cooperative and cooperative sensor/systems fusion and data processing.

(11)

Page 10 of 11

References

1. ICAO, “Global Air Navigation Plan for CNS/ATM Systems,” Doc 9750, 1998.

2. ICAO, “Manual of Air Traffic Services Data Link Applications,” Doc 9694-AN/955, 1999.

3. ICAO, “Manual on Required Communication Performance (RCP)”Doc 9869-AN/462, 2006.

4. EUROCONTROL,” European Organisation for the Safety of Air Navigation,” Edition 2.1,1999.

5. Ramasamy, S., and Sabatini, R., “Unifying Cooperative and Non-Cooperative UAS Sense-and-Avoid,’ Conference: Second International Workshop on Metrology for Aerospace (MetroAeroSpace 2015), At Benevento, Italy, 2015.

6. Laliberte, A.S., Rango, A. and Herrick, J.E., “Unmanned Aerial

Vehicles for Rangeland Mapping and Monitoring: A Comparison of Two Systems”, Proceedings of the ASPRS

Annual Conference, Tampa, Florida, USA, 2007.

7. S. Ramasamy, R. Sabatini, A. Gardi, “Towards a Unified Approach to Cooperative and Non-Cooperative RPAS Detect-and-Avoid.” Fourth Australasian Unmanned Systems Conference 2014 (ACUS 2014), Melbourne (Australia), December 2014. DOI: 10.13140/2.1.4841.3764

8. ICAO, “Unmanned Aircraft Systems (UAS),” Cir 328-AN/190, 2011.

9. Crespo, D.C., Mendes de Leon, P., “Achieving the Single European Sky: Goals and Challenges,” vol. 8, Kluwer Law International, 2011.

10. ICAO, “Roadmap for the integration of civil Remotely-Piloted Aircraft Systems into the European Aviation System,” A Strategic R&D Plan for the integration of civil RPAS into the European Aviation System, 2013.

11. Gu, W., Mittu, R., Marble, J., Taylor, G., Sibley, C., et el., “Towards Modeling the Behavior of Autonomous Systems and Humans for Trusted Operations,” The Intersection of Robust Intelligence and Trust in Autonomous Systems: Papers from the AAAI Spring Symposium, 2014.

12. ICAO., “Introduction to the Aviation System Block Upgrade (ASBU) Modules Strategic Planning for ASBU Modules Implementation,” 2013.

13. Baud, O., Gomord, P., Honoré, N., Lawrence, P., Ostorero, L., Paupiah, S., and Taupin, O., “Air Traffic Control Tracking Systems Performance Impacts with New Surveillance Technology Sensors,” Aerospace Technologies Advancements, 2010. DOI: 10.5772/7164

14. Huihuan, Q., Xinyu, W., Yangsheng, X., “Intelligent Surveillance Systems,” vol. 51. Springer Science & Business Media, 2011.

15. Aggarwal, J. K., “Multisensor Fusion for Computer Vision,” Vol. 99. Springer Science & Business Media, 2013.

16. Cappello, F, Subramanian, R, and Sabatini, R., “Multi-sensor Data Fusion Techniques for RPAS Navigation and Guidance”, AIAC 16: Multinatioinal Aerospace Programs-Benefits and Challenges. Engineers Australia, 2015.

17. Sabatini, R., Cappello, F., Ramasamy, S., Gardi, A. and Clothier, R., “An Innovative Navigation and Guidance System for Small Unmanned Aircraft using Low-Cost Sensors”, Journal of Aircraft Engineering and Aerospace Technology, 2015. 18. Dailey, D.J., Harn, P., and Lin P-J., “ITS Data Fusion,” Final

Research Report, Research Project T9903, Task 9, ATIS/ATMS Regional IVHS Demonstration, 1996.

19. Gad, A., and Farooq, M., “Data Fusion Architecture for Maritime Surveillance,” Proceedings of the Fifth International Conference on Information Fusion. Vol. 1. IEEE, 2002.

20. Fasano, G., Accardo, D., Moccia, A., Carbone, C., Ciniglio, U., Corraro, F., and Luongo, S., “Multi-Sensor Based Fully Autonomous Non-Cooperative Collision Avoidance System for Unmanned Air Vehicles,” Journal of Aerospace Computing, Information, and Communication, vol. 5, no. 10, pp. 338-360, 2008. DOI: 10.2514/1.35145

21. Rodriguez, L., Sabatini, R., Ramasamy, S., and Gardi, A., “A Novel System for Non-Cooperative UAV Sense-And-Avoid”, European Navigation Conference (ENC 2013), Vienna, Austria, 2013.

22. Spriesterbach, T.P., Bruns, K.A., Baron, L.I., and Sohlke, J.E., “Unmanned Aircraft System Airspace Integration in the National Airspace Using a Ground-Based Sense and Avoid System,” Johns Hopkins APL Technical Digest, Volume 32, Number 3, 2013.

23. Toet, A., and IJspeert, J.K., “Perceptual evaluation of different image fusion schemes,” Displays 24.1, 25-37, 2003.

24. Blum, R. S., and Liu, Z., “Multi-Sensor Image Fusion and Its Applications,” CRC press, 2005.

25. SESAR 2011., “Modernising the European Sky,” Brussels, Belgium: SESAR Consortium.

26. ICAO., “Doc 9750 – Global Air Navigation Capacity & Efficiency Plan 2013-2028,” 4 ed. Montreal, Quebec, Canada: International Civil Aviation Organization (ICAO). 2014a. 27. SESAR and the Environment. Brussels, Belgium: European

Commission. EU 2014. Clen Sky 2 - Joint Technical Programme. Brussels, Belgium: European Commission. EU 2010.

28. Sabatini, R., Gardi, A., Ramasamy, S., Kistan, T., and Marino, M., “Novel ATM and Avionic Systems for Environmentally Sustainable Aviation Conference: Practical Responses to Climate Change (PRCC) 2014,” Engineers Australia Convention 2014, At Melbourne, Australia, Volume: Transport (Aviation)• November 2014 DOI: 10.13140/2.1.1938.0808

29. Sabatini, R., “A Roadmap for Future Aviation Research in Australia: Improving Aviation Safety, Efficiency and Environmental Sustainability,” Conference: First International Symposium on Sustainable Aviation (ISSA 2015), At Istanbul, Turkey, DOI: 10.13140/RG.2.1.4833.2962

30. Ramasamy, S., Sabatini, R., and Gardi, A., “Novel Flight Management System for Improved Safety and Sustainability in the CNS+A Context, Conference: Integrated Communication,” Navigation and Surveillance Conference (ICNS 2015), At Herndon, VA, USA, DOI: 10.1109/ICNSURV.2015.7121225 31. Tagarev, T., and Ivanov, P., “Computational Intelligence in

Multi-Source Data and Information Fusion,” Issue: Information and Security, vol. 2, 1999.

32. Hall, D.L., “An Introduction to Multisensor Data Fusion,” Proceedings of the IEEE, vol. 85, issue 1, 1997.

33. Cappello, F., Sabatini, R., Ramasamy, S., Marino, M., “Particle Filter based Multi-sensor Data Fusion Techniques for RPAS Navigation and Guidance,” Conference: Second International Workshop on Metrology for Aerospace (MetroAeroSpace 2015), At Benevento, Italy, 2015.

34. Wan, Eric A., and Van Der Merwe, R., “The Unscented Kalman Filter for Nonlinear Estimation”, Adaptive Systems for Signal Processing, Communications, and Control Symposium 2000. AS-SPCC. The IEEE 2000. IEEE, 2000.

35. Van Der Merwe, R., “Sigma-point Kalman Filters for Probabilistic Inference in Dynamic State-space Models”, Dissertation, Oregon Health & Science University, 2004. 36. Oh, S-M., “Nonlinear Estimation for Vision-Based Air-to-Air

Tracking”, Diss. Georgia Institute of Technology, 2007. 37. Challa, S., “Fundamentals of Object Tracking”, Cambridge

(12)

Page 11 of 11

38. Doucet, A., & Johansen, A. M. (2009). “A Tutorial on Particle Filtering and Smoothing: Fifteen Years Later”, Handbook of Nonlinear Filtering, 12, 656-704.

39. Hussain, M.S., “Real-Coded Genetic Algorithm Particle Filters for High-Dimensional State Spaces”, (Doctoral dissertation, UCL (University College London)), 2014.

40. Lai, C.K., Lone, M., Thomas, P., Whidborne, J., Cooke, A., “On-Board Trajectory Generation for Collision Avoidance in Unmanned Aerial Vehicles”, Proceedings of the IEEE Aerospace Conference, pp. 1-14, 2011. DOI: 10.1109/AERO.2011.5747526

41. Angelov, P., “Sense and Avoid in UAS: Research and Applications,” John Wiley & Sons, 2012.

42. Cheng, Q. and Bondon, P., “A New Sampling Method In Particle Filter,” 17th

European Signal Processing Conference (EUSIPCO), 2009.

43. Parseh, M-J., Pashazadeh. S., and Seyedarabi, H., “Improved Particle Filtering Algorithm for Maneuvering Object Tracking Using Deformation Detection,” 2nd International Conference on Computer and Knowledge Engineering (ICCKE), October 18-19, 2012.

44. Mohri, M., Rostamizadeh, A., and Talwalkar, A., “Foundations of Machine Learning,” The MIT Press, ISBN: 9780262018258, 2012.

45. Bengio, Y., Courville, A., and Vincent, P., “Representation Learning: A Review and New Perspectives,” IEEE Trans Pattern Anal Mach Intell. 2013 Aug; vol. 35, issue 8, pp. 1798-828, DOI: 10.1109/TPAMI.2013.50.

46. Anthony, M. H. G., and Biggs, N., “Computational Learning Theory,” Cambridge University Press, 1997.

47. Welling, M. A, “First Encounter with Machine Learning,” Irvine, CA.: University of California, 1-93, 2011.

48. Cosenzo, K. A., Fatkin, L. T., and Branscome, T. A. “Cognitive Uncertainty and Work Shifts in a Real-World Multi-Task Environment (No. ARL-TR-3515),” Army Research Lab Aberdeen proving Ground MD: Human Research and Engineering Directorate, 2005.

49. Shahbazian, E., Rogova, G., and DeWeert, M. J. “Harbour Protection through Data Fusion Technologies,” Springer Science & Business Media, 2009.

50. Konopko, K., and Janczak, D., “Maneuver Detection Algorithm Based on Probability Density Function Estimation with Use of RBF and HRBF Neural Network,” In Photonics Applications in Astronomy, Communications, Industry, and High-Energy Physics Experiments II (pp. 625-629). International Society for Optics and Photonics, 2004.

51. Cuadrado-Gallego, J.J., Braungarten, R., Dumke, R.R., Abran, A. (Eds.), “Software Process and Product Measurement International Conference,” IWSM–MENSURA 2007, Palma de Mallorca, Spain, November 5-8, 2007.

52. Demmel, S., Gruyer, D., & Rakotonirainy, A., “V2V/V2I Augmented Maps: State-Of-The-Art and Contribution to Real-Time Crash Risk Assessment,” In Proceedings of 20th Canadian Multidisciplinary Road Safety Conference, The Canadian Association of Road Safety Professionals, Ontario, Canada, 2010.

53. Dontabhaktuni Jayakumar, D. P., & Arunakumari, D. “Satellite Image Fusion in Various Domains,” International Journal of Advanced Engineering, Management and Science (IJAEMS), vol. 1, issue 3, June 2015.

54. Garzelli, A., Aiazzi, B., Baronti, S., Selva, M., and Alparone, L. “Hyperspectral Image Fusion,” In Proceedings of the Hyperspectral 2010 Workshop (pp. 17-19), 2010.

55. Pramanik, S., & Bhattacharjee, D. “Multi-sensor Image Fusion Based on Moment Calculation,” 2nd IEEE International Conference on Parallel Distributed and Grid Computing (PDGC), pp. 447-451, December 2012.

56. Toet, A., “Cognitive Image Fusion and Assessment,” INTECH Open Access Publisher, 2011.

57. Toet, A., & Hogervorst, M., “Real-time Full Color Multiband Night Vision,” INTECH Open Access Publisher, 2010.

58. Stathaki, T. “Image Fusion: Algorithms and Applications,” Academic Press, 2011.

References

Related documents

Executing all steps of the object thinning requires ve custom image processing instructions ( LOWPASS , CLEAR , HISTOGRAM , THRESHOLD , and SKELETON ) and. the three

The purpose of this population-based multi-round cross-sectional study is to estimate the change in treatment contact coverage as a result of implementing a district level

We have received numerous calls regarding the “every three months” supervisory visit between the Physical Therapist of record and the Physical Thera- pist Assistant delivering care

This research was conducted among 70 dentistry students at Kashan university of Medical Sciences, in a course called English for Dental Purposes (EDP) during the first semester of

If the professor and the Dean of Students agree that the evidence is sufficient to warrant a charge of academic dishonesty, and if it is determined this is the second violation, the

Hence, the equilibrium of a decentralized system results in too low levels of local public investments compared to the social optimum, even though median voters delegate policy

Connecting Solutions The aluminium assembly system with unlimited construction possibilities....