(ProQuest: ... denotes non-US-ASCII text omitted.)
Zheng Dou 1 and Xiaochun Xu 1 and Yun Lin 1 and Ruolin Zhou 2
Academic Editor:Zhijun Zhang
1, College of Information and Communication Engineering, Harbin Engineering University, Room 148, Building 21, 145 Nantong Street, Harbin, Heilongjiang 150001, China
2, Department of Electrical and Computer Engineering, Western New England University, Springfield, MA, USA
Received 20 January 2014; Accepted 2 April 2014; 23 April 2014
This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
1. Introduction
Information fusion is a useful technique to integrate heterogeneous data from different information sources. By increasing comprehensiveness while decreasing uncertainty of information, information fusion can be used to improve the quality of decision using the redundancy and complementariness of different information sources. As one of the most important methods in information fusion, the Dempster-Shafer evidence theory (D-S theory) [1, 2], which is an improvement of the Bayesian theory, has been widely used in information systems [3-11]. A significant improvement of the D-S approach over traditional probabilistic approach is that it allows for the allocation of a probability mass to sets or intervals and can handle both stochastic uncertainty and subjective uncertainty. The D-S evidence theory is a flexible and powerful mathematical tool for handling uncertain, incomplete, and imprecise information for at least the following three reasons. Firstly, by representing the imprecision and uncertainty of a body of knowledge via the notion of evidence, belief can be committed to a singleton or a compound set. Secondly, the evidence combination rule of the D-S evidence theory can provide an interesting operator to integrate multiple information acquired from different data sources. Finally, the decision on optimal hypothesis choice can be made in a rational and flexible manner.
In drying industry process, the supervision on working conditions of sensors is very important and difficult, and its role is to detect, locate, and isolate the fault sensor as quickly and accurately as possible. But, due to the complexity of sensor and the uncertainty of working environment, the monitoring data is usually uncertain, incomplete, or imprecise, which leads to the reduction of accuracy rate. Therefore, in this paper, a two-layer information fusion structure based on BP Neural Network and D-S evidence fusion method was presented for the supervision on working conditions of sensor in drying industry process. Firstly, according to the monitoring data obtained from different sensor sources, BP Neural Network was used to establish the basic belief assignment function of evidence for every single sensor source. Then, the D-S evidence combination rule was used to fuse those evidences. Finally, according to the fusion result, the working conditions of sensor could be described effectively and accurately. In this fusion process, on the one hand, the BP Neural Network could provide the ability of self-learning, self-adaptation, and fault tolerance; on the other hand, the D-S evidence method could express and handle the uncertain, incomplete, and imprecise information. Therefore, this method could further improve the accuracy and robustness of sensor monitoring system, which was proved by the numerical simulation results.
2. Preliminaries
2.1. Dempster-Shafer Evidence Theory
The mathematical basis of evidence theory, which was introduced by Dempster [1] and extended by Shafer [2], pays attention to the question of belief in the proposition systems. "Belief" in a proposition is not the same as the "chance" of the proposition being true. When forming propositions, evidence can be considered as a similar way, and the Dempster-Shafer (D-S) theory pays attention to "evidence," "weights of evidence," and "belief in evidence." Obviously, the belief structure in the evidence theory conforms with the Bayesian Probability Model [2], and thus the evidence theory can be viewed as a generalization and improvement of the classic probability theory. Because of its ability in dealing with uncertainty and imprecision problems, the D-S theory can be widely used in many fields [3-11]. Formally, the evidence theory concerns with the following preliminary notations.
Framework of Discernment. Firstly, evidence theory supposes a set of hypotheses θ as the framework of discernment, which can be defined as follows: [figure omitted; refer to PDF] where the set θ is composed of N exclusive and exhaustive hypotheses. In this paper, it represents the temperature sensors. Assume the power set P(θ) is composed of the 2N propositions of θ as follows: [figure omitted; refer to PDF] where ∅ denotes the empty set. Then, the subset containing only one element is called singleton.
Mass Functions, Focal Elements, and Kernel Elements. When the framework of discernment is determined, the mass function m can be defined as a mapping of the power set P(θ) to a number between 0 and 1, which is shown as follows: [figure omitted; refer to PDF] and it also satisfies the following conditions: [figure omitted; refer to PDF]
The mass function m is also called the basic probability assignment (BPA) function, and m(A) represents the proportion of all relevant and available evidences that supports the claim that a particular element of θ belongs to the set A but not to a particular subset of A . Any subset A of P(θ) satisfying that m(A)>0 is called a focal element,and C=...m(A)...0;0 A is called a kernel element of mass function m in θ .
Belief and Plausibility Functions . The belief function Bel is defined as [figure omitted; refer to PDF]
The plausibility function Pl is defined as [figure omitted; refer to PDF]
The belief function Bel(A) measures the total amount of probability that must be distributed among the elements of A . It reflects inevitability and signifies the total degree of belief of A , which constitutes a lower limit function on the probability of A . On the other hand, the plausibility function Pl(A) measures the maximal amount of probability that can be distributed among the elements of A , which describes the total belief degree related to A and constitutes an upper limit function on the probability of A . The relationship between Bel(A) and Pl(A) is shown in Figure 1, and the interval [Bel(A) , Pl(A) ]is named as belief interval.
Figure 1: Schematic diagram of Bel(A) and Pl(A) .
[figure omitted; refer to PDF]
Rule of Evidence Combination . Suppose m1 and m2 are two mass functions formed based on the information from two different information sources in the same frame of discernment θ and that Dempster's rule of combination (also called orthogonal sum), noted by m=m1 [ecedil]5;m2 , is the first one within the framework of evidence theory which can combine two BPA m1 and m2 to yield a new BPA: [figure omitted; refer to PDF] where k represents a basic probability mass associated with conflicts among the sources of evidence. Here, k can be determined by summing the products of mass functions of all pairwise sets without intersection and it is often interpreted as a measure of conflict between the data sources. The larger the value of k is, the more conflicting the sources are and the less informative their combination is.
2.2. BP Neural Network Theory
The BP Neural Network [12, 13] is one of the most important and popular techniques in the field of Neural Network, and it is also a kind of supervised learning neural networks, the principle behind which involves using the steepest gradient descent method to get any small approximation. A general model of the BP is shown in Figure 2.
Figure 2: Structure of the BP Neural Network.
[figure omitted; refer to PDF]
In Figure 2, there are three layers in BP Neural Network (BPNN): input layer, hidden layer, and output layer. Two nodes of each pair of adjacent layers are directly connected, to form a link. Each link has a weighted value representing the correlation between two nodes. Assuming there are n input neurons, then the weighted values can be updated using a training process described by the following equations in two steps.
(1) Hidden Layer Stage. The outputs of all neurons in the hidden layer can be calculated as follows: [figure omitted; refer to PDF] where vij are the weights of neurons, netj is the activation value of the jth node, yj is the output of the hidden layer, and fH is called as the activation function of a node, which is usually a sigmoid function described as follows: [figure omitted; refer to PDF]
(2) Output Layer Stage. The outputs of all neurons in the output layer can be calculated as follows: [figure omitted; refer to PDF] where ωjk are the weights value of output and fO is the activation function, which is usually a line function. All weights are initially assigned with random values and modified by the delta rule according to the data of learning samples.
3. The Fault Detection Model Based on D-S Evidence Theory
3.1. Detection Model of Sensor Fault
As discussed above, the D-S evidence theory has a strong ability to deal with uncertain, incomplete, and imprecise information. However, there is no general method to calculate BPA in D-S evidence theory. Therefore, in this paper, a structure of three layers is proposed to detect, locate, and isolate the fault sensor, which is shown in Figure 3.
Figure 3: Detection model of sensor fault.
[figure omitted; refer to PDF]
The first layer is data layer, which is used to gather and acquire data. Here, it is supposed that there are N sensors for supervising the drying industry process.
The next two layers are very important, so they are described in detail in the following parts.
3.2. Description of Data Layer
The second layer is data-fusion layer, which is also called data preprocessing. In this part, BP Neural Network is used to get BPA of evidence, because it has many advantages such as robustness for uncertain model, strong matching for nonlinear model, short training period, high accuracy of values, and easily-adjusted network. The data-fusion layer is a two-input and one-output process. Two-input is the supervising data provided by sensors i and j , and one-output is mij =({OKi ,OKj }) , abbreviated as mij =(OK) , which means "the sensors i and j are working well." If the number of sensors is N , and the output number of BP Neural Network is CN2 .
3.3. Description of Decision Layer
The third layer has actually united different frameworks of discernment. The inputs are the outputs of the first layer, and the prior knowledge acquired from different sensors is used to calculate the evidence on the different frameworks of discernment. However, according to the requirements of evidence theory, the combination rule is true only under the unified frameworks of discernment. Therefore, in the second layer, the different frameworks of discernment must be united. As well known, it is possible to combine the two evidences within the different frameworks of discernment θ and θ[variant prime] , because they are compatible. In order to combine and merge these evidences, the relationships between θ and θ[variant prime] must be defined. There are two operations, refinement and coarsening [14-16], which can express the correspondences in the form of compatibility rules. In this paper, the BPAs defined on different framework of discernment are united into a common framework of discernment by the refinement operation, and the BPA of each sensor defined on its own framework of discernment is calculated by the coarsening operation. In fact, a refinement operation unifies compatible elements of θ[variant prime] to an element of θ , and a coarsening operation is the antagonist relation.
A basic probability assignment of sensor Si is defined on the set θi ={OKi ,KOi } , where OKi means "sensor i is working well" and KOi means "sensor j is faulty." Meantime, this framework of discernment is defined as θij , which is the Cartesian product of θi and θj : [figure omitted; refer to PDF]
Therefore, mij is defined on the set of θij : [figure omitted; refer to PDF] where d(Vi ,Vj ) represents a normalized distance between sensor data Vi and Vj . Of course, the function d , such as residual generation methods or multivariate statistical methods, can also be used.
Supposing Rk is the refinement operation from θij to θijk and θijk is the Cartesian product of θij and θik , then the combination rule of mij and mik is intersection operation rule described as follows: [figure omitted; refer to PDF]
Supposing there are M sensors in practice, then there are CM2 outputs data preprocessed by BPNN. When it is refinement operation, the two different frameworks of discernment must be compatible. For example, mij and mik can be refined, but mij and mkl cannot be refined. Therefore, there are in total CM2CM-22 /2(M>3) kinds of combination modes which cannot be refined.
Supposing Si is the coarsening operation from θijk to θi and Nij and Nik are the kernels of mij and mik , ∃(A,B)∈(Nij ,Nik ) , then [figure omitted; refer to PDF] where D∈2θi and [figure omitted; refer to PDF]
Because the combination operation performs all intersections within focal elements of each refined belief assignment, it must guarantee that all possible intersection operations should have been done. The intersection operations between evidence sources can be expressed only needing to declare the reference set of the corresponding belief assignments, so it is very easy to add new evidence sources (new sensors) without affecting existing functions.
Now, it is possible to get the belief interval of each sensor by (5) and (6), which can be described as belief interval [Bel(OKi ),Pl(OKi )] . And then, the new belief interval [Bel(OKi ),Pl(OKi )]combine can be calculated by using (8). Thus, the working state of every sensor can be known.
4. Numerical Simulation Analysis in Drying Industry Process
Supposing there are four temperature sensors θ={H1 ,H2 ,H3 ,H4 } to supervise the drying process and the sampling data of every single sensor is the input of BP Neural Network, then there are C42 =6 outputs, which are m12 =(OK) , m13 =(OK) , m14 =(OK) , m23 =(OK) , m24 =(OK) , and m34 =(OK) . Because of the limitation of space, the preprocessing result of BP Neural Network and part of intermediate result will be given directly in Table 1 without detailed description.
Table 1: Fault preprocessing result of BP Neural Network.
| m 12 ( A ) | m 13 ( A ) | m 14 ( A ) | m 23 ( A ) | m 24 ( A ) | m 34 ( A ) |
Fault decision result | 0.215 | 0.213 | 0.219 | 0.946 | 0.281 | 0.276 |
According to the discussion in Sections 3.2 and 3.3, there are C62 =15 kinds of combination modes, but C42C4-22 /2=3 kinds of combination modes cannot be refined. Therefore, there are 15-3=12 kinds of combination modes that can be refined. Furthermore, according to the method presented in Section 3.3, each two preprocessing results are firstly operated by refinement, then the refinement results are operated by coarsening to get the BPA of each sensor, and finally those BPAs are fused by using evidence theory combination rule to get the new intervals [Bel(OK),Pl(OK) ] shown in Table 2.
Table 2: Fault decision result of D-S evidence fusion.
Combination mode | Belief interval of evidence | |||
[Bel1 Pl1 ] | [Bel2 Pl2 ] | [Bel3 Pl3 ] | [Bel4 Pl4 ] | |
m 12 [ecedil]5; m 13 | [0.306 0.183] | [0.278 0.948] | [0.352 0.753] |
|
m 12 [ecedil]5; m 14 | [0.382 0.348] | [0.244 0.818] |
| [0.244 0.303] |
m 12 [ecedil]5; m 23 | [0.201 0.156] | [0.193 0.947] | [0.284 0.788] |
|
m 12 [ecedil]5; m 24 | [0.397 0.282] | [0.910 0.896] |
| [0.215 0.240] |
m 13 [ecedil]5; m 14 | [0.188 0.258] |
| [0.749 0.828] | [0.278 0.232] |
m 13 [ecedil]5; m 23 | [0.489 0.427] | [0.268 0.872] | [0.291 0.920] |
|
m 13 [ecedil]5; m 34 | [0.293 0.374] |
| [0.810 0.865] | [0.417 0.348] |
m 14 [ecedil]5; m 24 | [0.201 0.338] | [0.881 0.948] |
| [0.263 0.361] |
m 14 [ecedil]5; m 34 | [0.352 0.400] |
| [0.936 0.979] | [0.338 0.374] |
m 23 [ecedil]5; m 24 |
| [0.834 0.750] | [0.953 0.848] | [0.489 0.427] |
m 12 [ecedil]5; m 13 |
| [0.865 0.876] | [0.809 0.750] | [0.415 0.352] |
m 12 [ecedil]5; m 13 |
| [0.822 0.920] | [0.788 0.850] | [0.395 0.343] |
Evidence fusion results | [0.023 0.041] | [1.000 1.000] | [1.000 1.000] | [0.067 0.079] |
| ||||
Fault diagnosis results | Sensors 1 and 2 are fault, Sensors 3 and 4 are not working well. |
According to Table 2, through analyzing every pairwise combination, it can be concluded that the difference of upper and lower limits of belief interval [Bel(OK),Pl(OK) ] is very significant, which shows the great uncertainty of sensor state. Therefore, the results of those pairwise combinations cannot be used to decide the sensor state.
In order to fully exploit the maximum of available information and reduce the uncertainty of sensor state, the results of those pairwise combinations should be further fused by evidence combination rule. The new interval [Bel(OK),Pl(OK) ] is shown as the font of bold italic in Table 2.
For example, the fusion result of sensor one is [0.026, 0.041]; that is to say, the sum of belief degree of all pieces of evidence that precisely support the proposition "sensor one is working well" is 0.026 and precisely support the proposition "sensor one is fault" is 1-0.026=0.974 , so the fusion result indicates that "sensor one is fault." Similarly, the sum of belief degree of the proposition "sensor two is working well" is 1.0000 and that of the proposition "sensor two is fault" is 1-1.0000=0.0000 , so the fusion result indicates that "sensor two is working well." To sum up, sensors one and four are fault, and sensors two and three are working well.
5. Conclusions
In this paper, a modular and generic framework for multiple fault detection and isolation of sensors was presented with a two-layer structure. In data layer, through fully exploiting the sensor data, the data preprocessing could be realized by BP Neural Network to overcome the disadvantages caused by the change of input sensor data and calculate the BPA of evidence. In decision layer, a modular and generic framework of sensor network for multiple fault detection was presented, and the different but compatible frameworks of discernment were united without affecting the existing relationship, according to the refinement and coarsening operations. Furthermore, new evidence (sensor) could be added to the process very easily and effectively. Therefore, our method had much better expandability, modularity, and flexibility.
After combining all sensors with combination rule, the total uncertainty of sensor state was greatly reduced and the fault sensors could be further exploited by the final fusion results. Numerical simulation results also proved that our new method could be used in practice.
However, if the number of sensors is larger and the combination modes are more, there will be a heavy computation burden. Therefore, this method should be further simplified and optimized in the further study.
Acknowledgments
This work was supported by the Nation Nature Science Foundation of China (nos. 61301095 and 61201237), Nature Science Foundation of Heilongjiang Province of China (no. QC2012C069), and the Fundamental Research Funds for the Central Universities (nos. HEUCF130810 and HEUCF130817).
Conflict of Interests
All the authors declared that there was no conflict of interests regarding the publication of this paper.
[1] A. P. Dempster, "Upper and lower probabilities induced by a multivalued mapping," Annals of Mathematical Statistics , vol. 38, pp. 325-339, 1967.
[2] G. Shafer A mathematical Theory of Evidence , Princeton University Press, Princeton, NJ, USA, 1976.
[3] L. Dymova, P. Sevastjanov, "An interpretation of intuitionistic fuzzy sets in terms of evidence theory: decision making aspect," Knowledge-Based Systems , vol. 23, no. 8, pp. 772-782, 2010.
[4] Y. Deng, F. T. S. Chan, "A new fuzzy dempster MCDM method and its application in supplier selection," Expert Systems with Applications , vol. 38, no. 8, pp. 9854-9861, 2011.
[5] X. Y. Deng, Q. Liu, Y. Hu, Y. Deng, "Topper: topology prediction of transmembrane protein based on evidential reasoning," The Scientific World Journal , vol. 2013, 2013.
[6] X. Y. Su, Y. Deng, S. Mahadevan, Q. L. Bao, "An improved method for risk evaluation in failure modes and effects analysis of aircraft engine rotor blades," Engineering Failure Analysis , vol. 26, pp. 164-174, 2012.
[7] D. Zhu, W. Gu, "Sensor fusion in integrated circuit fault diagnosis using a belief function model," International Journal of Distributed Sensor Networks , vol. 4, no. 3, pp. 247-261, 2008.
[8] R. Feng, S. Che, X. Wang, N. Yu, "A credible routing based on a novel trust mechanism in ad hoc networks," International Journal of Distributed Sensor Networks , vol. 2013, 2013.
[9] F. Browne, N. Rooney, W. Liu, D. Bell, H. Wang, P. S. Taylor, Y. Jin, "Integrating textual analysis and evidential reasoning for decision making in engineering design," Knowledge-Based Systems , vol. 52, pp. 165-175, 2013.
[10] D. Niu, Y. Wei, Y. Shi, H. R. Karimi, "A novel evaluation model for hybrid power system based on vague set and Dempster-Shafer evidence theory," Mathematical Problems in Engineering , vol. 2012, 2012.
[11] Y. Zhao, J. Li, L. Li, M. Zhang, L. Guo, "Environmental perception and sensor data fusion for unmanned ground vehicle," Mathematical Problems in Engineering , vol. 2013, 2013.
[12] Z. Yudong, W. Lenan, "Stock market prediction of S&P 500 via combination of improved BCO approach and BP neural network," Expert Systems with Applications , vol. 36, no. 5, pp. 8849-8854, 2009.
[13] H. Azami, M.-R. Mosavi, S. Sanei, "Classification of GPS satellites using improved back propagation training algorithms," Wireless Personal Communications , vol. 71, no. 2, pp. 789-803, 2013.
[14] F. Janez, A. Appriou, "Theory of evidence and non-exhaustive frames of discernment: plausibilities correction methods," International Journal of Approximate Reasoning , vol. 18, no. 1-2, pp. 1-19, 1998.
[15] J.-P. Steyer, L. Lardon, O. Bernard, "Sensors network diagnosis in anaerobic digestion processes using evidence theory," Water Science and Technology , vol. 50, no. 11, pp. 21-29, 2004.
[16] W. Haitao, L. Qun, Z. Qiji, "Information fusion of neural networks and evidence theory in fault diagnosis of equipments," Computer Engineering and Applications , vol. 22, pp. 13-219, 2004.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Copyright © 2014 Zheng Dou et al. Zheng Dou et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Abstract
Due to the complexity and dangerousness of drying process, the fault detection of temperature sensor is very difficult and dangerous in actual working practice and the detection effectiveness is not satisfying. For this problem, in this paper, based on the idea of information fusion and the requirements of D-S evidence method, a D-S evidence fusion structure with two layers was introduced to detect the temperature sensor fault in drying process. The first layer was data layer to establish the basic belief assignment function of evidence which could be realized by BP Neural Network. The second layer was decision layer to detect and locate the sensor fault which could be realized by D-S evidence fusion method. According to the numerical simulation results, the working conditions of sensors could be described effectively and accurately by this method, so that it could be used to detect and locate the sensor fault.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer