6
APPLICATION OF SUPPORT VECTOR MACHINE BASED FAULT DIAGNOSIS Ming Ge 1 , Guicai Zhang, Ruxu Du and Yangsheng Xu Department of Automation and Computer-Aided Engineering The Chinese University of Hong Kong Hong Kong Abstract: The fault diagnosis is important in continuously monitoring the performance and quality of manufacturing processes. Overcoming the drawbacks of threshold approach, artificial neural network may extract the symptom of the faults through learning from the samples, but it is difficult to design its structure. Moreover, it needs a large numbers of samples in practice. In this paper, support vector machine approach was proposed to overcome these limitations based on statistics learning theory, and a new fault diagnosis system is developed. The experimental results showed that it is an efficient and practical on-line intelligent monitoring system for the stamping processes. Copyright © 2002 IFAC Keywords: Machine Learning, Intelligent Knowledge-based System, Fault Diagnosis, Pattern Identification, Manufacturing Processes 1 Please address the corresponds to this author at [email protected] 1. INTRODUCTION The detection and isolation of the faults in manufacturing processes is of great practical significance. The quick and correct fault diagnosis avoids production loss and product degradation. Manufacturing processes usually involve many process variables that interact with one another. When any of these variables deviate beyond their specified limits, a fault occurs. The early approaches of fault detection are based on the threshold checking. Later, pattern classification and model identification were used. Currently, a large number of approaches are available such as Machine Learning (ML) and statistics, see (Gertler, 1998). From a practical point of view, the threshold method depends on the priori experience. Possessing the intelligence, ML overcomes this drawback. From a mathematical point of view, ML is to estimate a function f using n sets of training samples. Ideally, f shall correctly predict the unseen examples. Traditionally, the classical ML methods are based on the Empirical Risk Minimization (ERM) principle (Cortes and Vapnik, 1995). Unfortunately, it has been shown that the ERM cannot ensure the minimization of the actual risk. Moreover, the complexity of the learning machine has relationship with not only the system to be studied but also the amount of training samples. The case of the over- fitting of Artificial Neural Network (ANN) gives an appropriate example to reveal the failure of ERM principle. Vapnik (1995) found two correlated factors may lead to the failure: the ANN insufficient training sample and unreasonable structure design. As to fault diagnosis system of manufacturing processes, lacking enough training samples is common because there is shortage of data during the available time before failure. As an example, let us consider the sheet metal stamping processes. It involves transient elastic and plastic deformation of the sheet metals as well as the static and dynamic behaviour of the press. Nearly 40 process variables would effect the production and the product. Aiming at meeting the ever-increasing demands on the quality and productivity, the work on the development of the intelligent fault diagnosis system was carried out, see (Ge, et al. 2001; Koh, et al., 1999). However, the same problems were met as described above. It is necessary to develop a learning machine with reasonable structure and the capacity dealing with small training sample size. For certain simple type of algorithms, statistical learning theory can rather precisely identify the factors that need to be taken into account to learn successfully. Springing from the statistical learning theory, support vector machine (SVM) makes the theory not only a tool for the theoretical analysis but also a tool for creating practical algorithms for estimating multidimensional functions. SVM has Copyright © 2002 IFAC 15th Triennial World Congress, Barcelona, Spain

application of support vector machine based fault diagnosis

  • Upload
    haquynh

  • View
    216

  • Download
    2

Embed Size (px)

Citation preview

Page 1: application of support vector machine based fault diagnosis

APPLICATION OF SUPPORT VECTOR MACHINE BASED FAULT DIAGNOSIS

Ming Ge1, Guicai Zhang, Ruxu Du and Yangsheng Xu

Department of Automation and Computer-Aided Engineering The Chinese University of Hong Kong

Hong Kong

Abstract: The fault diagnosis is important in continuously monitoring the performance and quality of manufacturing processes. Overcoming the drawbacks of threshold approach, artificial neural network may extract the symptom of the faults through learning from the samples, but it is difficult to design its structure. Moreover, it needs a large numbers of samples in practice. In this paper, support vector machine approach was proposed to overcome these limitations based on statistics learning theory, and a new fault diagnosis system is developed. The experimental results showed that it is an efficient and practical on-line intelligent monitoring system for the stamping processes. Copyright © 2002 IFAC Keywords: Machine Learning, Intelligent Knowledge-based System, Fault Diagnosis, Pattern Identification, Manufacturing Processes

1 Please address the corresponds to this author at [email protected]

1. INTRODUCTION The detection and isolation of the faults in manufacturing processes is of great practical significance. The quick and correct fault diagnosis avoids production loss and product degradation. Manufacturing processes usually involve many process variables that interact with one another. When any of these variables deviate beyond their specified limits, a fault occurs. The early approaches of fault detection are based on the threshold checking. Later, pattern classification and model identification were used. Currently, a large number of approaches are available such as Machine Learning (ML) and statistics, see (Gertler, 1998). From a practical point of view, the threshold method depends on the priori experience. Possessing the intelligence, ML overcomes this drawback. From a mathematical point of view, ML is to estimate a function f using n sets of training samples. Ideally, f shall correctly predict the unseen examples. Traditionally, the classical ML methods are based on the Empirical Risk Minimization (ERM) principle (Cortes and Vapnik, 1995). Unfortunately, it has been shown that the ERM cannot ensure the minimization of the actual risk. Moreover, the complexity of the learning machine has relationship with not only the system to be studied but also the amount of training samples. The case of the over-fitting of Artificial Neural Network (ANN) gives an

appropriate example to reveal the failure of ERM principle. Vapnik (1995) found two correlated factors may lead to the failure: the ANN insufficient training sample and unreasonable structure design. As to fault diagnosis system of manufacturing processes, lacking enough training samples is common because there is shortage of data during the available time before failure. As an example, let us consider the sheet metal stamping processes. It involves transient elastic and plastic deformation of the sheet metals as well as the static and dynamic behaviour of the press. Nearly 40 process variables would effect the production and the product. Aiming at meeting the ever-increasing demands on the quality and productivity, the work on the development of the intelligent fault diagnosis system was carried out, see (Ge, et al. 2001; Koh, et al., 1999). However, the same problems were met as described above. It is necessary to develop a learning machine with reasonable structure and the capacity dealing with small training sample size. For certain simple type of algorithms, statistical learning theory can rather precisely identify the factors that need to be taken into account to learn successfully. Springing from the statistical learning theory, support vector machine (SVM) makes the theory not only a tool for the theoretical analysis but also a tool for creating practical algorithms for estimating multidimensional functions. SVM has

Copyright © 2002 IFAC15th Triennial World Congress, Barcelona, Spain

Page 2: application of support vector machine based fault diagnosis

recently been gaining popularity due to two respects. First, it is quite satisfying from a theoretical point of view: support vector learning is based on some simple ideas and provides an obvious intuition of what learning from the examples is about. Second, it can lead to high performance in practical applications. Thus, SVM has been applied to many real-world problems, such as pattern identification, regression, even nonlinear equalization, etc., see (Gutta, et al., 2000; Trafalis and Ince, 2000; Sebald and Bucklew, 2000). For the newest development, reader is referred to the Internet site www.support-vector.net. Aiming to improve the fault diagnosis success rate and the response performance, in the paper, a new fault diagnosis system for the stamping processes using SVM approach was developed. The rest of this paper is structured in the following manner. In section 2, the principle of SVM on its classification mechanism was discussed, then, introduced the new fault diagnosis system. Section 3 addresses the details on the fault diagnosis system for the stamping processes and the experimental system setup and implementing. Then, the experimental results are presented in the Section 4. Finally, Section 5 contains the conclusions.

2. THEORTICAL BACKGROUND 2.1 The Principle of SVM The basic theory of SVM has been well understood, see (Cristianini and Taylor, 2000). Briefly, given a set of training samples (x1, y1), (x2, y2), …, (xn, yn), where, x∈RN denotes the input vector and y∈-1, 1 represents output class. The key ideas of the SVM is finding out a optimal hyperplane

bfy +⋅== xwwx ),( (1) where, w∈ RN and b∈ R are the weighting factors, f : RN –> -1,1, which can maximize the class margin not only reduces the empirical risk but also controls the generalization capacity. Using the soft margin concept, the problem is equivalent to solve the following optimization problem:

∑∑==

+=+⟩⋅⟨n

ii

n

iib

CC1

2

1,, 21

21min ξξ www

ξw (2)

s. t. ,1),( ii by ξ−≥+⟩⟨⋅ ixw ξi ≥ 0 i = 1,2,…, n where, C is a constant corresponding to the value ||w||2 and ξ is a variable giving a soft classification boundary. A more generalized form of SVM involves the use of a kernel function, K. The kernel function must satisfy the Mercer condition, such that

⟩⋅⟨= )()()( zxzx, ϕϕK (3)

where, x, z ∈ Ω, and the function ϕ is a mapping from Ω to a dot product space S. It can be shown that the above optimisation problem is equivalent to solve the quadratic optimization problem:

)(21)(min

1 1,ji x,xα KyyL jij

n

i

n

jiii ααα∑ ∑

= =−= (4)

s.t. ∑=

=n

iiiy

1

0 ≤ αi < C i = 1, 2, …, n After the unique solution α* are found, it is easy to acquire the corresponding weight factors. The optimal hyplanplane can be represented as:

*** ),(),,( bKybfSVi

ii += ∑∈

xxαx i* α (5)

The SVM is constructed by the classification decision function

))(sgn(),,( ** bKybgSVi

ii* += ∑

∈x,xαx i

* α (6)

SVM has a number of unique features. First, it has an exceptional ability to classify the unseen samples (x, y). The classification work is carried out in the kernel-induced feature space. In order to make the classes linear separable, normally, the inputs are mapped into a feature space if it is impossible to carry out the task in their original space. Replaced the dot products with the kernel , the inputs are mapped into a high dimensional feature space from their original space non-linearly. The type of the kernel decides the dimension in the feature space. For example, if adopting the polynomial kernel of order 2 for a d dimension vector, the dimension in the feature space is d(d+3)/2. Thus, it is possible to depict the characteristics of the input more accurately and uniquely using more information. In practice, various kernel functions can be employed as long as they comply with the Mercer condition, such as polynomial, radial basis function and sigmoid function, etc. In the meanwhile, the quadratic optimisation process guarantees the solution must be the global minima.

)( zx,K

Second, it does not require a large amount of training samples. Scholkopf, et al., (1999) pointed out that the actual risk of the learning machine, designed according to the ERM principle, consists of two parts: empirical risk of the training samples confidence interval expressed as:

( hnRR emp /)()( Φ+≤ ww ) (7) where, R(w) is actual risk, Remp(w) is empirical risk of the training samples, Φ is confidence interval, n is the amount of training sample and h is Vapnik-Chervonenkis (VC) dimension. It is noted that when n/h is small (for example less than 20, the training sample is small size), Φ has a little big value. Then, the performance is poor to represent R(w) with

Page 3: application of support vector machine based fault diagnosis

Remp(w). As a result, according to ERM principle, the big training sample size is a premise to acquire a satisfied learning machine. The support vector technique was first introduced by Cortes and Vapnik (1995) to solve the classification problems with small size training samples. SVM actually suggests a trade-off between the quality of the hyperplane and the complexity of the classifier with small training sample size because its formulation embodies the structural risk minimization principle as opposed to the ERM approach commonly employed in statistical learning. Therefore, SVM not only constructs an optimal hyperplane which has a strong generalization capacity taking into account the small size training sample, but also overcomes the overfitting and difficulties in structure design of the learning machine. Finally, after the training, SVM does not require complicated calculation and hence, can be done in real-time. The quadratic optimisation problem Equation 4 has a unique solution α*. According to Karush-Kuhn-Tucker complementarity conditions, the solution has to satisfy:

0]1)([ =−+⟩⋅⟨ byii ixwα (8) that implies only αi

* corresponding for inputs xi lying on the margin are non-zero, while the other ones are zero. These inputs are called support vector (SV), thus, only few of the points are involved in determining the optimal hyperplane Eq. 5 and the class decision Eq. 6, that reduces computational complexity and load. Moreover, the construction of the optimal hyperplane does not require extra computation load even the dimension in the feature space increases due to the introduction of kernel function. By replacing the dot product with an appropriately chosen kernel function, it is implicit that performing a non-linear mapping into a high dimensional feature space without increasing the number of tuneable parameters, the kernel function computes the dot product of the feature vectors corresponding to the two inputs. An appropriate kernel can make the system more efficient. Theoretically, it seems that polynomial is more suitable due to its simple calculation. All of these features make the SVM very attractive for condition monitoring and diagnosis in practice. 2.2 Fault Diagnosis System based on SVM SVM is designed for the classification of two classes. The fault diagnosis, however, has to deal with multi- type faults. Thus, the fault diagnosis system based on SVM could be designed with a hierarchical architecture illustrated as Figure 1. In the first layer, the corresponding SVM detects the fault occurring efficiently. If people only want the system to implement fault detection, the task then is finished by here. Otherwise, the real-time data will be transferred to the second layer in abnormal case for the fault isolation. The second layer’s SVM is

designed to identify only one particular fault from the others. Consequently, same operation was implemented on the following layers, each layer will isolate one kind of fault from the left. Therefore, the diagnosis procedure keeps towards till the type of fault is determined. SVM on each layer can be depicted as Figure 2

Fault detection Fault isolation

Fig. 1. The architecture of the fault diagnosis system

Fig. 2. The structure of a SVM 2.3 Fault Diagnosis System Implementation As a summary, to design a fault diagnosis system based on SVM, the procedure should be as follows: Step 1. Get training samples Ωi = ((x11, yi), …, (x1j, yi)) i = 1, 2, …, M; j = 1, 2, …, n where, Ωi ∈ Ω = Ωi | i : one kind of faults, j

is the amount of training samples in each kind of fault.

Step 2. for i = 1:M if i = 1 (layer 1, normal case), then yi = 1;

otherwise, yi = -1; end if i = 2 (layer 2, fault type A), then yi = 1;

otherwise, yi = -1; end …… if i = M-1 (layer M-1, the last fault type),

then yi = 1; otherwise, yi = -1; end end Step 3. Define a kernel function ; )( zx,K C could be determined by cross-validation. Step 4. On each layer in off-line, Equation 4 αi

* (αi* ≠ 0) → xi (corresponding

support vectors according to Equation 8) → acquiring b* (by an arbitrary SV).

Step 5. xi, αi*, b* ⇒ Equation 6 → ),,( *

i bg *αx

Step 6. → fault diagnosis system. ),,( *i bg *αx

Note: x → y: y can be acquired from x;

Page 4: application of support vector machine based fault diagnosis

x ⇒ y: substitutes x into y; To guarantee the good performance, the more frequently fault type occurs, the upper layer the corresponding SVM locates.

Same as the typical one, a fault diagnosis system for stamping process should have two capabilities: detection and isolation. In the meanwhile, the stamping processes are complicated and high-speed. A stroke is completed in a fraction of a second. Aiming to prevent the precise moulds and ensure the quality, the factor of real-time must be put on the first propriety of its fault diagnosis system except the accuracy of detection and isolation. Therefore, in the first layer of the system, the corresponding SVM needs not only efficiently to detect the fault occurring but also to inform the system to cease the operation immediately. Consequently, the fault type will be isolated by the following SVMs. As to the signal type, it is noted that an appropriate signal with high signal-to-noise ratio (SNR) not only represents the physical characteristic visibly but also can enhance the robust of the monitoring system. It is reported that tonnage signals and strain signals are the most commonly used ones for monitoring relying on its meaningful information and robust physical property, hence, the strain signal was employed in the works.

3. EXPERIMENT DESIGN AND IMPLEMENT 3.1 Experiment setup It is known that stamping operations can be divided roughly into three categories: blanking, bending and deep drawing. In this work, the experiment is carried out on a C-frame double column press (SEYI: SN1-25), and a kind of blanking operation was conducted. The maximum stamping force is 25 tons and the maximum speed is 110 Strokes Per Minute (SPM). The raw signals are acquired from a strain sensor (Kistler 9232A) mounted on the middle of the press frame. The signals were sampled using a National Instruments data acquisition card (PCI 4452), installed in a Pentium III 550MHz PC. Since stamping is a transient process, in order to capture the signal every time at the same instance, a trigger

Fig. 3. The experiment system

mechanism was developed. It is a proximity sensor installed near the main shaft. Every time when the key way of the shaft rotates passing the sensor, a pulse is generated and sent to computer to trigger the start of the sampling. The sampling rate was 5 kHz. To get rid of the meaningless data, in the meanwhile, ensure covering the entire stroke effectively, 512 data were drawn from the data acquired per stroke. Figure 3 shows the experimental system. 3.2 The patterns of different kinds of faults The part material is mild steel and the blank dimension (the perimeter of the part) is approximately 32 cm. The normal working condition is under 85spm with 1.1mm workpiece. When the working condition changes, called fault develops, some deviation occurs. Table 1 gives the conditions to be detected and isolated in the work. Figure 5 illustrates the strain signals under these conditions. It is noted that condition D and E had comparatively obvious variation among the conditions, while the others are similar. It is a tough job to diagnose the faults.

Table 1: Fault types for stamping processes

Fault Type Description -- No defect: workpiece thickness = 1.1 mm A Misfeed: workpiece is not aligned with

the dies B Slug: the slug(s) left on the surface of

sheet metal C Too thick: workpiece thickness = 1.2 mm D Too thin: workpiece thickness = 1.0 mm E Material missing the feeder can not

convey the materials forwards

Trigger

Fig. 5. The strain signal under different of conditions

Strain sensor

3.3 Experiments implementation From the above discussion, the SVs and optimal hyperplane and the decision function of the corresponding SVM will be acquired after provided some factors. Two major factors among these are the

Page 5: application of support vector machine based fault diagnosis

type of the kernel function and the amount of training samples. As to the first factor, the principle in the work is that the kernels need to ensure the accurate and efficient which are the basic requirements in fault diagnosis system. Vapink (1995) reported that the type of kernel employed by SVM applied on character recognition is inconsequential as long as the capacity is appropriate for the amount of training data and complexity of the classification boundary. In the work, three kinds of kernel, polynomial and RBF and sigmoid, as Equation 9 ~ 11 were applied:

pK )512/(),( zxzx ⋅= (9)

(10) )512/||exp(),( 22 σzxzx −−=K (11) )512/)(tanh(),( cvK −⋅= zxzx

As to the second factor, aiming to determine which amount of training samples is appropriate, the validation test has been taken. There are 240 samples with 40 ones in each kind of condition ready for use. Each sample is 512-dimension vector. Picking a serial of amount of samples of each kind of condition randomly act as training samples. The amount is from 1 to 11 with polynomial kernel (p=2). Figure 6 shows the classification success rate of these tests. It is noted that the success rate arrives 96.67%, a good result, when the amount of training sample in each kind of condition. After that, the oscillation is not obviously and the trend is better and better with the training sample increasing. It proves the significant advantage of the approach in industry application: small train sample size. The performance with 9 and 10 of training samples are same: 97.78%, and it is 98.33% for 11 case. However, the support vectors of the latter are 41 in total, while it is 32 for 9 case. To reduce the computational load and calculation time, in the later

experiments, the amount of training sample in each condition is set as 9.

Fig. 6. The results of the validation testes under

different amounts of training samples using polynomial kernel (p=2).

4. EXPERIMENT RESULTS Picking 9 samples of each kind of condition randomly act as training samples so that there are 54 training samples in total. The SVs will be chosen among these 54 training samples. In the meanwhile, the order p in the polynomial was tuned from 1 to 3, which is a parameter to control the capacity of the SVM. In theoretical, the greater p is, the more complex classification boundary the SVM has. The σ in RBF was tuned from 0.4 to 0.6, while, v was tuned from 1 to 3 with c in both 0 and 1 in sigmoid function. For each experiment, the training samples as well as the test samples are same. The results are represented in Table 2 and 3 and 4.

Table 2 Experimental results with polynomial kernel

Number of SV Number of mistake Order

SVM1 SVM2 SVM3 SVM4 SMV5 -- A B C D E Success

rate p=1 9 13 5 3 3 1 3 0 0 0 0 97.78% p=2 7 12 7 4 3 0 3 0 1 0 0 97.78% p=3 6 16 7 4 4 1 3 0 1 0 0 97.22%

Table 3 Experimental results with RBF kernel

Number of SV Number of mistake Order

SVM1 SVM2 SVM3 SVM4 SMV5 -- A B C D E Success

rate σ=0.4 11 15 10 8 5 1 0 0 0 0 0 99.44% σ=0.5 9 13 8 6 4 1 3 0 0 0 0 97.78% σ=0.6 10 14 8 5 4 1 3 0 0 0 0 97.78%

Table 4 Experimental results with sigmoid kernel (c = 0, 1)

Number of SV Number of mistake Order

SVM1 SVM2 SVM3 SVM4 SMV5 -- A B C D E Success

rate v=1 10 13 5 4 3 1 3 0 1 0 0 97.22% v=2 8 9 5 4 2 1 3 0 1 0 0 97.22% v=3 6 7 3 3 2 1 2 0 1 0 0 97.78%

Page 6: application of support vector machine based fault diagnosis

The results reflect the similarity of the strain signals under different conditions. Due to the obvious variation the conditions D and E can be isolated by the system without mistakes. However, some signals of the normal and material missing (A) conditions have been classified to the slug (C), as well as some signals of too thick condition have been classified to the normal in some experiments. From studying the relevant signals in time domain, it is found they were too similar and overlapping with each other so that they could not be separated even in the certain high dimensional feature space. In the meanwhile, it is noted the performance of success rate is robust on choosing the type of kernel function in classification. The variation of c value in the sigmoid kernel did not impact on its performance in the experiments. It is obvious that the numbers of SVs using sigmoid kernel are the least in the experiments. Its success rate is more similar to the polynomial kernel’s, but the former costs longer fault diagnosis time than the latter since it involves the exponential calculation. With RBF kernel it can acquire the best recognition result while the number of its SVs is the most in the experiments. Although the performance of the RBF case is slightly better than the polynomial case, the fault diagnosis time, for example, in the same worst case, of former (σ=0.6) is 220ms while the latter one (p=2) is 160ms, hence, polynomial kernel has a better trade-off point for on-line fault diagnosis task. Moreover, as a reference, the fault diagnosis task was carried out with the backpropagation (BP) approach, the typical one of traditional ML, using the same set of samples. Here, 2-layer BP was applied with different neurons in the first layer and different amount training samples because there are no standard principle to design the network. Table 5 showed the results. Obviously, complexity structure may bring better result, but it is still worse than the performance of SVM. When the training sample is not enough, the result is too bad to be accepted.

Table 5 Results of classification with BP approach Structure (1st - 2nd)

Amount training sample in each

condition

Success rate

20 - 6 30 90.00% 25 - 6 30 93.33% 25 - 6 10 88.89%

5. CONCLUSIONS Based on the discussions above, the following conclusions can be drawn: 1. Derived from the statistic learning theory, SVM

realizes SRM principle in its design. Different with other pattern recognition algorithms, it overcomes the local minima and overfitting, and it has a great generalization capacity. This makes the system intelligent to deal with complicated processes.

2. SVM nonlinearly maps the input space into a higher dimensional feature space via simple kernel representations. In the feature space, a

kernel function based classifier with maximum margin is constructed so that the classification capacity is optimized. These guarantee the system to carry out the tasks accurately.

3. An SVM classifier is determined only by a sparse set of support vectors, and these vectors are automatically selected from the training data during the learning process. Together with optimal α, b, and K, the efficient algorithm ensures on-line fault diagnosis application.

4. The perfect characteristics of the small training sample size and the robust performance on the parameters selection makes the fault diagnosis system more practical.

5. The fault diagnosis system using SVM has potential in application for the other manufacturing processes.

REFERENCES Cortes, C. and V. Vapnik (1995). Support Vector

Networks, Machine Learning, 20:273 – 297. Cristianini, N. and J. Shawe-Taylor (2000). A

Introduction to Support Vector Machines and other Kernel-based Learning Methods, Cambridge University Press, Cambridge, UK.

Ge, M., G.C. Zhang and Y. Xu (2001). Fault Diagnosis of Stamping Processes using Wavelet Packet and Neural Networks, CIMCA’2001, Las Vegas, USA.

Gertler, J.J. (1998). Fault Detection and Diagnosis in Engineering Systems, Marcel Dekker, Inc., New York.

Gutta, S., J.R.J. Huang, P. Jonathon, H. Wechsler and L.R. Rabiner (2000). Mixture of Experts for Classification of Gender, Ethnic Origin, and Pose of Human Faces, Neural Networks IEEE Transactions on, Vol. 11, pp. 948 –960.

Koh, C.K.H., J. Shi, W. J. Williams and J. Ni (1999). Multiple Faults Detection and Isolation Using the Haar Transform, Part 2: Application to the Stamping Process, Trans. of ASME Journal of Manufacturing Science and Engineering, Vol. 121, No. 2, pp. 295-299.

Scholkopf, B., C.J.C. Burges and A.J. Smola (1999). Advances in Kernel Methods: Support Vector Learning, MIT Press, Cambridge.

Sebald, D.J. and J.A. Bucklew (2000). Support Vector Machine Techniques for Nonlinear Equalization, IEEE Transactions on Signal Processing, Vol. 48, pp. 3217-3226.

Trafalis, T.B. and H. Ince (2000). Support Vector Machine for Regression and Applications to Financial Forecasting, IJCNN 2000, Proceedings of the IEEE-INNS-ENNS International Joint Conference on, Vol. 6, pp. 348 –353.

Vapnik, V.N. (1995). The Nature of Statistical Learning Theory, Springer Verlag, NY.