- Research
- Open Access

# A robust SVM classification framework using PSM for multi-class recognition

- Jinhui Chen
^{1}Email author, - Tetsuya Takiguchi
^{2}and - Yasuo Ariki
^{2}

**2015**:7

https://doi.org/10.1186/s13640-015-0061-x

© Chen et al.; licensee Springer. 2015

**Received:**27 August 2014**Accepted:**13 February 2015**Published:**11 March 2015

## Abstract

Our research focuses on the question of classifiers that are capable of processing images rapidly and accurately without having to rely on a large-scale dataset, thus presenting a robust classification framework for both facial expression recognition (FER) and object recognition. The framework is based on support vector machines (SVMs) and employs three key approaches to enhance its robustness. First, it uses the perturbed subspace method (PSM) to extend the range of sample space for task sample training, which is an effective way to improve the robustness of a training system. Second, the framework adopts Speeded Up Robust Features (SURF) as features, which is more suitable for dealing with real-time situations. Third, it introduces region attributes to evaluate and revise the classification results based on SVMs. In this way, the classifying ability of SVMs can be improved.

Combining these approaches, the proposed method has the following beneficial contributions. First, the efficiency of SVMs can be improved. Experiments show that the proposed approach is capable of reducing the number of samples effectively, resulting in an obvious reduction in training time. Second, the recognition accuracy is comparable to that of state-of-the-art algorithms. Third, its versatility is excellent, allowing it to be applied not only to object recognition but also FER.

## Keywords

- PSM
- SVMs
- SURF
- Region attributes
- Object recognition
- Facial expression recognition

## 1 Introduction

During the past decade or two, significant effort has been put into developing methods of training algorithms for pattern recognition, which is an attractive research subject in the field of computer vision due to the great potential for it to be used in many applications in a variety of fields, including object recognition, biological feature recognition, and human behavior analysis. Therefore, the need for this kind of technology in various different fields keeps propelling research forward year after year.

As the main detectors, AdaBoost and SVMs are widely used in this field of research. In 1997, Freund and Schapire [1] supplied the AdaBoost algorithm for realizing the learning framework of boosted trees, which could be derived from the Probably Approximately Correct (PAC) learning proposed by Valiant [2]. Since then, great advances have been made based on AdaBoost, especially a milestone work by Viola and Jones [3].

But some ideal strong classifiers usually require a large number of training samples and very time-consuming training experiments. Even now, many researchers are still trying to solve these problems. Li et al. [4] proposed a new learning SURF cascade for ameliorating boosting cascade frameworks. It improved the training efficiency, but the need for large-scale data gathering and extensive preparation creates a critical bottleneck. On the other hand, similar problems also exist in methods based on SVMs. There are too many examples, which will not be enumerated one by one here. Therefore, collecting many training samples and the associated long training time lead to considerable work and difficulty for researchers in the field of pattern recognition. Since training is a critical infrastructure for recognition engines, the research on training is significant for learning machines. Hence, there is a great need to solve the problem mentioned above.

Unfortunately, some researchers usually ignore these problems and argue that they just care about the recognition speed because the training is an offline task. However, diverse data appear everyday, and some may not be well covered by existing classifiers; thus, we have to update these classifiers frequently. This problem of having to retrain and refresh classifiers for unknown image data to alleviate possible hit-miss results is well known.

Similarly, we believe Google is a powerful search engine, and one of the most important reasons is that it refreshes its pagerank and indexing frequently. Moreover, its superior technological background guarantees its update speed is fast enough. Therefore, it is still very important that research on solving both the problems associated with collecting many training samples and those associated with long training time continue until effective, practical solutions are developed.

There are three main ways that the research described in this paper will contribute to the research being carried out in this field. The first is that the efficiency of SVM learning can be improved. Experiments show that the proposed approach is capable of reducing the number of samples effectively, resulting in an obvious reduction in training time. The second benefit is that the recognition accuracy is comparable to state-of-the-art algorithms. Third, this new system’s versatility is excellent, allowing it to be applied not only to object recognition but also facial expression recognition.

In the remainder of this paper, we first revisit related works in Section 2. Then, we describe the normalization of samples in Section 2 and the classifying framework in Section 2, respectively. Section 2 describes the experiments, and conclusions are drawn in Section 2.

## 2 Related work

We will first revisit related works on object recognition and facial expression recognition in this section. On one hand, facial expression recognition is a typical multi-class classification problem in computer vision. There are many precursors who have focused on FER research, and the latest ones, such as, Liu et al’s STM-ExpLet [6] and Huang et al’s new feature extraction algorithm for FER [7] have pushed the research forward. But many difficulties still exist in this research, because the subjects in the images usually have variable facial appearances and they can adopt a wide range of head poses. These problems are difficult to be overcome. Moreover, classifiers usually have to rely on a large-scale dataset for training. Unfortunately, current approaches of FER usually ignore these problems and do not present a robust feature set and a corresponding robust classifying framework that allows the expression to be discriminated cleanly under these situations. Reviewing [8] makes it clear that this situation has not been well improved. Looking deeper into the experimental reports of these works [9-13], we also find that the best precision achieved by any of these state-of-the-art methods is no higher than 31.7%, when the method is evaluated by real-world scenarios. Therefore, FER is still an extremely challenging task in computer vision. The first need is a robust feature and the corresponding high-quality training framework.

On the other hand, object recognition is also another hot research topic in computer vision due to its many applications. Great advances have been made in the past decade, especially since the milestone work by Viola and Jones [3]. But we must note that, even with the Viola-Jones method, in order to realize good generalization performance, more training data are required in the learning procedure. Although it is quite easy to collect many training samples from the Internet nowadays, collecting a large number of samples for training these object detectors [14-16] by search engines is easier said than done. To our knowledge, almost all existing object detectors require large amounts of data for training. Meanwhile, many methods are based on boosting cascade frameworks, and, as we know, almost all existing cascade frameworks are trained based on two conflicted criteria (false-positive-rate and hit-rate) for the detection-error tradeoff. Also known is the fact that training is usually required to achieve a very low false-positive rate per scan window (FPPW) such as 10^{−6} [17], which means that hundreds of millions or even billions of negative samples should be processed during the training procedure. Therefore, training ideal classifiers is a very time-consuming task. Usually, many researchers have to obtain mirror images of samples for training with the help of third-party software tools.

Our approach outperforms the methods advocated in a recent line of papers that use third-party software tools to obtain mirror images of samples for training in their object/facial-expression recognition systems, which we briefly review here. To the best of our knowledge, our approach is the first to apply the proposed method to both object recognition and facial expression recognition. Also, our approach is the first to employ the PSM directly for detector training without using any tools. Experiments show that this has the greatest impact on the performance of training efficiency, because time can be saved which would otherwise be spent on collecting vast amounts of data from the Internet or using third-party software to deal with the samples in order to get mirror images of these samples.

## 3 PSM for extending sample space

The PSM is derived from the perturbation method, and it can be applied to reducing the size of processing data. For example, it can be used to normalize facial and object data, which is usually adopted as a many-one mapping model. However, in this paper, what we are proposing is a one-many mapping model. Namely, we use it to extend the subspace of samples and the technical details of this one-many mapping model are discussed in this section. In addition, there are also many existing methods based on virtual images, which seem similar to ours, but most of them rely on pixel-level transformation (such as [18,19] etc.). Therefore, after processing with this approach, some features might be damaged easily. Moreover, they require some manual work; and in the training period, the program has to read a large number of virtual image files again, which leads to time waste. Our approach involves a classification framework that is capable of computing robustly and effectively while avoiding the problems mentioned above.

### 3.1 Training-sample normalization

*m*×

*n*pixels, and the original samples are normalized by mean value and variance of pixel transformation. Therefore, the image

*I*after normalization can be obtained according to the following equation:

*σ*is the standard deviation at the locations

*x*and

*y*, which can be calculated via

*a*,

*b*) is used to adjust the value of pixels. In this paper, we used regular samples in experiments; therefore,

*a*was set as 1, and

*b*was set as 0.

*μ*is the mean value of pixels, and it can be computed through image traversal using the following equation:

### 3.2 Changing orientational factors

After the calculation of subsection 2, we can thus extend the subspace of samples by changing the facial directions of the images. In this paper, we use the method that is proposed by Chen et al. [20] on the Procrustes analysis [21] to reconstruct three-dimensional face model and the obtain three-dimensional data. We indicate it in Algorithm ??. For more details, please refer to Appendix.

*E*

_{ r }is below a threshold

*ε*or

*K*landmarks are processed over, the while loop would be stopped and the three-dimensional data will be output. Here,

*β*=(

*β*

_{1},

*β*

_{2},⋯,

*β*

_{ m })

^{ T }is the shape parameter and

*m*is the dimensionality of the shape parameter, which is used to adjust three-dimensional shape data.

*S*

_{3D }is a 3×

*n*matrix,

*P*is a 2×3 orthographic projection matrix,

*T*is a 3×

*n*translation matrix consisting of

*n*translation vectors

*t*=[

*t*

_{ x },

*t*

_{ y },

*t*

_{ z }]

^{ T }, and

*R*

_{ θ }is a 3 × 3 rotation matrix where the yaw angle is

*θ*. In this paper,

*θ*is set as ±15°, ±30°, and ±60°. Thus, through Algorithm ??, we can reconstruct the three-dimensional data

*X*=(

*x*,

*y*,

*z*)

^{ T }from the original images. Hence, according to the transformation matrix formula,

we can convert the facial directions to extend the subspace of the training samples. Here, *T* and *R* are the shear mapping transformation matrix and the rotation matrix respectively, and *S* represents the scaling matrix.

### 3.3 Changing illumination attributes

where *V*
_{1} is the changing feature, *V*
_{2} is the result after the changes, *n* is the dimensionality of the feature vector, *w* is the weight coefficient, and *e* is the basis of illumination-change-factor vectors.

In this paper, *e* is obtained through processing the luminance-normalized rendering images by principal component analysis (PCA), wherein, *m* is the principal component (*m*=1,⋯,8). The rendering images are gained by the treatment of three-dimensional images obtained in subsection 2.

## 4 Classifying framework

This section will provide the framework used for SVM learning through adopting SURF features. Moreover, we will also employ the region attributes of images to revise the incorrect recognition of classifiers relying on visual features. We will describe them separately in this section.

### 4.1 Feature description

SURF is a scale- and rotation-invariant interest point detector and descriptor. It is faster than SIFT [22] and more robust against different image transformations. In this paper, we adopt an 8-bin T2 descriptor to describe the local feature, which is inspired by [23]. Unlike [23], however, we further allow different aspect ratios for each patch (the ratio of width and height) because this can make increase the speed of image traversal. We also imported diagonal and anti-diagonal filters because this can improve the description capability of the SURF descriptors.

*d*

_{ x }as the horizontal gradient image, which can be obtained using the filter kernel [−1,0,1], and

*d*

_{ y }is the vertical gradient image, which can be obtained using the filter kernel [−1,0,1]

^{ T }; Define

*d*

_{ D }as the diagonal image and

*d*

_{ AD }as the anti-diagonal image, both of which can be computed using two-dimensional filter kernels diag (−1,0,1) and antidiag (−1,0,1). Therefore, 8-bin T2 is able to be defined as \(v= (\sum (\left |d_{x}\right |+d_{x}),\sum (\left |d_{x}\right |-d_{x}),\sum (\left |d_{y}\right |+d_{y}),\sum (\left |d_{y}\right |-d_{y}),\sum (\left |d_{D}\right |+d_{D}),\sum (\left |d_{D}\right |-d_{D}),\sum (\left |d_{\textit {AD}}\right |+d_{\textit {AD}}),\sum (\left |d_{\textit {AD}}\right |-d_{\textit {AD}}))\). Here,

*d*

_{ x },

*d*

_{ y },

*d*

_{ D }, and

*d*

_{ AD }can be computed individually by filters shown in Figure 3a(1), a(2), b(1), and b(2) respectively in use of integral images, the details about how to compute two-dimensional Haar responses with integral images; please refer to [3].

The recognition template for SURF is 40×40 with four spatial cells, allowing the patch size ranging from 12×12 to 40×40 pixels. We slide the patch over the recognition template with four pixels forward to ensure enough feature-level difference. We further allow different aspect ratio for each patch (the ratio of width and height). The local candidate region of the features is divided into four cells. The descriptor is extracted in each cell. Hence, concatenating features in four cells together yields a 32-dimensional feature vector. About feature normalization, in practice, *L*
_{2} normalization followed by clipping and renormalization (*L*
_{2}
*H*ys) [24] is shown working best.

### 4.2 Classifier construction

The classifier of our framework is built based on one-versus-rest SVMs (OVR-SVMs). OVR strategy consists of constructing one SVM per class, which is trained to distinguish the samples of one class from the samples of all the remaining classes. Normally, classification of an unknown object is carried out by adopting the maximum output among all SVMs. The proposed method is based on OVR-SVMs classifiers and implemented by redeveloping liblinear SDK [25].

*f*

_{ j }(

*x*) denotes the output of the SVM trained to separate the class

*ω*

_{ j }from the other classes (total samples are

*M*). Then, for each sigmoid the parameters,

*c*

_{ j }and

*d*

_{ j }are optimized by minimizing the local negative log-likelihood:

Here are *N* outputs of the sigmoid function, where *h*
_{
k
} is the output of the sigmoid function with the probability *p*
_{
k
} event. In order to solve this optimization problem, Platt [26] applied a model-trust minimization algorithm based on the Levenberg-Marquardt algorithm. But in [27], Lin et al. pointed out that there are some problems in this method, meanwhile they proposed another minimization algorithm based on Newton’s method with backtracking line search.

*c*

_{ j }and

*d*

_{ j }are optimized by minimizing the global negative log-likelihood.

The optimization of parameters *c*
_{
j
} and *d*
_{
j
} is done with the intention of obtaining the lowest error rate on testing dataset. The reason why we use the negative log-likelihood is not only because it can optimize the parameters *c*
_{
j
} and *d*
_{
j
} but also because it can be used for comparing the various probability estimates; in other words, it can evaluate the error rate on machine learning and reject some of the unsatisfactory candidate expression regions described by SURF features.

### 4.3 Region attribute estimation

*C*=

*n*×

*m*blocks, and the feature vector of each block is computed. These vectors are used to construct a matrix

*X*, which is named as region attributes. Each column data of

*X*can be extracted from each block that is normalized by equalizing the value and variance of the luminance, while the norm is set as 1. The region attribute is estimated using the following score equation.

where *φ* is the eigenvector of *X*, *λ* is the eigenvalue of *X*, and *δ*
^{2} denotes the image noise correct divisor. When *δ*
^{2}=0, it means that the distances of all feature vectors of the current image projecting into subspace are unified; in the other words, the noise is negligible. *X* is estimated image region attributes, and \(\bar {X}\) is the average feature vector (AFV) of samples. If the value of distance is smaller, the score is higher, namely, the probability of miss-recognition is lower.

In this paper, the most significant way to enhance the learning efficiency of OVR-SVM classifiers is based on two conflicted criteria. This method is inspired by the boosting cascade (see Algorithm ??): As an error rate evaluating threshold *e* (*e*
^{
n
}=(1−*d*)), its function is similar to false-positive rate in the boosting cascade [2]. Recognition rate is evaluated by *d*; i.e., it is a meter threshold, whose function is similar to the hit rate in the boosting cascade. They are used for the recognition-error tradeoff where *e*<0.5. The classifying result will be considered as a miss recognition, and the OVR-SVMs classifying model is executed repeatedly until a given Boolean condition (*d*≤0.2) is met. In this way, it can guarantee that the classification results remain the same between the SVM classifiers and the distance metric on region attributes.

In order to make the framework more robust, we also adopt two important approaches. 1) If their results are coincident, the region attribute score *d* and recognition-error tradeoff parameter *e* will be updated for the next stage (Algorithm ??, step 3 and step 5); 2) When the results are not coincident, the current image will be put into the negative sample set automatically (Algorithm ??, step 7), so that the classifier can be updated at the next learning iteration stage. Therefore, the proposed framework is an adaptive learning framework that can cover the new data better than the conventional SVM-based methods. At the same time, this framework presents a mutual feedback mechanism for SVMs and the distance metric, which is more robust than a single classifying model. This is very important for avoiding some miss-recognition results that are individually categorized by the classifiers.

## 5 Experiments

At first, our method was proposed for facial expression recognition. But in practice, we found that it can be successfully applied to not only facial expression recognition but also object recognition. Therefore, this section will summarize the experimental data for both expression recognition and object recognition. The details of the implementation, dataset, and evaluation results will be shown here.

### 5.1 Implementation

We implemented all training and recognition programs in C++ on RHEL (Red Hat Enterprise Linux) 6.5 OS. In expression recognition, the facial detection part used the source code of Open CV, which was based on the Viola and Jones framework [3]. The expressional recognition part was implemented based on the proposed framework. In object recognition, all of the recognition systems were based on the proposed approaches. The experiments were done on the PC (Core i7-2600 3.40 GHz CPU and 8 GB RAM), and the training procedure was fully automatic. For SURF extraction, we adopted the integral image to speedup the computation as described in subsection 2. For machine learning, we built the OVR-SVMs through redeveloping liblinear software [25].

### 5.2 Experimental dataset

In the training stage, it is necessary to construct a mini-sized training set for machine learning, which will be applied to fix the parameters of sigmoid and softmax function. In the testing stage, we also need to build the testing set for evaluation. The easiest way to do this is to apply the same dataset to both the training and testing stages in a way of cross-validation. But, as pointed out by Platt [26], using the same data twice can sometimes lead to a disastrously biased estimate. Moreover, it cannot be proved that the approach is broadly practical. Therefore, in experiments, we used different datasets in the training and testing stages separately. The details of the training set and testing set are shown as follows:

**Training database set**

- 1)
*Expression recognition*: We used the Cohn-Kanade expression database (CK +) [29], which is a set of frontal face images posed by 123 people, but not all of the people posed each type of expression we needed. Therefore, we also collected some samples online using an image search engine. Finally, we obtained 240 initial facial samples for each type of emotion. All of these facial samples were normalized to 90×100 pixels. - 2)
*Object recognition*: a) 213 samples of JAFFE [30] were used for face training, which were normalized to 90×100 pixels; b) 600 side view car-training samples from the PASCAL VOC**2****0****0****7**dataset [31] were used for car training, which were normalized to 100×250 pixels; c) 600 samples were collected using the Google search engine for sheep training, which were normalized to 200×200 pixels.

In the training stages, the training data of current processing category were adopted as positive sample data; the other categories’ data were used for negative data.

**Testing database set**

- 1)
*Expression recognition*: In order to evaluate both of the real-life and ideal situations, we used two parts of testing sets. One part was obtained from soap operas, because many public databases were processed by providers in advance or for the other reasons, such as the images cannot represent real-life scenes, because they are not continuous images etc. Hence, we had to use some video clips from comedy dramas, which had a total of ten persons whose facial expressions were similar to the training samples. The images of these actors and actresses are on eight video clips having a length of 120 s. We marked this set as Test Set A. The other testing set was the JAFFE database [30], whose facial samples are totally different from the CK + database. The 213 JAFFE images were mixed randomly, and one image could be used repeatedly (to ensure that there are enough images for different video making). These images were also made into eight 120-s-long videos, and we marked this set as Test Set B. - 2)
*Object recognition*: 80 facial samples collected from the FDDB [32], 80 car-testing samples collected from the PASCAL VOC**2****0****0****5**database [31], and 80 samples of sheep collected from NUS-WIDE [33] were mixed and made into three clips of 9-min-long videos. All of the testing videos were normalized to the size of 640×480 and the frame rate of 60 frames per second (FPS). These videos were used to do evaluation experiments.

### 5.3 Experimental evaluation

#### 5.3.1 Expression recognition

**Training experiments**The training database of all methods was mentioned above, but only the proposed method did not adopt any process to obtain plenty of mirror samples. Hence, it reduced a mass of samples and took only 49.8 min to complete the whole process. Besides, the training procedure was fully automatic. The training results are shown in Table 1.

However, in order to enhance the generalization performance of comparison methods, we had to deal with the samples by some transformations (mirror reflection and rotate the images by horizontal and vertical angles ±15°, ±30°, and ±60° etc.). Finally, we obtained each class 30,960, total 123,840 facial samples for training classifiers. Therefore, they are very time-consuming tasks.

**Testing experiments**Figure 5 indicates the expression-recognition rate for different feature detectors based on the proposed framework. The aim of this experiment was evaluating the performance of the proposed detector using different methods of feature extraction. Hence, this experiment was done without a PSM model. The results showed that feature detectors using SURF and SIFT obtained more accurate recognition rates, but the average speed of the SIFT detector’s version was only 16.8 FPS. In comparison, the speed of the SURF’s version reached 39.4 FPS. Theoretically, 16.8 FPS is too slow to deal with complex scenes, such as real-time scenes. Our framework adopted an 8-bin T2 descriptor as the descriptor. It obtained similarly accurate recognition results compared to the accuracy of the original SURF’s version and even the SIFT’s one, but it surpassed the others in regard to feature extraction speed. In fact, in our experiments, 8-bin T2 descriptors had almost the same accuracy as the original SURF; however, the speed of original SURF version was only about 19 FPS, which was also extremely slow. Therefore, the feature descriptor based on 8-bin T2 SURF is the best choice for our framework.

*p*

_{ i }denotes the accuracy of

*i*th expression, and

*L*is the total of expressional categories). Because it can denote the mean level of recognition rate better than mean average precision (mAP) for the event containing different sample numbers in each independent class.

**Experimental results of expression recognition for Test Set A**

Table 2 shows the recognition rate of evaluation experiments for Test Set A. Since the human races and facial expressions of Test Set A’s people were similar to those of the training samples, meanwhile, the region attribute model was effective for Test Set A in which there are videos from real life. Consequently, its accuracy was quite better than the Test Set B’s. The maximum recognition precision of the proposed method was 86.3%, and the worst result was 69.3%.

On the other hand, Table 3 shows the recognition accuracies for Test Set B. Due to the variation and complexity of facial expressions across different cultures and races, the region attribute model was not effective for facial recognition. The results for this test set were not better than Test Set A’s. But on the whole, the results of both test sets show that the proposed method was the more accurate version of these methods. Note that the proposed method used training samples without any image-mirror process here. Namely, based on the mini-size training set, the proposed method can also obtain a better result; thus, this model allows for generating ideal strong classifiers without the need for large volumes of training samples. Hence, under these experimental conditions, the validity of the proposed methods was proved.

#### 5.3.2 Object recognition

**Training experiments**As our methods effectively reduce a great number of samples, it took very little time to complete the training process: 2.3 min (face), 10.1 min (car), and 10.6 min (sheep), respectively. The related data are shown in Table 4.

**Testing experiments**Figure 8 shows the experimental results. In Figure 8a, the component selections of the proposed method were evaluated. The middle one denoted the results of the proposed method without PSM, namely, the OVR-SVM + SURF model. The bottom one indicated the recognition result based on OVR-SVM. The OVR-SVM + PSM + SURF model was also the most accurate version of our object classifier. This also proves the outstanding versatility of the proposed method because it can analyze both human behaviors and object categories.

Figure 8b shows recognition results evaluated by the latest detectors. Li et al. [4] claimed they took about 47 min on their PC (Core i7 3.2 GHz CPU and 12 GB RAM) to obtain their ideal facial detectors, which could obtain a precision of 94% (the total of their training samples was 63,000). However, with our approach, using just 213 samples, we were able to achieve similar results. In the other experiment, we gained a little better result, using the same database as Li et al. did for car recognition. On the other hand, for sheep recognition, [34] adopted SIFT features based on SVM to obtain the best accuracy of 81.7% (versus our 86.7%). They did not provide their training time, but there is a reason to believe that the proposed method is better because the amount of their samples is hundreds of times more than ours. We applied similar classifiers; moreover, it was demonstrated that SURF extraction is much faster and more efficient than SIFT in [5].

## 6 Conclusions

This paper brings together effective normalization measures, visual features, and image attributes to construct a robust classification framework that minimizes the amount of training data needed while also improving the training efficiency. It can solve the question how to make classifiers be capable of processing images rapidly and accurately even without having to rely on a large-scale dataset. Hence, it is important to those with closely related research interests.

PSM is an effective approach for alleviating the trouble of collecting large amounts of training samples. By carrying out a large number of experiments, we found that SURF is the most suitable feature descriptor for our classifier, and the region attributes of images can revise some incorrectly detected classifiers caused by visual features. Combining these approaches, a robust classification framework can be constructed, which offers three major advantages. First, it can minimize the amount of training data and improve the training efficiency. Second, the recognition accuracy is comparable to state-of-the-art algorithms. Third, this framework can apply to not only facial expression recognition but also object recognition. The experiments proved the proposed method was valid in regard to training efficiency, recognition accuracy, and versatility.

In future research, considering a possible implementation in a real-life scenario, we are inclined to consider these points: 1) we will try to use region attributes as binary latent variables, which are incorporated into the SVM model for inference, and 2) we will ameliorate methods for the construction of SVMs to improve accuracy and to make our method capable of handling more complex tasks.

## 7 Appendix

### 7.1 Additional explanation for Algorithm 1

*P*in

*R*

^{ P }is represented by

*l*landmarks. Two figures

*P*:

*l*×

*p*and

*P*

^{′}:

*l*×

*p*

^{′}are said to have the same shape, if they are related by a special similarity transformation:

*p*×

*p*

^{′}, |

*Γ*|=1, a translation matrix

*γ*:

*p*

^{′}×

*l*, a positive scaling factor

*α*, and

*I*

_{ l }is a vector of ones. By using the generalized Procrustes analysis, it is possible to derive a consensus shape for a collection of figures [21], which is then used in registering new shapes into alignment with the collection by an Affine transformation. In a 3D model, the geometry is defined as a shape vector

*S*

_{3D}∈

**R**

^{3}, which contains the

*x*,

*y*, and

*z*coordinates of

*n*vertices. And Equation 13 is adjusted as follows

*β*=(

*β*

_{1},

*β*

_{2},⋯,

*β*

_{ m })

^{ T }is the shape parameter and

*m*is the dimension of the shape parameter which was determined to represent the shape of the 3D model. Given the input image indicated as

*S*

_{2D}=(

*x*

_{1},

*y*

_{1},⋯,

*x*

_{ n },

*y*

_{ n })∈

**R**

^{2}, the shape parameter

*β*needs to be determined such that it minimizes the shape residual between the projected 3D facial shape generated by the shape parameter and the input 2D facial shape. The optimal shape and pose parameters (

*β*,

*R*

_{ θ },

*T*) are obtained from

*S*

_{3D}is a 3×

*n*matrix that is reshaped from the 3

*n*×1 model shape vector obtained using Equation 13,

*P*is a 2×3 orthographic projection matrix,

*T*is a 3×

*n*translation matrix consisting of

*n*translation vectors

*t*=[

*t*

_{ x },

*t*

_{ y },

*t*

_{ z }]

^{ T }, and

*R*

_{ θ }is a 3×3 rotation matrix where the yaw angle is

*θ*. The 3D shape creation is indicated as follows:

- 1.
Initialization: set

*β*_{0}=0 and*k*=1. - 2.
Alignment:

*S*_{2D}is aligned with the 2D shape obtained by projecting the frontal 3D shape (*s*_{ i }) onto the*x*−*y*plane. - 3.
Update

*R*_{ θ }and*T*with the fixed shape parameter by*m**i**n*∥*P*(*R*_{ θ }*S*_{3D}+*T*)−*S*_{2D}∥^{2}, and reconstruct (*S*_{3D})_{ k }using the shape parameter*β*_{ k }. - 4.
Verify whether

*E*_{ r }≤*φ*or*k*>*N*; if not, go to step 3 and*k*=*k*+1. - 5.
Reconstruct

*S*_{3D}using the final shape parameters.

When *E*
_{
r
} is below a threshold (e.g. in [21], Gower suggested setting *φ*=10^{−4}) or the landmarks are processed over, the reconstruction would be stopped, and the consensus shape would be output.

**Note:** Part content of this Appendix was published in the ACM International Conference on Multimedia 2013.

## Declarations

### Acknowledgements

The authors would like to sincerely thank the editors and the anonymous reviewers for their constructive comments and valuable suggestions, which helped to improve this paper.

## Authors’ Affiliations

## References

- Y Freund, R Schapire, A Desicion-theoretic Generalization of On-line Learning and an Application to Boosting. J. Comput. Syst. 55(1), 119–139 (1997).View ArticleMATHMathSciNetGoogle Scholar
- LG Valiant, A theory of the learnable. Commun. ACM. 27(11), 1134–1142 (1984).View ArticleMATHGoogle Scholar
- P Viola, M Jones, in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), 1. Rapid object detection using a boosted cascade of simple features, (2001), pp. 511–5181.Google Scholar
- J Li, Y Zhang, in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR). Learning SURF cascade for fast and accurate object detection, (2013), pp. 3468–3475.Google Scholar
- H Bay, T Tuytelaars, L Van Gool, in Proc. Eur. Conf. Comput. Vis. (ECCV). SURF: speeded up robust features, (2006), pp. 404–417.Google Scholar
- M Liu, S Shan, R Wang, X Chen, in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR). Learning expressionlets on spatio-temporal manifold for dynamic facial expression recognition, (2014), pp. 1749–1756.Google Scholar
- H-F Huang, S-C Tai, Facial expression recognition using new feature extraction algorithm. Electron. Lett. Comput. Vision Image Anal. 11(1), 41–54 (2012).MathSciNetGoogle Scholar
- J. J. K. S. RolandGoecke Abhinav Dhall, T Gedeon, in Proc. ACM The Int. Conf. on Multimodal Interaction (ICMI). Emotion recognition in the wild challenge 2014, (2014).Google Scholar
- O Rudovic, V Pavlovic, M Pantic, in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR). Multi-output Laplacian dynamic ordinal regression for facial expression recognition and intensity estimation, (2012), pp. 2634–2641.Google Scholar
- G Zhao, M Pietikainen, Dynamic texture recognition using local binary patterns with an application to facial expressions. IEEE Trans. PAMI. 29(6), 915–928 (2007).View ArticleGoogle Scholar
- L Wang, Y Qiao, X Tang, in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR). Motionlets: mid-level 3D parts for human motion recognition, (2013), pp. 2674–2681.Google Scholar
- P Scovanner, S Ali, M Shah, in Proc. ACM Multimedia Conf. (MM). A 3-dimensional sift descriptor and its application to action recognition, (2007), pp. 357–360.Google Scholar
- A Klaser, M Marszalek, C Schmid, in Proc. British Machine Vis. Conf. (BMVC). A spatio-temporal descriptor based on 3D-gradients, (2008), pp. 275–110.Google Scholar
- S Maji, AC Berg, in Proc. IEEE Int. Conf. Comput. Vis. (ICCV). Max-margin additive classifiers for detection, (2009), pp. 40–47.Google Scholar
- Q Zhu, M-C Yeh, K-T Cheng, S Avidan, in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), 2. Fast human detection using a cascade of histograms of oriented gradients, (2006), pp. 1491–1498.Google Scholar
- L Bourdev, J Brandt, in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), 2. Robust object detection via soft cascade, (2005), pp. 236–243.Google Scholar
- P Viola, MJ Jones, Robust real-time face detection. Int. J. Comput. Vision. 57(2), 137–154 (2004).View ArticleGoogle Scholar
- W-S Chu, C-R Huang, C-S Chen, Gender classification from unaligned facial images using support subspaces. Inf. Sci. 221, 98–109 (2013).View ArticleGoogle Scholar
- D Decoste, B Schölkopf, Training invariant support vector machines. Machine Learning. 46(1-3), 161–190 (2002).View ArticleMATHGoogle Scholar
- J Chen, Y Ariki, T Takiguchi, in Proc. ACM Multimedia Conf. (MM). Robust facial expressions recognition using 3D average face and ameliorated adaboost, (2013), pp. 661–664.Google Scholar
- JC Gower, Generalized procrustes analysis. Psychometrika. 40(1), 33–51 (1975).View ArticleMATHMathSciNetGoogle Scholar
- DG Lowe, in Proc. IEEE Int. Conf. Comput. Vis. (ICCV), 2. Object recognition from local scale-invariant features, (1999), pp. 1150–1157.Google Scholar
- J Li, T Wang, Y Zhang, in Proc. IEEE Int. Conf. Comput. Vis. (ICCV) Workshops. Face detection using SURF cascade, (2011), pp. 2183–2190.Google Scholar
- N Dalal, B Triggs, in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), 1. Histograms of oriented gradients for human detection, (2005), pp. 886–8931.Google Scholar
- R-E Fan, K-W Chang, C-J Hsieh, X-R Wang, C-J Lin, LIBLINEAR: a library for large linear classification. J. Machine Learning Res. 9, 1871–1874 (2008).MATHGoogle Scholar
- J Platt, in Proc. Advances in Large Margin Classifiers. Probabilities for SV machines, (2000), pp. 61–74.Google Scholar
- H-T Lin, C-J Lin, RC Weng, A note on Platt’s probabilistic outputs for support vector machines. Machine learning. 68(3), 267–276 (2007).View ArticleGoogle Scholar
- Z Sun, N Ampornpunt, M Varma, S. v. n. Vishwanathan, in Proc. Adv. Neural Inf. Process. Syst. (NIPS). Multiple kernel learning and the SMO algorithm, (2010), pp. 2361–2369.Google Scholar
- P Lucey, JF Cohn, T Kanade, J Saragih, Z Ambadar, I Matthews, in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR) Workshops. The extended Cohn-Kanade dataset (CK+): a complete dataset for action unit and emotion-specified expression, (2010), pp. 94–101.Google Scholar
- M Lyons, J Gyoba, S Akamatsu, Automatic Classification of Single Facial Images. EEE Trans. PAMI. 21(12), 1357–1362 (1999).View ArticleGoogle Scholar
- M Everingham, L Van Gool, CKI Williams, J Winn, A Zisserman, The pascal visual object classes (VOC) challenge. Int. J. Comput. Vision. 88(2), 303–338 (2010).View ArticleGoogle Scholar
- V Jain, EG Learned-Miller, FDDB: a benchmark for face detection in unconstrained settings. UMass Amherst Technical Report (2010).Google Scholar
- T-S Chua, J Tang, R Hong, H Li, Z Luo, Y-T Zheng, in Proc. ACM Int. Conf. Image and Video Retrieval (CVIR). NUS-WIDE: a Real-World Web Image Database from National University of Singapore, (2009).Google Scholar
- X Li, CGM Snoek, in Proc. ACM Multimedia Conf. (MM). Classifying tag relevance with relevant positive and negative examples, (2013), pp. 485–488.Google Scholar

## Copyright

This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly credited.