Acessibilidade / Reportar erro

Arboreal Identification Supported by Fuzzy Modeling for Trunk Texture Recognition

ABSTRACT

Due to the natural variability of the arboreal bark there are texture patterns in trunk images with values belonging to more than one species. Thus, the present study analyzed the usage of fuzzy modeling as an alternative to handle the uncertainty in the trunk texture recognition, in comparison with other machine learning algorithms. A total of 2160 samples, belonging to 20 tree species from the Brazilian native deciduous forest, were used in the experimental analyzes. After transforming the images from RGB to HSV, 70 texture patterns have been extracted based on first and second order statistics. Secondly, an exploratory factor analysis was performed for dealing with redundant information and optimizing the computational effort. Then, only the first dimensions with higher cumulative variability were selected as input variables in the predictive modeling. As a result, fuzzy modeling reached a generalization ability that outperformed some algorithms widely used in classification tasks. Therefore, the fuzzy modeling can be considered as a competitive approach, with reliable performance in arboreal trunk texture recognition.

Keywords:
soft computing; image processing; pattern matching; bioinformatics

RESUMO

Devido à variabilidade natural da casca arbórea, há padrões de textura em imagens de tronco com valores pertencentes a mais de uma espécie. Logo, o presente estudo analisou o uso da modelagem fuzzy como uma alternativa para lidar com a incerteza no reconhecimento de padrões, em comparação com outros algoritmos de aprendizado de máquina. Para as análises experimentais foram utilizadas um total de 2160 amostras, pertencentes a 20 espécies arbóreas da floresta decídua brasileira. Depois de transformar as imagens do sistema RGB para modelo HSV, 70 padrões de textura foram extraídos com base em estatísticas de primeira e segunda ordem. Na sequência, foi realizada uma análise fatorial exploratória para tratar informações redundantes e otimizar o esforço computacional. Então, apenas as primeiras dimensões com maior variabilidade acumulada foram selecionadas como variáveis de entrada na modelagem preditiva. Como resultado, a modelagem fuzzy alcançou uma capacidade de generalização superior a de algoritmos amplamente usados em tarefas de classificação. Portanto, a modelagem fuzzy pode ser considerada uma abordagem com desempenho competitivo e confíavel no reconhecimento da textura em imagens do tronco arbóreo.

Palavras-chave:
computação não-rígida; processamento de imagens; correspondência de padrões; bioinformática

1 INTRODUCTION

The usage of computational intelligence in the feature extraction and pattern recognition from biological data has been increasingly studied for supporting the arboreal identification. However, as the studies carried out have focused on the processing of leaf images, these techniques are not applicable when the leaf structure is not available, as occurs with deciduous species at certain times of the year.

As an alternative, the texture recognition in tree trunk images still has few outcomes reported in the literature, in which the predictive modeling has been performed using machine learning algorithms based on k-Nearest Neighbors 2626 A. Porebski, N. Vandenbroucke & L. Macaire. Iterative feature selection for color texture classification. In Image Processing, 2007. ICIP 2007. IEEE International Conference on, volume 3, pp. III-509. IEEE (2007).), (3030 Y.Y. Wan, J.X. Du, D.S. Huang, Z. Chi, Y.M. Cheung, X.F. Wang & G.J. Zhang. Bark texture feature extraction based on statistical texture analysis. In Intelligent Multimedia, Video and Speech Processing, 2004. Proceedings of 2004 International Symposium on, pp. 482-485. IEEE (2004)., Artificial Neural Networks 1919 Z.K. Huang, C.H. Zheng, J.X. Du & Y.y. Wan. Bark classification based on textural features using artificial neural networks. In International Symposium on Neural Networks, pp. 355-360. Springer (2006)., Support Vector Machine 66 J. Boman. Tree species classification using terrestrial hotogrammetry. Master’s thesis, Department of Computing Science, Umea University, Umea (2013).), (1313 S. Fiel & R. Sablatnig. Automated identification of tree species from images of the bark, leaves and needles. In 16th Computer Vision Winter Workshop. Citeseer, p. 67. Citeseer (2011).), (1818 Z.K. Huang. Bark classification using RBPNN based on both color and texture feature. International Journal of Computer Science and Network Security, 6(10) (2006), 100-103., and Decision Tree 1010 A. Bressane, J.A.F. Roveda & A.C. Martins. Statistical analysis of texture in trunk images for biometric identification of tree species. Environmental monitoring and assessment, 187(4) (2015), 212..

By analyzing statistical properties in tree trunk images, 1010 A. Bressane, J.A.F. Roveda & A.C. Martins. Statistical analysis of texture in trunk images for biometric identification of tree species. Environmental monitoring and assessment, 187(4) (2015), 212. found that, due to the natural variability of the arboreal bark, commonly its texture patterns have some values belonging to more than one species, i.e, there is an overlap between neighboring subspaces. As a consequence, this overlapping in the pattern matching can lead to an ambiguity during predictive modeling.

In these cases, there is some uncertainty with regard to what species the sample belongs to, undermining the texture analysis by means of predictor variables with a sharply defined boundary. Therefore, the present study aims to analyze the usage of fuzzy modeling as an approach to deal with the uncertainty in the trunk texture recognition, in comparison with other machine learning algorithms.

In the mid-1960s, the fuzzy set theory has been developed by 3232 L.A. Zadeh. Fuzzy sets. Information and control, 8(3) (1965), 338-353. as an extension of the classical set theory to provide a mathematical treatment for complex phenomena, becoming it popular after 1980s 3333 L.A. Zadeh. Is there a need for fuzzy logic? Information sciences, 178(13) (2008), 2751-2779.), (2525 W. Pedrycz & F. Gomide. Fuzzy systems engineering: toward human-centric computing. John Wiley & Sons (2007).. For that, the fuzzy modeling is a soft-computing method capable of processing uncertain knowledge or data. Thus, by affording a convenient formalism for integrating different kinds of variables, by means of an user-friendly structure with transparency and interpretability, the usage of fuzzy modeling is becoming more and more common, with several applications in the environmental sciences over the years (e.g. 77 A. Bressane, J.A. Bagatini, C.H. Biagolini, J.A.F. Roveda, S.R.M.M. Roveda, F.H. Fengler & R.M. Longo. Neuro-fuzzy modeling: a promising alternative for risk analysis in urban afforestation management. Brazilian Journal of Forest Science. IN PRESS.), (88 A. Bressane, C.H. Biagolini, P.S. Mochizuki, J.A.F. Roveda & R.W. Lourenço. Fuzzy-based methodological proposal for participatory diagnosis in the linear parks management. Ecological Indicators, 80 (2017), 153-162.), (99 A. Bressane, P.S. Mochizuki, R.M. Caram & J.A.F. Roveda. A system for evaluating the impact of noise pollution on the population’s health. Reports in public health, 32(5) (2016).), (2222 D. Liu & Z. Zou. Water quality evaluation based on improved fuzzy matter-element method. Journal of Environmental Sciences, 24(7) (2012), 1210-1216.), (2323 L. Liu, J. Zhou, X. An, Y. Zhang & L. Yang. Using fuzzy theory and information entropy for water quality assessment in Three Gorges region, China. Expert Systems with Applications, 37(3) (2010), 2517-2521.), (2121 A. Lermontov, L. Yokoyama, M. Lermontov & M.A.S. Machado. River quality analysis using fuzzy water quality index: Ribeira do Iguape river watershed, Brazil. Ecological Indicators, 9(6) (2009), 1188-1197.), (44 J. Ascough, H. Maier, J. Ravalico & M. Strudley. Future research challenges for incorporation of uncertainty in environmental and ecological decision-making. Ecological modelling, 219(3) (2008), 383-399.), (22 V. Adriaenssens, B. De Baets, P.L. Goethals & N. De Pauw. Fuzzy rule-based models for decision support in ecosystem management. Science of the Total Environment, 319(1) (2004), 1-12.), (2828 W. Silvert. Fuzzy indices of environmental conditions. Ecological Modelling, 130(1) (2000), 111-119.).

According to 2020 H. Ishibuchi & T. Nakashima. Effect of rule weights in fuzzy rule-based classification systems. IEEE Transactions on Fuzzy Systems, 9(4) (2001), 506-515., the main applications of the fuzzy modeling used to be optimization and control problems. Nevertheless, nowadays many other areas can be highlighted, such as the development of intelligent systems for supporting the decision making, data mining, signal processing, diagnosis, forecasting, regression, and classification from numerical data using pattern recognition based on the graded membership 2929 H. Singh, M.M. Gupta, T. Meitzler, Z.G. Hou, K.K. Garg, A.M. Solo & L.A. Zadeh. Real-Life Applications of Fuzzy Logic. Adv. Fuzzy Systems, 2013 (2013), 581879-1.. Thereby, the fuzzy modeling can achieve a competitive performance when compared to other machine learning algorithms in classification tasks involving uncertainty, vagueness, partial true, which demand predictors without hard boundaries 33 C. Arunpriya & A.S. Thanamani. Fuzzy inference system algorithm of plant classification for tea leaf recognition. Indian Journal of Science and Technology, 8(S7) (2015), 179-184.), (2727 L.S. Riza, C.N. Bergmeir, F. Herrera & J.M. Benítez Sánchez. frbs: Fuzzy rule-based systems for classification and regression in R. Journal of Statistical Software, 65(6) (2015), 1-30..

2 METHODS

2.1 Data collection and feature extraction

The data were collected using a digital camera for capturing outer bark images at different heights of the trunk, at a 50 mm distance around the trees. Due to the three-dimensional shape of arboreal trunk, only a central area was used for extracting features, in order to avoid the distortion at the image edge. Then, using a moving mask with 512 x 512 pixels, 2160 samples were obtained, being 108 of each of the 20 tree species from the Brazilian native deciduous forest, shown in Figure 1.

Figure 1:
Tree trunk images (512x512 pixels) from: Anadenanthera falcata (Af), Anadenanthera macrocarpa (Am), Bauhinia forficate (Bf), Caesalpinia ferrea (Ca), Caesalpinia echinata (Ce), Cedrela fissilis (Cf), Caesalpinia peltophoroides (Cp), Ceiba speciosa (Cs), Centrolobium tomentosum (Ct), Enterolobium contortisiliquum (Ec), Erythrina speciosa (Es), Gochnatia polymorpha (Gp), Guazuma ulmifolia (Gu), Hymenaea courbaril (Hc), Inga vera (Iv), Piptadenia gonoacantha (Pg), Schizolobiun parahyba (Sp), Tibouchina granulosa (Tg), Tabebuia roseoalba (Tr), and Zanthoxylum kleinii (Zk).

To reduce the influence of the environmental conditions and image acquisition settings, before starting the feature extraction the images have been transformed from RGB (red-green-blue) system to HSV (hue-saturation-value) space. Then, features based on first and second order statistics were extracted using the V channel from the grayscale images.

The first-order statistical parameters included 6 texture features, equivalent to uniformity, entropy, skewness, smoothness, intensity, and standard deviation, described below from 1414 R.C. Gonzales & R.E. Woods. Digital image processing, volume 3. Pearson Prentice Hall (2008)..

As a measure of the proximity of the gray levels, the uniformity (u) is given by:

u = i = 0 L - 1 p 2 ( z i ) (2.1)

where L correspond to the number of gray levels in the image, z i is the intensity, and p(z i ) is the image histogram.

The first-order entropy (e) measures the randomness in the image, as in:

e = - i = 0 L - 1 p ( z i ) log 2 p ( z i ) (2.2)

The skewness is a measure of the asymmetry (µ 3), and smoothness (s) takes in to account the transition of gray shades, respectively obtained by:

μ 3 = i = 0 L - 1 ( z i - μ 1 ) 2 p ( z i ) (2.3)

and

s = 1 - 1 1 + μ 2 2 (2.4)

where (µ 1) is the intensity that returns the gray level average, and (µ 2) is the standard deviation, calculated by:

μ 1 = i = 0 L - 1 z i p ( z i ) (2.5)

and

μ 2 = 1 n - 1 i = 1 n ( z i - μ 1 ) 2 (2.6)

where n is the number of image pixels.

In turn, the second order statistics was composed of the contrast, correlation, energy, and homogeneity, measured at 16 positions (ϕ), correspondent to distance between pixels equal to 1, 3, 5 and 7, in the rotation angles 0, 45, 90 and 135 degrees, producing 64 texture features. These descriptors are described below from 1717 R.M. Haralick, K. Shanmugam & I. Dinstein. Textural features for image classification. IEEE Transactions on systems, man, and cybernetics, 3(6) (1973), 610-621. and 1515 R.C. Gonzales, R.E. Woods & S.L. Eddins. Digital image processing using MATLAB, volume 2. Gatesmark Publishing (2009)..

Contrast (c) compares the intensity of neighboring pixels, being calculated by:

c ϕ = i = 1 k j = 1 k ( i - j ) 2 p i j (2.7)

where k is the co-occurrence matrix dimension, p ij is probability of satisfying ϕ.

The correlation (r) measures the probability of occurrence of specified pixel pairs, given by:

r ϕ = i = 1 k j = 1 k ( i - m r o w ) ( j - m c o l ) σ r o w - σ c o l p i j (2.8)

Energy (() adds the squared elements in the co-occurrence matrix, and homogeneity (h) measures the closeness of gray levels in the spatial distribution over image, respectively obtained by:

ε ϕ = i = 1 k j = 1 k p i j 2 (2.9)

and

h ϕ = i = 1 k j = 1 k p i j 1 + | i - j | (2.10)

From the foregoing, the total number of measured variables amounted to 70 texture features. Then, taking into account that some features may be highly correlated, an Exploratory Factor Analysis (EFA) has been performed. As a multivariate analysis technique, the EFA finds a coordinate system that maximizes the variance shared among variables, enabling to reduce the data dimensionality and prevent the use of redundant information 1212 A.B. Costello & J.W. Osborne. Best practices in exploratory factor analysis: Four recommendations for getting the most from your analysis. Pan-Pacific Management Review, 12(2) (2009), 131-146..

In the new m-dimensional space found by EFA, the standardized original variables (z) correspond to linear combinations of underlying factors (z'), given by 3131 A.G. Yong & S. Pearce. A beginner´s guide to factor analysis: Focusing on exploratory factor analysis. Tutorials in Quantitative Methods for Psychology, 9(2) (2013), 79-94.:

z j = a j 1 z ' 1 + a j 2 z ' 2 + . . . + a j m z ' m (2.11)

For that, the EFA was carried out using the Spearman’s coefficient, a non-parametric alternative regarded as robust for general distributions (non-normal data), the principal factors as extraction method, and the communalities (h i ) based on the squared multiple correlations, as in:

h i = j = 1 m l i j 2 (2.12)

where l ij is the correlation between the i th principal factor with j th original variable (texture feature), previously standardized by means of:

z i = x i - x ¯ σ (2.13)

where x i is the measured original variable, x¯ and σ are respectively its mean and standard deviation.

Thus, the features extracted from tree trunk images have been reduced to fewer latent variables (principal factors), which were used as predictors for generating fuzzy if-then rules in the texture pattern recognition.

2.2 Fuzzy modeling for the pattern recognition

The development of the fuzzy modeling for classification tasks is relatively recent in comparison to other applications. Notwithstanding, several approaches have already been proposed, including space partitioning 1111 Z. Chi, H. Yan & T. Pham. Fuzzy algorithms: with applications to image processing and pattern recognition, volume 10. World Scientific (1996)., neural-network-based methods 2424 D. Nauck & R. Kruse. A neuro-fuzzy method to learn fuzzy classification rules from data. Fuzzy sets and Systems, 89(3) (1997), 277-288., clustering techniques 11 S. Abe & R. Thawonmas. A fuzzy classifier with ellipsoidal regions. IEEE Transactions on fuzzy systems, 5(3) (1997), 358-368., genetic algorithms 1616 A. Gonzblez & R. Pérez. SLAVE: A genetic learning system based on an iterative approach. IEEE Transactions on Fuzzy Systems, 7(2) (1999), 176-191., and fuzzy partition using certainty grades 2020 H. Ishibuchi & T. Nakashima. Effect of rule weights in fuzzy rule-based classification systems. IEEE Transactions on Fuzzy Systems, 9(4) (2001), 506-515..

For the predictive modeling in the present study, we used a fuzzy rule-based classification system, created and described by 2727 L.S. Riza, C.N. Bergmeir, F. Herrera & J.M. Benítez Sánchez. frbs: Fuzzy rule-based systems for classification and regression in R. Journal of Statistical Software, 65(6) (2015), 1-30. as FRBCS.W algorithm, made available in R programming language by means of the ‘frbs’ package. The FRBCS.W algorithm has been developed based on the Ishibuchi’s method 2020 H. Ishibuchi & T. Nakashima. Effect of rule weights in fuzzy rule-based classification systems. IEEE Transactions on Fuzzy Systems, 9(4) (2001), 506-515..

As aforementioned, the Ishibuchi’s method is a learning method from numerical data that consists of the fuzzy partitioning with certainty grades. In the learning process, the antecedent part of rules is determined by a grid-type fuzzy partition. This partitioning occurs by dividing the input space of the predictor variables (x i ) into regular fuzzy regions, resulting in uniform and symmetrical intervals correspondent to the antecedent terms (a ij ), as can be seen in Figure 2.

Figure 2:
Grid-type fuzzy partition: (a) partitioning of the predictor variable - x i ; (b) intervals of certainty and uncertainty that comprises the fuzzy region.

By using the grid-type fuzzy partition, the total number of rules (N) is determinate by amount of possible combinations of the antecedent terms. For that, rulebase is generated by pattern matching, calculating membership degrees (φ) of the training data in the antecedents terms (a ij ) of each predictor variable (x i ). In turn, the consequent part is defined as the dominant categorical variable (C j ) in the decision area formed by the fuzzy if-then rule (2020 H. Ishibuchi & T. Nakashima. Effect of rule weights in fuzzy rule-based classification systems. IEEE Transactions on Fuzzy Systems, 9(4) (2001), 506-515.):

Rule R j : IF x 1 is a 1 j AND ... AND x m is a m j THEN C j with C F j , j = 1 , 2 , . . . , N (2.14)

where x is a m-dimensional vector of predictor variables (x i ), CF j is the certainty grade of the rule R j , and C j is the dominant categorical variable, determined taking into account:

p c l a s s C j φ j ( x p ) = max p c l a s s k φ j ( x p ) : k = 1 , 2 , . . . , c (2.15)

where x p = (x p1 ,...,x pm ) is a new pattern, and c is the number of output classes.

After generating the predictive model, the classification of new instances is based on a single winner rule, which is determined by (2020 H. Ishibuchi & T. Nakashima. Effect of rule weights in fuzzy rule-based classification systems. IEEE Transactions on Fuzzy Systems, 9(4) (2001), 506-515.):

φ j ( x p ) . C F j = max φ j ( x p ) . C F j : j = 1 , 2 , . . . , N (2.16)

where φj (x p ) is the compatibility grade of the instance in the rule R j , and CF j is the certainty grade, a real number in the interval [0, 1] that works as the weight of rule, given by:

C F j = β c l a s s C j ( R j ) - β ¯ k = 1 c β c l a s s k ( R j ) (2.17)

where

β ¯ = k C j β c l a s s k ( R j ) ( c - 1 ) (2.18)

and

β c l a s s k ( R j ) = x p c l a s s k φ j ( x p ) , k = 1 , 2 , . . . , c (2.19)

2.3 Benchmarking experiment

From the database with 2160 samples, we used 70% of this total, randomly selected, for the machine learning process. During this process a 5-fold cross-validation was carried out over learning dataset, in order to find the best control parameters setting. Then, a hold-out validation has been performed using the remaining 30% as testing dataset for assessing the generalization ability of the Fuzzy Rule-Based Classification System (FRBCS) in the trunk texture pattern recognition (Figure 3).

Figure 3:
Split of database for the learning process and to assess the generalization ability based on testing dataset.

Furthermore, as a reference for assessing the performance from the fuzzy-based approach, a benchmarking experiment has been carried out using the same database for training, checking and testing other algorithms shown in Table 1.

Table 1:
Algorithms and setting of control parameters adjusted during the learning process, which provide the best results in the cross-validation over the checking dataset.

Based on the testing results, the learning algorithms performance has been assessed according to the overall accuracy (θ), which measures the ratio of samples correctly classified by the total number of samples (n T ), as in:

θ = n T - 1 i = 1 n s p T P s p i (2.20)

where TP spi is the total number of true positive samples, and n sp is the total number of tree species.

3 RESULTS AND DISCUSSION

Regarding the requirements for data preprocessing using multivariate analysis, the Cronbach’s alpha equivalent to 0.9 indicated an excellent internal consistency, and the Kaiser-Meyer-Olkin equal to 0.97 confirmed a good sampling adequacy, verifying sufficient conditions to perform the Exploratory Factor Analysis (EFA), whose result is shown in Figure 4.

Figure 4:
Eigenvalues and cumulative variability explained by the first 20 latent variables (principal factors) produced from the Exploratory Factor Analysis.

From the Figure 4, we find that the first 20 principal factors explain 99.0% of the cumulative variability. Therefore, the EFA was capable of reducing the data dimensionality and at the same time retaining almost all information available in the 70 original variables. Thus, these principal factors were used as predictor variables in the modeling process, affording the results shown in Table 2.

Table 2:
Performance of the machine learning algorithms in the benchmarking experiment, using the first 20 principal factors as predictor variables.

In general, each machine learning algorithm has properties which can provide better performance than others, depending on the characteristics of the case under analysis. Thus, the performance from the algorithms in the benchmarking experiment has been discussed taking into account such properties. In this sense, by analyzing Table 2 it is noted three performance groups according to the accuracy over testing dataset.

With accuracy less than 80%, in the first group are the Single Decision Tree (SDT) and Cascade-Correlation Neural Network (CNN). The CNN is a self-organizing network that determines its own size and topology, by adding neurons to the architecture. The SDT also grows adding nodes to its structure, both for reaching greater preciseness during the learning process. As a consequence, these algorithms can lead to an overfitting to the train data, losing some generalization ability. Then, we use an overfitting control pruning the models to minimum cross-validated error over checking dataset.

Despite this, the CNN performance decreases from 86.2% during training to 78.5% in the testing, and SDT from 91.6% to 72.3%. Therefore, these findings can be considered as an indicator of the complexity of the arboreal trunk texture, making hardier the classification task.

Capable of handling this issue better than the single tree-based model (SDT), the Decision Tree Forest (Random Forest), Stochastic Gradient Boosting (TreeBoost), and Boosted Rule-Based Model (C5.0) are in the second group of algorithms with medium-performance in the testing (from 80 to 90%), along with k-Nearest Neighbors (KNN).

The Random Forest and TreeBoost are ensembles based on different strategies of creating a collection of decision trees. The Random Forest uses the bagging (Bootstrap Aggregating) technique for creating trees grown in parallel, which afforded a generalization ability of 89.5%. On the other hand, the TreeBoost uses a sequential training (boosting) that resulted in a series of trees with 87.3% accuracy. Similarly, C5.0 is a voting classification algorithm also based on a boosting technique to create a collection of rules that achieved 85.3% accuracy. The boosting usually provides more accuracy than bagging strategy, except when there is some noise in data, such as outliers 55 E. Bauer & R. Kohavi. An empirical comparison of voting classification algorithms: Bagging, boosting, and variants. Machine learning, 36(1-2) (1999), 105-139.. Therefore, as the Random Forest outperforms the boosting-based models in the present analysis, we can consider some influence of outliers. Notwithstanding, as the bark texture in the arboreal trunk is a biological feature subject to imperfections, these outliers has not been removed because they can be caused by a natural variability. In turn, the KNN is a non-parametric algorithm of instance-based learning, in which a pattern is recognized by majority voting according to the similarity with the k nearest neighbors. By using kernel functions to weight the vote of the neighbors, the KNN provides 89.7% accuracy, slightly higher than ensemble-based models.

The third group with high-performance, more than 90% of accuracy over testing dataset, has been formed by the Support Vector Machine (SVM), Probabilistic Neural Network (PNN), Fuzzy Rule-Based Classification System (FRBCS), and Multilayer Perceptron Neural Network (MLP).

The SVM operates by finding an n-dimensional hyperplane in order to optimize the separation of different data classes. Although similar to the artificial neural networks (ANN) in some aspects, the SVM is less prone to overfitting and has good adequacy for dealing with high dimensional spaces and outliers, because it selects the most suitable features and considers only the most relevant points. Besides that, the SVM has a solution global and unique whilst the ANN can suffer from multiple local minima. Thus, in our analysis the SVM provides a significant improvement in comparison with most of the learning algorithms, reaching 96.2% accuracy over testing dataset.

Among the artificial neural networks, the PNN performs the classification based on the estimation of probability density functions, capable of dealing with erroneous data and computing nonlinear decision boundaries as complex as necessary, in order to approach the Bayes optimal, i.e., to minimize the error in a probabilistic manner as much as possible. Thus, relatively insensitive to outliers, the PNN achieves virtually the same performance than SVM, with 96.1% accuracy over testing dataset. In turn, the MLP allows nonlinear mappings, using logistic activation functions and back-propagation algorithm for adjusting the neural network weights. To prevent overfitting, we use the MLP architecture with minimum validated error during the learning process, resulting in a substantial generalization ability correspondent to 90.8% accuracy, but still even lower than PNN one.

Regarding the FRBCS, to be the focus of the present study, in what follows we approach a more detailed description on the machine learning process, before presenting the accuracy over testing dataset. During the training we found that the gaussian curve membership function afforded a performance better than ones achieved with triangular and trapezoidal-shaped functions. Then, using gaussian functions for the fuzzy partitioning, variations of the number of antecedent terms have been assessed in combination with minimum and product t-norm (Figure 5).

Figure 5:
Performance of different setings of the fuzzy rule-based classification model, from the variations of the number of antecedent terms in combination with minimum and product t-norm.

Analyzing Figure 5, it is noted that, for both t-norms (minimum and product), about 10 antecedent terms were sufficient for the fuzzy classifier to reduce the error to zero during the training, but a higher accuracy over checking dataset required a greater number of terms. In that regard, one of the main aspects to highlight is the difference of performance provided by minimum and product t-norm.

Both product and minimum t-norm allowed aggregating the predictor variables via fuzzy intersections, modeling the simultaneous occurrence of patterns that characterize the same arboreal species. However, the product t-norm operates multiplying all the membership values and, in contrast, the minimum t-norm takes into account only the lowest membership during the aggregation process (Figure 6).

Figure 6:
Aggregation process of predictor variables (x i ) in the rules 1 (R 1) and 2 (R 2), using minimum and product t-norm.

In Figure 6 we have a case in which a given sample has features (pattern values) belonging to more than one arboreal species, i.e, a sample with pertinence in both consequent classes of the rules 1 and 2, but with different membership degrees. By using the minimum t-norm the most critical condition given by the lowest membership become decisive, and hence we have a more rigorous classifier, but which can be naive by disregarding the other predictor variables.

As a consequence, for the case in Figure 6 the minimum t-norm would result in the arboreal species identification supported by the rule 2 (φmin (R 2) . φmin (R 1)). However, the sample has higher membership in the majority of the fuzzy regions correspondent to the consequent of the rule 1, as computed by the product t-norm (φprod (R 1) φprod (R 2)). Thus, by taking account all the predictors, the product t-norm seems to afford a more assertive predictive modeling, so that it provided better performance than minimum t-norm in all settings assessed (see Figure 5).

During the learning process we can note a tendency of accuracy improvement over checking dataset with the increase of the number of fuzzy regions, which was more significant up to about 15 antecedent terms. This improvement seems to occur due to the increase of the decision areas (D j ) formed by each fuzzy if-then rule, as can be seen in Figure 7.

Figure 7:
Increase in decision areas formed by the fuzzy if-then rules as a consequence of the increment of the number of antecedent terms.

Nevertheless, after a certain point there was a performance fluctuation that demanded an exhaustive search for the best accuracy over checking dataset (93.5%), which was found using gaussian curve membership function, product t-norm, and 23 antecedents terms. Then, by using this setting the fuzzy-based model reaches 94.0% accuracy over testing dataset.

4 CONCLUSIONS

In the present study we analyzed the enforceability of fuzzy-based pattern recognition for dealing with complexity related to the natural variability of texture in the arboreal trunk, which can cause uncertainties due to ambiguity in the pattern matching.

By providing a nonlinear and smooth discriminant function, with the differential of taking into account the graded membership of a given sample in different classes (arboreal species), the Fuzzy Rule-Based Classification System (FRBCS) afforded a high generalization ability, which outperformed the most of assessed learning algorithms, including ensembles with a lot of classifiers and kernel-based models, such as some artificial neural networks, widely used in pattern recognition tasks. Therefore, the fuzzy modeling can be considered an alternative approach, with a competitive and reliable performance for arboreal trunk texture recognition, in order to support the tree species identification using computational intelligence.

ACKNOWLEDGMENTS

Supported by the Coordination for the Improvement of Higher Education Personnel (CAPES).

REFERENCES

  • 1
    S. Abe & R. Thawonmas. A fuzzy classifier with ellipsoidal regions. IEEE Transactions on fuzzy systems, 5(3) (1997), 358-368.
  • 2
    V. Adriaenssens, B. De Baets, P.L. Goethals & N. De Pauw. Fuzzy rule-based models for decision support in ecosystem management. Science of the Total Environment, 319(1) (2004), 1-12.
  • 3
    C. Arunpriya & A.S. Thanamani. Fuzzy inference system algorithm of plant classification for tea leaf recognition. Indian Journal of Science and Technology, 8(S7) (2015), 179-184.
  • 4
    J. Ascough, H. Maier, J. Ravalico & M. Strudley. Future research challenges for incorporation of uncertainty in environmental and ecological decision-making. Ecological modelling, 219(3) (2008), 383-399.
  • 5
    E. Bauer & R. Kohavi. An empirical comparison of voting classification algorithms: Bagging, boosting, and variants. Machine learning, 36(1-2) (1999), 105-139.
  • 6
    J. Boman. Tree species classification using terrestrial hotogrammetry. Master’s thesis, Department of Computing Science, Umea University, Umea (2013).
  • 7
    A. Bressane, J.A. Bagatini, C.H. Biagolini, J.A.F. Roveda, S.R.M.M. Roveda, F.H. Fengler & R.M. Longo. Neuro-fuzzy modeling: a promising alternative for risk analysis in urban afforestation management. Brazilian Journal of Forest Science. IN PRESS.
  • 8
    A. Bressane, C.H. Biagolini, P.S. Mochizuki, J.A.F. Roveda & R.W. Lourenço. Fuzzy-based methodological proposal for participatory diagnosis in the linear parks management. Ecological Indicators, 80 (2017), 153-162.
  • 9
    A. Bressane, P.S. Mochizuki, R.M. Caram & J.A.F. Roveda. A system for evaluating the impact of noise pollution on the population’s health. Reports in public health, 32(5) (2016).
  • 10
    A. Bressane, J.A.F. Roveda & A.C. Martins. Statistical analysis of texture in trunk images for biometric identification of tree species. Environmental monitoring and assessment, 187(4) (2015), 212.
  • 11
    Z. Chi, H. Yan & T. Pham. Fuzzy algorithms: with applications to image processing and pattern recognition, volume 10. World Scientific (1996).
  • 12
    A.B. Costello & J.W. Osborne. Best practices in exploratory factor analysis: Four recommendations for getting the most from your analysis. Pan-Pacific Management Review, 12(2) (2009), 131-146.
  • 13
    S. Fiel & R. Sablatnig. Automated identification of tree species from images of the bark, leaves and needles. In 16th Computer Vision Winter Workshop. Citeseer, p. 67. Citeseer (2011).
  • 14
    R.C. Gonzales & R.E. Woods. Digital image processing, volume 3. Pearson Prentice Hall (2008).
  • 15
    R.C. Gonzales, R.E. Woods & S.L. Eddins. Digital image processing using MATLAB, volume 2. Gatesmark Publishing (2009).
  • 16
    A. Gonzblez & R. Pérez. SLAVE: A genetic learning system based on an iterative approach. IEEE Transactions on Fuzzy Systems, 7(2) (1999), 176-191.
  • 17
    R.M. Haralick, K. Shanmugam & I. Dinstein. Textural features for image classification. IEEE Transactions on systems, man, and cybernetics, 3(6) (1973), 610-621.
  • 18
    Z.K. Huang. Bark classification using RBPNN based on both color and texture feature. International Journal of Computer Science and Network Security, 6(10) (2006), 100-103.
  • 19
    Z.K. Huang, C.H. Zheng, J.X. Du & Y.y. Wan. Bark classification based on textural features using artificial neural networks. In International Symposium on Neural Networks, pp. 355-360. Springer (2006).
  • 20
    H. Ishibuchi & T. Nakashima. Effect of rule weights in fuzzy rule-based classification systems. IEEE Transactions on Fuzzy Systems, 9(4) (2001), 506-515.
  • 21
    A. Lermontov, L. Yokoyama, M. Lermontov & M.A.S. Machado. River quality analysis using fuzzy water quality index: Ribeira do Iguape river watershed, Brazil. Ecological Indicators, 9(6) (2009), 1188-1197.
  • 22
    D. Liu & Z. Zou. Water quality evaluation based on improved fuzzy matter-element method. Journal of Environmental Sciences, 24(7) (2012), 1210-1216.
  • 23
    L. Liu, J. Zhou, X. An, Y. Zhang & L. Yang. Using fuzzy theory and information entropy for water quality assessment in Three Gorges region, China. Expert Systems with Applications, 37(3) (2010), 2517-2521.
  • 24
    D. Nauck & R. Kruse. A neuro-fuzzy method to learn fuzzy classification rules from data. Fuzzy sets and Systems, 89(3) (1997), 277-288.
  • 25
    W. Pedrycz & F. Gomide. Fuzzy systems engineering: toward human-centric computing. John Wiley & Sons (2007).
  • 26
    A. Porebski, N. Vandenbroucke & L. Macaire. Iterative feature selection for color texture classification. In Image Processing, 2007. ICIP 2007. IEEE International Conference on, volume 3, pp. III-509. IEEE (2007).
  • 27
    L.S. Riza, C.N. Bergmeir, F. Herrera & J.M. Benítez Sánchez. frbs: Fuzzy rule-based systems for classification and regression in R. Journal of Statistical Software, 65(6) (2015), 1-30.
  • 28
    W. Silvert. Fuzzy indices of environmental conditions. Ecological Modelling, 130(1) (2000), 111-119.
  • 29
    H. Singh, M.M. Gupta, T. Meitzler, Z.G. Hou, K.K. Garg, A.M. Solo & L.A. Zadeh. Real-Life Applications of Fuzzy Logic. Adv. Fuzzy Systems, 2013 (2013), 581879-1.
  • 30
    Y.Y. Wan, J.X. Du, D.S. Huang, Z. Chi, Y.M. Cheung, X.F. Wang & G.J. Zhang. Bark texture feature extraction based on statistical texture analysis. In Intelligent Multimedia, Video and Speech Processing, 2004. Proceedings of 2004 International Symposium on, pp. 482-485. IEEE (2004).
  • 31
    A.G. Yong & S. Pearce. A beginner´s guide to factor analysis: Focusing on exploratory factor analysis. Tutorials in Quantitative Methods for Psychology, 9(2) (2013), 79-94.
  • 32
    L.A. Zadeh. Fuzzy sets. Information and control, 8(3) (1965), 338-353.
  • 33
    L.A. Zadeh. Is there a need for fuzzy logic? Information sciences, 178(13) (2008), 2751-2779.

Publication Dates

  • Publication in this collection
    Jan-Apr 2018

History

  • Received
    19 Mar 2017
  • Accepted
    22 Jan 2018
Sociedade Brasileira de Matemática Aplicada e Computacional Rua Maestro João Seppe, nº. 900, 16º. andar - Sala 163 , 13561-120 São Carlos - SP, Tel. / Fax: (55 16) 3412-9752 - São Carlos - SP - Brazil
E-mail: sbmac@sbmac.org.br