Acessibilidade / Reportar erro

Design of Waveguide Structures Using Improved Neural Networks

Abstract

In this paper, an improved neural networks (INN) strategy is proposed to design two waveguide filters (Pseudo-elliptic waveguide filter and Broad-band e-plane filters with improved stop-band). INN is trained by an efficient optimization algorithm called teaching–learning-based optimization (TLBO). To validate the effective of this proposed strategy, we compared the results of convergence and modeling obtained with a population based algorithm that is widely used in training NN namely Particle Swarm Optimization (PSO-NN). The results show that the proposed INN has given better results.

Index Terms
Improved neural networks; modeling; teaching–learning-based optimization; waveguide filters

I. INTRODUCTION

The full wave EM solvers [11 S.M. Ali, N.K. Nikolova, and M.H. Bakr, “Sensitivity Analysis with Full-Wave Electromagnetic Solvers Based on Structured Grids,” IEEE Transactions on Magnetics, vol. 40, pp.1521–1529, 2004.] have been used to design the microwave filter for a long time. With the increasing complexity of wireless and satellite communication hardware, the use of EM solver takes a considerable amount of time. Recently, artificial neural network (ANN) has been proven to be a fast and effective means of modeling complex electromagnetic devices. It has been recognized as a powerful tool for predicting device behavior for which no mathematical model is available or the device has not been analyzed properly yet. ANN can be trained to capture arbitrary input-output relationship to any degree of accuracy. Once a model is developed it can be used over and over again. The trained model delivers the output parameters very quickly. For these attractive qualities, ANN has been applied to different areas of engineering's [22 Y. Wang, M. Yu, H. Kabir, and Q.J. Zhang, “Application of Neural Networks in Space Mapping Optimization of Microwave Filters,” International Journal of RF and Microwave Computer Aided Engineering, vol. 22, pp. 159–166, 2012.] - [44 A. A. Deshmukh, S.D. Kulkarni, A.P.C. Venkata, and N.V. Phatak, “Artificial Neural Network Model for Suspended Rectangular Microstrip Antennas,” Procedia Computer Science, vol. 49, pp. 332–339, 2015].

Training of neural networks (NN) is an important step; it is based on optimization of weights of NN to minimize the mean square error (MSE) between the NN output and the desired output. Specialized learning algorithms are used to adapt these weight values. Among those algorithms, the most popular algorithm is the back-propagation method (BP) [55 D.J. Jwo, and K.P. Chin, “Applying Back-propagation Neural Networks to GDOP Approximation,” The Journal of Navigation, vol. 55, pp. 97–108, 2002.] based on a gradient descending. Lately, many populations based algorithms have been proposed for training a neural network such as Particle Swarm Optimization (PSO) [66 D. Gyanesh, K. P. Prasant, and K. P. Sasmita, “Artificial Neural Network Trained by Particle Swarm Optimization for Non-Linear Channel Equalization,” Expert Systems with Applications, vol. 41, pp. 3491–3496, 2014.], Genetic Algorithms [77 S. Ding, Y. Zhang, J. Chen, and J. Weikuan, “Research on Using Genetic Algorithms to Optimize Elman Neural Networks,” Neural Computing and Applications, vol. 23, pp. 293–297, 2013.] and other optimization algorithms [88 K. Khan, and A. Sahai, “A Comparison of BA, GA, PSO, BP and LM for Training Feed Forward Neural Networks in E-Learning Context,” International Journal of Intelligent Systems and Applications, vol.4, pp. 23 – 29, 2012.].

In this paper, we tried to improve the NN by training them by a recent and effective optimization algorithm called Teaching-Learning Based Optimization (TLBO) [99 R. V. Rao, V. Patel, “An improved teaching-learning-based optimization algorithm for solving unconstrained optimization problems,” Scientia Iranica, vol 20, pp. 710–720, 2014.]. The trained networks are applied to modeling two waveguide Filters (Pseudo-elliptic waveguide filter and Broad-band e-plane filters with improved stop-band).We use a multilayer perceptron neural network (MLP-NN) to three layers. Each sub-net in the NN architecture shown in Fig. 1, possesses Ne input neurons corresponding to the number of the geometry parameters of the structures, Nc neurons in the hidden layer and one output associated with the value of Sij(fk). The entire network consists of k distinct NNs corresponding to a particular frequency with k determined by the number of approximate points in the frequency interval. Frequency responses of S parameters obtained in simulations compose the network database. The connection weight from the neurons of the input layer to the neurons of the hidden layer is WE and the connection weight from the neurons of the hidden layer to the neurons of the output layer is WS.

Fig. 1
Neural network architecture.

II. TEACHING LEARNING BASED OPTIMIZATION

In 2011, Rao et al. [1010 R.V. Rao, V. J. Savsani, and D. P. Vakharia, “Teaching–Learning-Based Optimization: A Novel Method for Constrained Mechanical Design Optimization Problems,” Computer-Aided Design, vol.43, pp. 303–315, 2011.], [1111 R.V. Rao, V. J. Savsani, and D. P. Vakharia, “Teaching–Learning-Based Optimization: An Optimization Method for Continuous Non-Linear Large Scale Problems,” Information Sciences, vol.183, pp. 1–15, 2012.] proposed an algorithm, called Teaching-Learning-Based Optimization (TLBO), based on the traditional Teaching Learning phenomenon of a classroom. TLBO is a population based algorithm, where a group of students (i.e. learner) is considered as population and the different subjects offered to the learners are analogous with the different design variables of the optimization problem. The results of the learner are analogous to the fitness value of the optimization problem. In 2014, Rao and Patel [99 R. V. Rao, V. Patel, “An improved teaching-learning-based optimization algorithm for solving unconstrained optimization problems,” Scientia Iranica, vol 20, pp. 710–720, 2014.] are improving the basic TLBO to enhance its exploration and exploitation capacities by introducing the concept of number of teachers and adaptive teaching factor. By means of this modification, the entire class is split into different groups of learners as per their level (i.e. results), and an individual teacher is assigned to an individual group of learners. Each teacher tries to improve the results of his or her assigned group based on two phases: the teacher and the learner phase.

A. Teacher phase

In this phase, learners of each group take their knowledge directly through the teacher, where a teacher tries to increase the mean result value of the classroom to another value, which is better than, depending on his or her capability. This follows a random process depending on many factors. In this work, the value of solution is represented as (Xj,k)S, where j means the jth design variable (i.e. subject taken by the learners), k represents the kth population member (i.e. learner), and S represents the Sth group. The existing solution is updated according to the following expression

(1) ( X j , k ) s = ( X j , k + D j ) s + r a n d * ( X h X k ) s , i f f ( X ) k > f ( X ) h
(2) ( X j , k ) s = ( X j , k + D j ) s + r a n d * ( X k X k h ) s , i f f ( X ) h > f ( X ) k

The above equations are for a minimization problem, the reverse is true for a maximization problem. Dj is the difference between the current mean and the corresponding result of the teacher of that group for each subject calculated by:

(3) ( D j ) s = r a n d * ( X j , T e a c h e r T F * M j ) s

Where hk, Mj is the mean result of each group of learners in each subject and TF is the adaptive teaching factor in each iteration given by equation:

(4) T F = M X k b e s t

B. Learner phase

In this part, each group update the learners’ knowledge with the help of the teacher's knowledge, along with the knowledge acquired by the learners during the tutorial hours, according to:

(5) ( X j , k ) s = X j , k + r a n d * ( X j , k X j , p ) s + r a n d * ( X t e a c h e r E F * X j , k ) s , i f f ( X p ) > f ( X k )
(6) ( X j , k ) s = X j , k + r a n d * ( X j p X j , k ) s + r a n d * ( X t e a c h e r E F * X j , k ) s , i f f ( X k ) > f ( X p )

Where EF = exploration factor = round (1 + rand). The above equations are for a minimization problem, the reverse is true for a maximization problem.

III. Improved neural networks

The training of neural networks is to find an algorithm for optimized weights of networks to minimize the mean square error (MSE) described as follows

(7) M S E = 1 P T * P T N s ( Y s Y ) 2

Where PT the total number of training samples, YS is the output of the network and Y is the desired output.

(8) Y s = f 2 ( N c W S * f 1 ( N e W E * X ) )

With f2 and f1 are the activation functions (typically: sigmoid, tanh …), X is the input vector of NN.

Regarding the NN training, the most used training algorithm is the back-propagation (BP) algorithm, which is a gradient-based method. In this paper, we propose a new strategy of training a neural network based on a recently proposed algorithm called Teaching-Learning Based Optimization (TLBO) [99 R. V. Rao, V. Patel, “An improved teaching-learning-based optimization algorithm for solving unconstrained optimization problems,” Scientia Iranica, vol 20, pp. 710–720, 2014.]; the details of this strategy are presented in the next section.

A. Implementation of TLBO on the neural networks

The step-wise procedure for the implementation of (TLBO-NN) is given in this section.

Step.1. Define the neural network architecture (number of neurons in input layer Ne, number of neurons in hidden layer Nc and number of neurons in output layer Ns) and define the optimization problem:

  • -

    Design variables of the optimization problem (i.e. number of subjects offered to the learner): WE and WS the matrices of input connection weights and output connection weights respectively. WE matrix of Nc rows and Ne columns and WS matrix of Ns rows and Nc columns.

  • -

    The optimization problem (fitness function): find the optimal WE and WS which minimizes the mean square error (MSE) equation (7).

Step.2. Initialize the optimization parameters and initialize the population

  • -

    Population size (number of learners in a class) NP

  • -

    Number of generations (maximum number of allowable iterations) maxit

  • -

    Number of teachers NT

  • -

    The initial population according to the population size and the number of neurons and evaluate the corresponding objective function value. For simplification, the population is decomposed into two groups one represents the inputs weights population WEp and the second one represents the output weights population WSp.

(9) W E p = r a n d * ( N P * N c , N e )
(10) W S p = r a n d * ( N s , N P * N c )

Step.3. Select the best solution who acts as chief teacher for that cycle, and select randomly the other teachers.

Step.4. Arranged the teachers according to their fitness value, then assigns the learners to the teachers [99 R. V. Rao, V. Patel, “An improved teaching-learning-based optimization algorithm for solving unconstrained optimization problems,” Scientia Iranica, vol 20, pp. 710–720, 2014.].

Step.5. Calculate the mean result of each group of learners in each design variables: MWE, MWS.

Step.6. For each group, evaluate the difference between the current mean and the corresponding result of the teacher of that group for each design variables by utilizing the adaptive teaching factor [1212 R. V. Rao, D. P. Rai, “Optimization of fused deposition modeling process using teaching-learning-based optimization algorithm,” Engineering Science and Technology, an International Journal, 19, pp. 587–603, 2016.]. given by equation(4).

Step.7. For each group, update the solution using teacher- phase equations (1) (2), and learner-phase equations (5) (6).

Step.8. Combine all the groups.

Step.9. Repeat the procedure from step 3 to 8 until the termination criterion is met.

B. Application examples and results

In this part, the performance of the improved neural networks (INN) is investigated for modeling two microwave filters: Pseudo-elliptic waveguide filter Fig. 2 [1313 Q. Zhang, and Y. Lu, “Design of Wide-Band Pseudo-Elliptic Waveguide Filters with Cavity-Backed Inverters,” IEEE Microwave And Wireless Components Letters, vol. 20, pp. 604–606, 2010.] and Broad-band e-plane filters with improved stop-band Fig. 3 [1414 X. Zhengbin, G. Jian, Q. Cheng, and D. Wenbin D, “Broad-Band E-plane filters with improved stop-band performance,” IEEE Microwave and Wireless Components Letters, vol. 21, pp. 350–352, 2011.]. The dimensions of the first and second filters are listed in Table I and Table II respectively.

Fig. 2
The pseudo-elliptic waveguide filter, a) Top view, b) Side cross section view.
Fig. 3
Broad-band E-plane filter, a) The designs of broad-band E-plane filter, b) Fin-line post structure.
Table I
Dimensions for the Designed Filter (units: millimeters).
Table II
Dimensions for the Designed Filter (units: millimeters).

For modeling the two structures above-mentioned. We propose multilayer feed-forward neural network architecture with a single hidden layer. We begin by selecting the input parameters and creating database, the latter starts by creating a list of points from the matrix of bounds of the input parameters. The list of the database points is of size (Ne, PNe), Ne is the number of input parameters, and P is the number of points for each parameter. In this work, we choose an equal number of points for each variable P=4, Ne=5 for the first filter corresponding to (W1, W2, W3, V1, V2) with a database equal 1024 and Ne=3 for the second filter corresponding to (W1, W2, W3) with a database equal 64. The bounds of each parameter are presented in Table III. The number of output Ns=1 corresponding to the frequency responses of the Sij parameters. For the frequency range, we chose to be (8, 12 GHz) with a number of points K=41 and (26, 36 GHz) with K=34 for the first and the second filter respectively. The choice of the number of hidden neurons is strongly related to the nature of nonlinearity to model. In our application examples, the number of hidden neurons gives a good convergence of the algorithm and a good accuracy of the neural model formed are Nc=8 for the first filter and Nc=6 for the second filter. The activation functions are hyperbolic tangent function (Tansig), and linear function (Purelin) respectively. When the architecture of NN is selected, the next step is to train NN using TLBO algorithm section (III. A). We begin by initializing the connection weights equations (9) and (10). Once the learning is complete, we obtained the update WE and WS and we can approximate the S parameters response to any input parameter in the boundary range using equation (8).

Table III
INPUT PARAMETERS AND THEIR LIMITS

Fig. 4 and Table IV shows the convergence of PSO and TLBO algorithms for minimize the MSE of NN for the filters above-mentioned with the effect of the number of teachers in TLBO algorithm. The common parameters for algorithms (population size NP=50, number of iterations is 300 for the first filter, and 100 for the second filter), the other specific parameters of the algorithms are given below.

Fig. 4
Convergence of algorithms for minimizing the MSE, a) first filter, b) second filter.
Table IV
Comparative results of convergence of MSE

PSO Settings: c1and c2are constant coefficients c1 = c2 = 2, w is the inertia weight decreased linearly from 0.9 to 0.2.

TLBO Settings: for TLBO there is no such constant to set.

It is observed from Fig. 4 and Table IV that, the TLBO (NT=7) algorithm perform better in terms of convergence than the PSO algorithm, in which this algorithm requires less number of iterations to converge to optimum solution as compared to PSO algorithm. Fig.5 and Fig.6 gives the approximate parameters S11, S21 (magnitude and phase) for the first and second filters respectively, an excellent approximate can be observed.

Fig. 5
The approximate parameters S11, S21 (magnitude and phase) for the first filter
Fig. 6
The approximate parameters S11, S21 (magnitude and phase) for the second filter

IV. Conclusion

In this paper, Teaching–Learning-Based Optimization (TLBO) algorithm is proposed to training and testing feed-forward neural networks (FNN) for modeling waveguide filter structures (Pseudo-elliptic waveguide filter and Broad-band E-plane filters with improved stop-band). The results show the efficiency of TLBO algorithm, where TLBO algorithm converges to the global minimum faster than PSO algorithm. The main advantage of this algorithm does not require selection of the algorithmspecific parameters.

REFERENCES

  • 1
    S.M. Ali, N.K. Nikolova, and M.H. Bakr, “Sensitivity Analysis with Full-Wave Electromagnetic Solvers Based on Structured Grids,” IEEE Transactions on Magnetics, vol. 40, pp.1521–1529, 2004.
  • 2
    Y. Wang, M. Yu, H. Kabir, and Q.J. Zhang, “Application of Neural Networks in Space Mapping Optimization of Microwave Filters,” International Journal of RF and Microwave Computer Aided Engineering, vol. 22, pp. 159–166, 2012.
  • 3
    J. S. Sivia, A. P. S. Pharwaha, and T. S. Kamal, “Analysis and Design of Circular Fractal Antenna Using Artificial Neural Networks,” Progress in Electromagnetics Research B, vol. 56, pp. 251–267, 2013.
  • 4
    A. A. Deshmukh, S.D. Kulkarni, A.P.C. Venkata, and N.V. Phatak, “Artificial Neural Network Model for Suspended Rectangular Microstrip Antennas,” Procedia Computer Science, vol. 49, pp. 332–339, 2015
  • 5
    D.J. Jwo, and K.P. Chin, “Applying Back-propagation Neural Networks to GDOP Approximation,” The Journal of Navigation, vol. 55, pp. 97–108, 2002.
  • 6
    D. Gyanesh, K. P. Prasant, and K. P. Sasmita, “Artificial Neural Network Trained by Particle Swarm Optimization for Non-Linear Channel Equalization,” Expert Systems with Applications, vol. 41, pp. 3491–3496, 2014.
  • 7
    S. Ding, Y. Zhang, J. Chen, and J. Weikuan, “Research on Using Genetic Algorithms to Optimize Elman Neural Networks,” Neural Computing and Applications, vol. 23, pp. 293–297, 2013.
  • 8
    K. Khan, and A. Sahai, “A Comparison of BA, GA, PSO, BP and LM for Training Feed Forward Neural Networks in E-Learning Context,” International Journal of Intelligent Systems and Applications, vol.4, pp. 23 – 29, 2012.
  • 9
    R. V. Rao, V. Patel, “An improved teaching-learning-based optimization algorithm for solving unconstrained optimization problems,” Scientia Iranica, vol 20, pp. 710–720, 2014.
  • 10
    R.V. Rao, V. J. Savsani, and D. P. Vakharia, “Teaching–Learning-Based Optimization: A Novel Method for Constrained Mechanical Design Optimization Problems,” Computer-Aided Design, vol.43, pp. 303–315, 2011.
  • 11
    R.V. Rao, V. J. Savsani, and D. P. Vakharia, “Teaching–Learning-Based Optimization: An Optimization Method for Continuous Non-Linear Large Scale Problems,” Information Sciences, vol.183, pp. 1–15, 2012.
  • 12
    R. V. Rao, D. P. Rai, “Optimization of fused deposition modeling process using teaching-learning-based optimization algorithm,” Engineering Science and Technology, an International Journal, 19, pp. 587–603, 2016.
  • 13
    Q. Zhang, and Y. Lu, “Design of Wide-Band Pseudo-Elliptic Waveguide Filters with Cavity-Backed Inverters,” IEEE Microwave And Wireless Components Letters, vol. 20, pp. 604–606, 2010.
  • 14
    X. Zhengbin, G. Jian, Q. Cheng, and D. Wenbin D, “Broad-Band E-plane filters with improved stop-band performance,” IEEE Microwave and Wireless Components Letters, vol. 21, pp. 350–352, 2011.

Publication Dates

  • Publication in this collection
    Dec 2017

History

  • Received
    05 Apr 2017
  • Reviewed
    06 Apr 2017
  • Accepted
    21 Sept 2017
Sociedade Brasileira de Microondas e Optoeletrônica e Sociedade Brasileira de Eletromagnetismo Praça Mauá, n°1, 09580-900 São Caetano do Sul - S. Paulo/Brasil, Tel./Fax: (55 11) 4238 8988 - São Caetano do Sul - SP - Brazil
E-mail: editor_jmoe@sbmo.org.br