^{1}

^{*}

^{2}

^{*}

^{3}

In this paper, artificial neural networks are used for predicting single fiber efficiency in the process of removing smaller particles from gas stream by fiber filters. For this, numerical simulations are obtained of a classic model of literature for fiber efficiency, which is numerically solved along with the convection diffusion equation in polar coordinates for particle concentration, with associated initial and boundary conditions. A sufficient number of examples from two numerical simulations are employed to construct a database, from which parameters of a novel neural model are adjusted. This model is constructed based on the back propagation algorithm in order to map two features, namely Peclet number and packing density, which are extracted from the numerical simulations into the corresponding single fiber efficiency. The results indicate that the developed neural model can be trained in a reasonable computational time and is capable of estimating single fiber efficiency from examples of the test set with a maximum error of 1.7%.

Concern with the increasing air pollution has stimulated the development of new techniques that operate in the efficiency of solid-gas separation, mainly for smaller particles. Filtration by fibrous filters is one of the most widely used techniques to capture fine particles from a gas stream. The process takes place by passing the aerosol through the filter media, so that the particles will be deposited on the filter surface [

The performance of filtration is evaluated, essentially, for two parameters, the pressure drop and collection efficiency. For this, several authors have theoretically studied the efficiency of particle deposition on fibers. Lee and Liu [

Thus, it is evident from the works above that theoretical studies are of prime importance for understanding the performance of fibrous filters. Also, it is important to observe that, combined with artificial neural network (ANN), predictions from parameters of this process may become an important factor for innovation and development of new techniques. In addition, ANNs have several advantages that make them attractive tools: they can be trained to learn from examples; are easily updated and generalized; possess a large degree of freedom and accurate prediction at high speed [

Recently, ANNs models have been successfully applied to various branches of science and technology, such as medicine [

The objective of this work is to investigate the possibility of using a neural network to directly assess the single fiber efficiency. Thus, a much more robust measurement method can be developed. For this, in order to construct a database for developing a neural model, a mathematical model is implemented and numerical results are obtained for several Peclet numbers and packing density, which are important parameters in this process.

This section aims to present the mathematical modelling to simulate the process of particle deposition on a single fiber to construct a database for an ANN model.

As the structure of fibrous filters is complex, many works have adopted cell models in which fluid flow equations are solved in a two dimensional cell surrounding the single fiber, where the boundary conditions take into account neighboring fibers [^{−1/2}) is the dimensionless cell radius that is determined by the filter packing density a and U is the flow velocity scale [

In this study, the deposition of aerosol particles in fibrous filters is considered to occur exclusively because of diffusion, which is one of the major particle collection mechanisms, especially when dealing with smaller particles (see [

The single fiber efficiency is mathematically expressed by

where n is the particle concentration and Pe is the Peclet number.

To determine numerically the value of the integral in Equation (1), the following dimensionless convection diffusion partial differential equation is utilized, which in polar coordinates r and θ is given by [

where ur and uθ are, respectively, the radial and angular components of the gas velocity, expressed as

with Ψ being the dimensionless Kuwabara function [

For fibrous filtration, the third term on the right-hand side of Equation (2) is much smaller than the other terms and can be neglected to give [

The adopted boundary conditions are

Equation (4) associated with the boundary conditions (5)-(7) was discretized by finite-difference methodology on a staggered grid system; the diffusive terms were approximate by second order central differences and the non-linear convection terms were approximated by the TOPUS upwind scheme [

where f is a boundary face of a computational cell (see

corresponds to the normalized variables of Leonard [_{1}, c_{2} and c_{3} are expressed as c_{1} = −2b + 1, c_{2} = (5b − 10)/4 and c_{3} = (−b + 10)/4, where b is an adjustable real parameter chosen in the interval [−2, 2] to attend the total variation diminishing (TVD) condition of Harten [_{f} (Downstream), U_{f} (Upstream) and R_{f} (Remoteupstream) of a computational face f are prescribed according to the advection velocity v_{f} (u_{r} or u_{q}) at this face, as can be observed in

this study since the use of conventional schemes (central differences and first order upwind) has been unsatisfactory for predicting single fiber efficiency in the range 100 £ Pe £ 165, 0.2 £ a £ 0.3.

The linear system resulting from the discretization of Equation (4) was solved by the Gauss-Seidel iteration procedure. The numerical results were compared favorably with experimental data of literature and published in [

In this section, the basic concepts of ANNs, the back propagation algorithm and a new neural model for predic- ting h, in the process of removing aerosol particles, are introduced.

In short, ANNs are a form of artificial intelligence composed of a network of connected nodes (neurons). Formally, an ANN can be defined as nonlinear mapping of an input onto an output vector space. This is achieved through layers of neurons in which the input coordinates are summed according to specific weights and biases to produce a single output. Finding suitable architecture for ANNs is extremely critical; most works in the literature use trial and error to adjust the parameters of a neural network (see, for example, [

In this study, a four-layer feed-forward ANN with two hidden layers was employed and the back propagation algorithm was used for the training. No recursiveness was adopted here, that is, the input vector of a specific neuron layer was formed only by the values of the preceding layer. A schematic representation of a typical ANN structure is depicted in

The notation in a feed-forward network is as follows. If the activation function of i-th neuron in the j-th layer is indicated by Fi j(.), its output si, j can be calculated from outputs of the preceding layer si j-1 and from the corre-sponding biases bi,j and weights wi,k,j-1 (the second subscript k indicates the neuron in the (j − 1)-th layer from which the connection is being established), according to

After denoting the input and output network values by σ_{i}, i = 1, 2, 3, 4 and ξ_{j}, j = 1, 2, respectively, the mapping relationship of one onto the other can then be done by applying Equation (9). For instance, the mapping in

Equation (10) makes it clear that the relationship between σ_{i} and ξ_{j} is unambiguously defined by choosing the activation functions and by setting the biases and weights. One very important characteristic of ANNs, among others, is the so-called learning potential, that is, the possibility of adjusting the biases and weights through a convenient training rule to closely reproduce pre-assigned pairs of input/output values. Back propagation is probably the most used training heuristic and it is particularly well adapted to feed-forward architecture. It is based on the iterative application of a discrete gradient descent algorithm, computed from the first derivatives of a conveniently defined error function whose arguments are the parameters of the network (weights and biases). In general, the basic steps of the back propagation procedure, which is implemented in this work, are the following [

1) Initialize the ANN's parameters bi,j and wi,k,j with random numbers;

2) Take the p-th (σ_{i}^{p}, ξ_{i}^{p}) pair from a training data set with pre-assigned input/output pairs;

3) Calculate the outputs with the same input and form the pair (σ_{i}^{p}, ξ_{i}^{p});

4) Evaluate the error e between the desired and the obtained output values according to the L2 norm

5) Calculate the derivatives of the error with respect to bi,j and wi,k,j;

6) Modify the ANN parameters according to the steepest descent strategy and a specified learning rate γ

7) Iterate from 2 to 5, by successively modifying bi,j and wi,k,j, until a defined number of learning epochs (cycles) or a convenient stopping criterion has been achieved.

The performance of an ANN is affected by internal architecture (number of hidden layers and number of neurons in each one) and type of interconnections (feed-forward, recursiveness, winner-take-all, etc.). The exact shape of the activation function has limited effects on the overall performance and is usually set according to the needs of the training heuristics (a sigmoid function in the case of the back propagation method). There is no general mathematical theory but rather a number of empirical rules that must be considered when constructing such models.

The ANN implemented in this work has two inputs and one output. Various architectures were trained and tested; the neural model that provided good results, obtained by trial and error, was that containing two intermediate layers with six and three neurons. The number of neurons in the intermediate layers was adjusted to enable the ANN to learn complex tasks for the gradual extraction of significant information from the inputs [

where

The neuron of the output layer is responsible for estimating single fiber efficiency (η). The training procedure uses the acquisition of the Peclet number (Pe) and packing density (α) for the filter to a sufficient level for inference of the η. Two numerical simulations were performed with Pe and α varying uniformly in the ranges: a) 20 ≤ Pe ≤ 80 and 0.1 ≤ α ≤ 0.2; and b) 100 ≤ Pe ≤ 165 and 0.2 ≤ α ≤ 0.3. Equation (1) was used for estimating the η.

In both simulation cases, the input data matrix has 2 lines for 165 columns with the line number correspon- ding to the number of inputs (Pe and α) and the column to the examples used in the training of the ANN; each one corresponds to a known and fixed value of the η. The output matrix has 1 line for 165 columns, where 1 is the number of desired output and 165 is the number of examples. Generalization was considered: 88 different examples, from those used in the training process, were presented to the ANN in the first simulation; and 66 examples in the second simulation. Therefore, in the first calculation, 253 pairs of input/output values were obtained to construct the database, and in the second calculation 231 pairs were employed. The division of the examples in the training and test sets was done randomly and some divisions were tested. The ANN results for all the tested divisions were shown to be practically equal, with the same correlation coefficients for the data.

The purpose of this section is that it presents the training error of neural network developed and the results obtained for the data of the test.

The results of the training of the ANN have shown that it is capable of reproducing the input/output relationship of the training set data. To evaluate the generalization capacity of the ANN, the characteristic data of new nu-merical tests were presented to the neural networks and a good correlation between the input and output data was observed.

In attempting to reduce the error in a reasonable time, optimization of the ANN parameters (learning rate, epochs and number of neurons in the intermediate layers) was performed. The learning rate used in both tests was 0.1 and 12 × 10^{4} epochs were considered, with a training time of 36 minutes for the first case, and 44 minutes for the second one. As can be observed from ^{−4}).

^{2}) between the numerical values and the ANN values was 0.9996, showing that there was a strong relationship between the input and output data provided by the ANN; whereas, in the second range, the best representation for η is y = 1.0021x − 0.0003 with a correlation coefficient of 0.9995. These results indicate that training of the ANNs was successfully accomplished.

In summary, the results obtained through the ANN for the test set examples were quite satisfactory, with maximum relative errors of 1.7% for the first case and 0.9% for the second case.

_{Numeric}) and the η estimated by the ANN (η_{ANN}).

20 ≤ Pe ≤ 80; 0.1 ≤ α ≤ 0.2 | 100 ≤ Pe ≤ 165; 0.2 ≤ α ≤ 0.3 | ||
---|---|---|---|

η_{Numeric } | η_{ANN } | η_{Numeric } | η_{ANN } |

0.4009 | 0.3993 | 0.1610 | 0.1604 |

0.3131 | 0.3130 | 0.1424 | 0.1422 |

0.2609 | 0.2619 | 0.1251 | 0.1256 |

0.2034 | 0.2025 | 0.1764 | 0.1760 |

0.1832 | 0.1833 | 0.1339 | 0.1336 |

0.1579 | 0.1557 | 0.1790 | 0.1786 |

0.4072 | 0.4073 | 0.1413 | 0.1417 |

0.2332 | 0.2321 | 0.1505 | 0.1505 |

0.1704 | 0.1716 | 0.1604 | 0.1605 |

0.1597 | 0.1590 | 0.1702 | 0.1698 |

0.3254 | 0.3243 | 0.1582 | 0.1585 |

0.1749 | 0.1751 | 0.1815 | 0.1813 |

0.2009 | 0.2010 | 0.1439 | 0.1439 |

0.1873 | 0.1862 | 0.1790 | 0.1786 |

0.2154 | 0.2154 | 0.1244 | 0.1247 |

0.3047 | 0.3065 | 0.1512 | 0.1511 |

0.3115 | 0.3113 | 0.1175 | 0.1171 |

0.2612 | 0.2624 | 0.1315 | 0.1312 |

0.2052 | 0.2051 | 0.1500 | 0.1504 |

0.3390 | 0.3371 | 0.1214 | 0.1209 |

Artificial neural models were trained with data derived from numerical simulations in order to determine smaller particle capture efficiency in a single fiber. To construct a database, a sufficient number of examples were per-formed, from which the ANN parameters (weights and biases) were adjusted. A model was developed to map the parameters Pe and α with the aim of estimating the single fiber efficiency, η. The developed ANN was shown to be capable of estimating correct values for fiber efficiency from examples not contemplated in the training.

From this research, it can be inferred that, in the context of fiber filter efficiency, the developed neural model is an effective tool for performing nonlinear mapping of the parameters. The simplicity and efficiency of the pro-posed neural approach indicates that the present methodology can be effectively used in the process studied. Furthermore, it should be emphasized still that, although numerical data have been used in this work, the results obtained suggest that if the experimental data are available, these data could have been used to train the neural network which would produce a very powerful prediction tool.

The authors wish to express their thanks to the Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq-Grants No. 475390/2011-3) for the financial support.

Érica Regina Filletti,Juliana Maria da Silva,Valdemir Garcia Ferreira, (2015) Predicting of the Fibrous Filters Efficiency for the Removal Particles from Gas Stream by Artificial Neural Network. Advances in Chemical Engineering and Science,05,317-327. doi: 10.4236/aces.2015.53033