• Nu S-Au Găsit Rezultate

View of Feature Selection using Multi-Verse Optimization for Brain Tumour Classification

N/A
N/A
Protected

Academic year: 2022

Share "View of Feature Selection using Multi-Verse Optimization for Brain Tumour Classification"

Copied!
13
0
0

Text complet

(1)

Feature Selection using Multi-Verse Optimization for Brain Tumour Classification

Katkoori Arun Kumar1, Ravi Boda2, Dr. Praveen Kumar3, Dr. Vijendar Amgothu4

1Research Scholar, Electronics and Communication Engineering Department, KL University, Hyderabad, India.

e-mail: [email protected]

2Associate Professor, Electronics and Communication Engineering Department, KLEF Off Campus, Hyderabad, India.

e-mail: [email protected]

3Associate Professor, Electronics and Communication Engineering Department, KLEF Off Campus, Hyderabad, India.

e-mail: prawinpoola@klh.edu.in

4Associate Professor, Electronics and Communication Engineering Department, Malla Reddy Engineering College, Hyderabad, India.

e-mail: [email protected] Abstract:

Selecting the feature will reduce the dimensionality of the characteristics and the consistency of the reserved features can be maintained in the choir. The selection of features represents a machine-learning optimization challenge, reducing the number of features, eliminating obsolete, noisy, and repetitive data, and achieving suitable acknowledgement. In this paper, Multi-verse Optimization (MVO) is used to select the features from segmented MR image with Random Forest classifier. The steps involved for feature selection- pre- processing, tumor segmentation, feature extraction, feature selection and classification. Two different objective functions are used- image entropy is used as objective function for tumor segmentation and accuracy function is used as objective function for feature classification.

Experimental results show that performance of MVO+RF approach gives high accuracy compared to existing techniques like PSO and BAT.

Keywords:MR Image, Image segmentation, Feature extraction, feature selection, MVO, Random-Forest classifier.

Introduction:

Tumors are masses or lumps of tissue that resemble swelling. Although not all aTumors are cancerous, it is a good idea to consult a doctor if one appears. A tumor, according to the National Cancer Institute, is “an abnormal mass of tissue that results when cells divide more than they should or do not die when they should”. Cells in a healthy body grow, divide, and replace one another. The old cells die as new one‟s form. When a person is diagnosed with cancer, new cells form even when the body does not require them. When there are too many new cells, a cluster of cells, or tumor, can form. Some tumors are „benign‟

and made up of non-cancerous cells, while others are „malignant‟. Malignant tumors are cancerous, and cells can spread throughout the body.

MR Imaging is frequently employed in disease detection, diagnosis, and treatment monitoring and is based on advanced technology that excites and detects changes in the rotational axis of protons found in the water that makes up living tissues[13][17][18][19].

These tissuescan be distinguished by two relaxation times: T1and T2. T1 is called longitudinal relaxation time, is the time constant that governs how quickly excited protons

(2)

return to equilibrium. It is the time it takes for spinning protons to realign with the external magnetic field. T2 is called transverse relaxation time, is the time constant that governs how quickly excited protons reach equilibrium or fall out of phase with one another. It is the time it takes for spinning protons to lose phase coherence among nuclei spinning perpendicular to the main field.

Feature selection is a hot topic in machine learning, pattern recognition, and data mining [1][2][13]. Irrelevant and redundant features encourage additional search because they make patterns less detectable and rules required for forecasting or classification less obvious, in addition to the high overfitting risk.The selection of feature subsets necessitates determining the most appropriate feature to maximise prediction or classification accuracy.

The primary goal of this research was to identify the best feature subset. Typically, features are chosen based on computational time and the quality of the generated feature subset solutions. In fact, fast and accurate classification with the fewest features is frequently chosen. This appears to be achievable through feature selection. In this paper, MVO algorithm was chosen for determining the best feature subsets for classifying brain tumors as benign or malignant[13][17][18][19].

When creating a predictive model, feature selection is the process of reducing the number of input variables. It is preferable to reduce the number of input variables to reduce the computational cost of modelling and, in some cases, to improve model performance.

Statistical-based feature selection methods entail using statistics to evaluate the relationship between each input variable and the target variable and selecting the input variables with the strongest relationship with the target variable. Although the choice of statistical measures is dependent on the data type of both the input and output variables, these methods can be quick and effective. As a result, when performing filter-based feature selection, it can be difficult for ML practitioner to choose an appropriate statistical measure for a dataset.

Related work:

There are two popularly used feature selection methods: Filter and Wrapper. Filter approaches score and rank features based on specific statistical criteria, then select the features with highest ranking. Examples for this methods- T-test[4], Chi-Square test [5], PCA [6], etc. Despite the speed, the filter model is not strong against interactions between features and redundancies and may not produce the right subset of features.

The wrapper method considers the feature subset and regression model interactions. It uses a learning algorithm to find the most appropriate subset of features and to evaluate the accuracy of potential subsets in predicting the target. These wrapper methods are classified into two- Greedy and Stochastic. Greedy algorithms seek to find the best possible combination of features that results in the best performing model-which is computationally costly and therefore inefficient by using exhaustive quest. To solve large-scale combinatorial problems, stochastic algorithms are created. The leading function subset selection research problems are PSO, ACO, GA, etc. [7]. These methods are capable of capturing redundancy and interaction in features. It is not constrained by the monotonicity assumption. They produce the best function subset, but they are time-consuming to compute. However, the hybridization of filter and wrapper methods done, these are unreliable since, in the presence of component associations, an isolated significant feature can be as discriminatory as an insignificant one.

(3)

Materials and Methods:

The proposed method architecture is shown in figure 1. In this, input MRI image is given to pre-processing stage to eliminate the noise. Median filter is used in this stage. By using Multi-verse Optimization (MVO), threshold was calculated. And this threshold is used to segment the tumour part in input MRI image. Later then the features are extracted using GLCM (Gray-Level Co-Occurrence Matrix) and LBP (Local- Binary- Pattern) methods.

These features are given to the next stage i.e. feature selection. This is done by Multi-verse optimization algorithm. Then the type of tumor was classified by classifier.

Figure 1. Proposed Method Architecture

Feature extraction using GLCM:

It is a well-known texture feature extraction tool that can be used for image segmentation, image retrieval, image classification, pattern recognition, etc. [8] [9]

[10][11].This feature is extracted using GLCM by using the joint condition probability distribution of the image gray-level to reflect texture and calculating the local correlation of pixels to obtain the texture feature value. The authors [12] derived multi-scale texture features from very high-resolution panchromatic imagery by measuring various GLCM directions and window sizes.

It is a 2-D matrix at which two pixels in the image, divided by a certain vector, exist.

Gray-Level Co-occurrence Matrix is represented as below

𝐺𝑗𝑘 △ 𝑤,△ 𝑥 = 𝐾𝑆 𝑗, 𝑘 △ 𝑤,△ 𝑥) (1)

Input MRI image Tumor segmentation

using MVO Pre-processing

Feature Extraction using GLCM and LBP Feature Selection using

MVO

Benign Malignant

Classifier (Random-Forest)

(4)

where 𝐾 = 1

(𝐷−△𝑤)(𝐻−△𝑥) (2)

S 𝑗, 𝑘 △ 𝑤,△ 𝑥) = 𝐻−△𝑥ℎ=1 𝐷−△𝑤𝑑 𝐵 (3)

Where B= 1, if f(d,h)=j and f(d+△w, h+△x) = k

0, otherwise. (4)

It is M * M matrix in size, which means that rows and columns reflect the range of possible pixel values.

This matrix is calculated using two parameters- „dis‟ and „theta‟. „dis‟ is the pixel pair‟s relative distance (measured in pixel numbers 1,2,3, …….). The „theta‟ is the rotational angle or relative orientation (0o, 45o,90o,135o, etc.) as shown in figure 1.

Figure 2.Adjacencydirections (theta)

The statistic of pixel shifts within an area that represents the similarity and regularity of the pixel gray value in different directions is known as image texture directionality.

According to this, the statistic of pixel gray shift in different directions will represent the directional function of texture images [13]. This function is the union of structural and statistical characteristics at the large and small scales. As a result, the path calculation must represent the configuration of the image data to some degree, but it should also reflect the statistical properties of the image pixels. The direction measurement is used to divide the image in various directions and the gray-scale image value is measured in each direction, resulting in the quantitative difference in pixels in each direction.

The texture measure is dependent on second order statistical parameters by using GLCM to define texture characteristics. Table 1 lists thedifferent categories of statistics that were mentioned.

Table I.Features extracted using GLCM matrix

Parameter Description Mathematical formula

ASM- Angular Second Moment

Not have distinguishing

ability 𝒇𝟏 = 𝒈𝟐𝒋𝒌

j k

(5)

Contrast It is rotationally variant and

has discriminating ability 𝒇𝟐 = 𝒓𝟐[ 𝒈𝒋𝒌]

𝑴−𝟏

𝒓=𝟎

j k Correlation It has rotational-dependant

feature and has strong discriminating ability

𝒇𝟑 = (𝒋𝒌)𝒈𝒋 𝒌 𝒋𝒌− 𝝁𝒘𝝁𝒙 𝝈𝒘𝝈𝒙

where 𝜇𝑤 𝑎𝑛𝑑 𝜇𝑥 are means, 𝜎𝑤 𝑎𝑛𝑑 𝜎𝑥 are standard deviations Variance It is rotational-invariant and

has discriminating ability 𝒇𝟒 = (𝒋 − 𝜇)𝟐𝒈𝒋𝒌

𝒌 𝒋

IDM- Inverse Difference Moment

It is like angular second

moment 𝒇𝟓 = 𝒈𝒋𝒌

𝟏 + (𝒋 − 𝒌)𝟐

𝒌 𝒋

Sum average It is similar to variance and

rotationally-variant 𝒇𝟔 = 𝒓𝒈𝒋+𝒌(𝒓)

𝟐𝑴−𝟐

𝒓=𝟎

Sum variance It is similar to variance.

𝒇𝟕 = (𝒓 − 𝒇𝟔)𝟐𝒈𝒋+𝒌(𝒓)

𝟐𝑴−𝟐

𝒓=𝟎

Sum entropy It‟s characteristics similar to

entropy 𝒇𝟖 = − 𝒈𝒘+𝒙(𝒓) 𝒍𝒐𝒈(𝒈𝒘+𝒙(𝒓))

𝟐𝑴−𝟐

𝒓=𝟎

Entropy It is almost rotationally variant and has strong discriminating ability

𝒇𝟗 = − 𝒈𝒋𝒌𝒍𝒐𝒈(𝒈𝒋𝒌)

𝒌 𝒋

IMC-1 It is similar to sum average but is different for different groups and significant rotational variant

𝒇𝟏𝟎 = 𝒇𝟗 − 𝑯𝒘𝒙𝟏 𝒎𝒂𝒙{ 𝑯𝒘, 𝑯𝒙}

where

𝑯𝒘𝒙𝟏= − 𝒈𝒋𝒌𝒍𝒐𝒈 (𝒈𝒘(𝒋)𝒈𝒙(𝒌)

𝒌 𝒋

IMC-2 It is rotation variant and

costly compared to others 𝒇𝟏𝟏 = 𝟏 − 𝒆𝒙𝒑(−𝟐(𝑯𝒘𝒙𝟐− 𝒇𝟗 where

𝑯𝒘𝒙𝟐

= − 𝒈𝒘(𝒋)𝒈𝒙(𝒌)𝒍𝒐𝒈 (𝒈𝒘(𝒋)𝒈𝒙(𝒌)

𝒌 𝒋

Feature Extraction using LBP:

The LBP operator marks the pixels in a decimal image, which is called LBPs or LBP codes that are coded around each pixel in the local structure. [14] The resulting purely negative values are labelled with „0‟, while others are enclosed with „1‟. Each pixel is equivalent to its 8 neighbours in the 3 * 3 neighbourhood by subtracting the centre value of a pixel. A binary is achieved for each pixel, as all these binary values are concatenated in a clockwise direction, starting from one on the left top. For marking the given pixel then the corresponding decimal value of the binary number produced. The resulting binary numbers are LBP Codes.

(6)

One constraint for LBP operators is that they cannot catch dominant features of larger architectures in the smaller 3 * 3 neighbourhood [15]. This operator was later expanded to use neighbourhoods of varying sizes to cope with texture at various scales. A local area is characterised as a collection of sample points uniformly spaced on a circle centred on the pixel being marked, and the sampling points not flowing in the pixels are interpolated using bilinear interpolation such that any radius and number of sample points are permitted in the neighbourhood.

Figure3.Example of LBP operator

For a given pixel at (x,y), the LBP is defined as below

LBP P,R(x,y) = 𝑷−𝟏𝑷=𝟎𝒘 (𝒈𝒍𝒏 − 𝒈𝒍𝒄) 𝟐𝑷 (5) Where 𝒈𝒍𝒄is the gray-level of central pixel.

gl𝑛 is the gray-level of its neighbours.

P is surrounding pixels in the circle neighbourhood.

R is radius of neighbourhood.

The function w(x) is defined as w(x) ={ 1, if x >0

0, ifx < 0 (6)

In the clockwise direction, the sampling is normal around the central pixel according to a certain radius value, which defines the scattered sampling point‟s spatial resolution. If the

“R” grows, an LBP has more general knowledge on the pattern of texture and therefore requires more neighbors. When a big radius is combined with a small P, the re-sampling can lead to extreme artifacts.In this paper, the number of neighbours is set to 8 for the purpose of balancing knowledge and calculation, and the radius is 1.

Multi-Verse Optimization (MVO):

The evolution of the universe was a massive eruption, according to Big Bang Theory.

There was not even one organism in the universe, according to the mentioned hypothesis. The big bang theory is the explanation behind all the world‟s things. A recent and popular

(7)

hypothesis called „multiverse‟ has been suggested by physicians [16].The multiverse hypothesis says that there was not only one big bang, but a lot of big bangs, and that all bore a new world. It has been concluded that there are many other worlds other than our universe.

There is also a chance that these various universes could hit each other.

There are three types of holes exist in the universe- white holes, black holes and worm holes. The „white holes‟ are supposed to be the key factor in the world‟s creation, and they were never observed. The big bang is the white hole itself. when two competing universes merge, a big bang is formed. White holes vary completely from „black holes‟. These hole‟s property is that with its overwhelming gravitational strength it draws every little thing into itself. Due to „worm holes‟, different worlds remain in contact. In multiverse theory, objects are moved from universe to universe or in the same universe with the help of worm holes.The world is expanding in space due to inflation, which is a continuous rate of expansion. In the meantime, the fast inflation rate is required when building various bodies, such as planets and stars. To achieve a peaceful state in various worlds, the black, white and worm holes must be kept in touch.

The following guidelines for optimization should be followed by the MVO algorithm.

 If the rate of inflation is high, there would be a lot of white holes.

 Where the rate of inflation is greater, the existence of black holes would be less possible.

 If the universe‟s inflation rate is high, objects will be sent into white holes.

 If the universe‟s inflation rate is minimal, objects will be obtained from black holes.

 Objects will travel to the best universe through wormholes at random, regardless of inflation rate.

All concepts like white, black, and tunnels can be mathematically modelled by using the Roulette-Wheel mechanism. At each iteration, organise the worlds in order to choose a white hole using the Roulette-Wheel mechanism.

Let

𝑍 =

𝑧11 𝑧12⋯ 𝑧1𝑡

⋮ ⋱ ⋮

𝑧𝑚1 𝑧𝑚2 ⋯ 𝑧𝑚𝑡

(7) where t- number of parameters

m- number of universes.

And 𝑍𝑖𝑗 = 𝑧𝑘

𝑗, 𝑟𝑎𝑛𝑑 1 < = 𝑁𝑂𝑅(𝑍𝐼)

𝑧𝑖𝑗, 𝑟𝑎𝑛𝑑 1 > = 𝑁𝑂𝑅(𝑍𝐼) is the jth term of the ith universe (8)

Where ZI- ith universe

NOR(ZI)- normalized inflation rate.

rand1- random variable is either 1 or 0 𝑧𝑘𝑗- jth term of the kth universe

(8)

Each universe is given an objective function, and the objective value is used to calculate the inflation rate. This algorithm explores the search spaces with black and white holes, while exploiting the search spaces with worm holes.

The process of objects inter-universe trade is that high inflation universes often try to disposal objects and return them with low inflation to the receiving universes. In addition, low-inflation universes use high inflation universes as objects to achieve a stable and inflation-modified world. The inflation rates in all universes are at the end of this mechanism and all universes are stable. The universes are initialised as normal and then ordered according to their inflation during this process. At each point, the exchanging of particles between the local universes through worm hole tunnels is sustained towards the best universe.

This can be represented as shown below

𝑍𝑖𝑗 =

𝑧𝑗 + 𝑇𝐷𝑅 ∗ ((𝑢𝑗 − 𝑙𝑗) ∗ 𝑟𝑎𝑛𝑑4 + 𝑙𝑗 ) 𝑟𝑎𝑛𝑑3 < 0.5 𝑧𝑗 − 𝑇𝐷𝑅 ∗ ((𝑢𝑗 − 𝑙𝑗) ∗ 𝑟𝑎𝑛𝑑4 + 𝑙𝑗 ) 𝑟𝑎𝑛𝑑3 >= 0.5 (𝑟𝑑2) < 𝑊𝐻𝐸𝑃

𝑧𝑖𝑗 (𝑟𝑑2) >= 𝑊𝐻𝐸𝑃

(9)

Where 𝑧𝑗 – jth universe 𝑧𝑖𝑗- jth term in ith universe 𝑢𝑗- upper bound of jth variable

𝑙𝑗 – lower bound of jth variable

WHEP- wormhole existing probability and is defined as 𝑊𝐻𝐸𝑃 = 𝑚𝑖𝑛 − 𝑞 ∗ (𝑚𝑖𝑛 −𝑚𝑎𝑥

𝑄 ) (10)

𝑇𝐷𝑅- Travelling distance rate

𝑇𝐷𝑅 = 1 − (𝑞1/𝑤

𝑄1/𝑤) (11)

rd1, rd2, rand3, and rand4 - random variables in the interval [0 1]

q- current iteration

Q- maximum number of iterations

w- the accuracy of exploitation and is taken as 1.

_________________________________________________________________________________

Pseudo code of MVO algorithm

_________________________________________________________________________________

Initialize population’s maximum size, WHEP, TDR ,Maximum number of iterations;

Initialize uj,lj;

(9)

Initialize BestUniverse;

SU = sorted universe;

NOR= Normalized universe’s inflation rate;

While iter< max_iter do

Find the universe’s cost;

for every universe ‘i’ do for every particle ‘j’ do rd1=rand [0 1];

if rd1 < NOR (Ui) then

white_hole index = Roulette Wheel selection (-NOR);

U(black_hole index, j) = SU(white_hole index, j);

end rd2= rand [0 1];

if rd2 < WHEP then rand3= rand [0 1];

rand4 = rand[0 1];

if rand3 < 0.5 then

𝑍𝑗𝑖= 𝑧𝑗 + 𝑇𝐷𝑅 ∗ ((𝑢𝑗 − 𝑙𝑗) ∗ 𝑟𝑎𝑛𝑑4 + 𝑙𝑗 );

else

𝑍𝑗𝑖 = 𝑧𝑗 − 𝑇𝐷𝑅 ∗ ((𝑢𝑗 − 𝑙𝑗) ∗ 𝑟𝑎𝑛𝑑4 + 𝑙𝑗 );

end end end end end

__________________________________________________________________________________

Detailed steps of Tumor segmentation using MVO algorithm are shown below:

1. Consider input MR image which is to be segmented. (Tumor) 2. Perform pre-processing using median filtering.

3. Initialize MVO parameters like WHEP, TDR, population‟s maximum size, number of iterations, etc.

4. Define cost function i.e. entropy.

5. By optimizing the cost function, the optimal threshold value can be obtained using MVO technique. (see MVO pseudo code above)

6. Best universe describes the Global threshold (Gbest) with the best cost value.

7. Using Gbest, the tumour is segmented.

Detailed steps of feature selection using MVO algorithm are shown below:

1. Consider the segmented MR image as input.

2. Using GLCM, get the texture features.

3. Using LBP, get the texture features.

4. Concatenate the features of step 2 and step 3.

5. Initialize MVO parameters like WHEP, TDR, population‟s maximum size, number of iterations, etc.

6. Define cost function i.e. Accuracy.

(10)

7. By optimizing the cost function, the optimal threshold value can be obtained using the formulae. (see MVO pseudo code above)

8. Best universe describes the Gbest with the best cost value.

9. Apply to Random-forest classifier to know the given input MR image is benign or malignant.

Experimental Results:

We have taken different slices of the human brain and single patient‟s MRI images are analysed. The MRI images are taken from different datasets of the same patient at different times and in different view-focuses. These are taken from Kaggle dataset, using the link htttp://www. kaggle.com/navoneel/brain-mri-images-for-brain-tumor-detection/.

In this, 240 datasets are considered. Among them, 100 datasets are related to malignant and140 are related to benign type of tumor. The extensive simulation has been done to validate the significance of the proposed algorithm. These MRI images are T1- weigthed type, which are widely considered in medical diagnosis of tumor. These experiments are done in MATLAB on PC.

First, the input MRI image is pre-processed using median filter and then using MVO algorithm, a threshold was found. This threshold is applied to segment the tumour from the input MRI image. We run this algorithm for 30 times.

(a) (b) (c)

Figure 4. Input MRI images

(a) (b) (c)

Figure 5. Pre-processed MRI images

(11)

(a) (b) (c)

Figure 6. Tumor segmentation using MVO

The MVO parameters are set as follows:

 Population size=20 (i.e. 20 universes are considered)

 Maximum number of iterations=30

 „w‟ is chosen as <0.05

 WHEP is increased linearly from 0.2 to 1

 TDR is decreased from 0.6 to 0

In figure 4, three input MRI images are considered. These are given to the pre- processed stage. The outputs after median filter and skull stripping, are shown in figure 5.Then by using MVO, a threshold was selected. Threshold values of figure 6 (a), (b), and (c) are 158, 162, 149 respectively.

Later, the features are extracted using GLCM and LBP techniques. A total of 42 features were extracted using the above-mentioned texture-based methods.Then, features are selected using the MVO technique. The performance is calculated using Random-Forest classifier. Total 30 experiments were done. During this, the parameters like execution time, training accuracy, testing accuracy, and number of features selected are observed. Here three optimization techniques are considered- PSO, BAT, and MVO. This is shown in following table.

Table II. Performance Analysis of 3 optimization techniques.

Optimization technique

Experiments Execution time(sec.)

Training Accuracy (%)

Testing Accuracy (%)

Number of selected features

PSO Exp. 1 4.80 88.30 82.41 30

Exp. 30 3.90 87.13 82.41 32

BAT Exp. 1 5.43 88.73 84.71 28

Exp. 30 4.62 87.73 83.50 31

MVO Exp. 1 3.51 88.5 84.71 29

Exp. 30 3.68 87.7 83.67 28

From the above table, it is observed that training and testing accuracies are good for proposed MVO techniques compared to existing PSO and BAT algorithms. And the number

(12)

of features selected are also less. So proposed MVO-RF technique takes less computing time with less features, more accuracy is obtained.

Conclusion:

The construction of an effective classification model is essential for problems of classification with various dimensions and sample sizes. Choosing attributes and choosing the grouping system are the key activities. In this paper, the MVO method is used for feature selection and then fitness values were tested by a Random-Forest classifier. Experimental findings indicate that this MVO+RF approach efficiently streamlined the collection of features and total number of required parameters, which achieves an increased classification accuracy in comparison with standard PSO and BAT algorithms. This approach will provide the optimization of the selection process as an excellent pre-processing technique, since it improves classification accuracy and, at the same time, minimizes the computing time required.

References:

1. Xuan Z, Xin G, Jiajun W, Hui Y, Zhiyong W, Zheru C.Eye tracking data guided feature selection for image classification. Pattern Recognition 2017;63(2):56-70.

2. Bischl B, Lang M, Kotthoff L, Schiffner J, Richter J, Studerus E, Casalicchio G, Jones Z M.mlr: Machine learning in R. Journal of Machine Learning Research 2016;17(170):1-5.

3. Dash M, Choi K, Scheuermann P, Liu H. Feature selection for clustering-a filter solution.In Proceedings of the Second IEEE International Conference on Data Mining, 2002, pp.115-122.

4. Deqing W, Hui Z, Rui L, Weifeng L, Datao W. t-Test feature selection approach based on term frequency for text categorization. Pattern Recognition Letters 2014;45(2):1-10.

5. Jin X, Xu A, Bie R, Guo P. Machine learning techniques and chi-square feature selection for cancer classification using SAGE gene expression profiles. Lecture Notes in Computer Science 2006;3916:106-115.

6. Yijuan Lu, Cohen Ira, Xiang S Z, Tian Qi. Feature selection using principal feature analysis.

In Proceedings of the 15th ACM international conference on Multimedia, 2007, pp. 301-304.

7. Chuan L, Wenyong W, Qiang Z, Xiaoming S, Martin K. A new feature selection method based on a validity index of feature subset. Pattern Recognition Letters 2017;92:1-8.

8. Zhang, Y. Optimisation of building detection in satellite images by combining multispectral classificationand texture filtering. ISPRS J. Photogramm. Remote Sens. 1999, 54, 50–60.

9. Sahoo, M. Biomedical Image Fusion and Segmentation using GLCM. In Proceedings of the 2nd NationalConference-Computing, Communication and Sensor Network (CCSN), Orissa, India, 29–30 October 2011;pp. 34–39.

10. Kekre, H.; Thepade, S.D.; Sarode, T.K.; Suryawanshi, V. Image retrieval using texture features extracted fromGLCM, LBG and KPE. Int. J. Comput. Theory Eng. 2010, 2, 695–700.

11. De Almeida, C.W.D.; De Souza, R.M.C.R.; Candeias, A.L.B. Texture Classification Based on Co-OccurrenceMatrix and Self-Organizing Map. In Proceedings of the IEEE International Conference on Systems, Man andCybernetics, Istanbul, Turkey, 10–13 October 2010; pp.

2487–2491.

12. Pacifici, F.; Chini, M.; Emery, W.J. A neural network approach using multi-scale textural metrics from veryhigh-resolution panchromatic imagery for urban land-use classification.

Remote Sens. Environ. 2009, 113,1276–1292.

13. Pravin Kshirsagar and Sudhir Akojwar (2017), “Classification of ECG-signals using Artificial Neural Networks”, Researchgate.net

14. S. Liao, M. W. K. Law and A. C. S. Chung, "Dominant Local Binary Patterns for Texture Classification," in IEEE Transactions on Image Processing, vol. 18, no. 5, pp. 1107-1118, May 2009.

(13)

15. Shihab Hamad Khaleefah1, Salama A. Mostafa2, Aida Mustapha3, Mohammad Faidzul Nasrudin, “Review of localbinary pattern operators in image feature extraction” in Indonesian Journal of Electrical Engineering and Computer Science Vol. 19, No. 1, July 2020, pp. 23~31

16. Mirjalili, S.; Mirjalili, S.M.; Hatamlou, A. Multi-Verse Optimizer: A nature-inspired algorithm for globaloptimization. Neural Comput. Appl. 2016, 27, 495–513.

17. Pravin R Kshirsagar, Anil N Rakhonde, Pranav Chippalkatti, “ MRI IMAGE BASED BRAIN TUMOR DETECTION USING MACHINE LEARNING”, Test Engineering and Management, January- February 2020 ISSN: 0193-4120, Vol. 81, Page No. 3672 – 3680.

18. Pravin Kshirsagar et.al (2016), “Brain Tumor classification and Detection using Neural Network”, DOI: 10.13140/RG.2.2.26169.72805.

19. Pravin R. Kshirsagar, Arpit D. Yadav, Kirti A. Joshi, Pranav Chippalkatti, Rinali Y. Nerkar (2020) “Classification and Detection of Brain Tumor by using GLCM Texture Feature and ANFIS” Journal of Research in Image and Signal Processing, 5(1), 15- 31 http://doi.org/

10.5281/zenodo.3732939.

Referințe

DOCUMENTE SIMILARE

The objective of the present paper is to study the effect of the preparation condition of MCM-48 samples to be used as catalysts for reducing the toxicity of the

Results obtained reveal the reliability and good predicatively of neural network model for the prediction of the size of Ag-NPs in green method.. (Received May 22, 2013;

Our main objective in this paper is to study the sufficient conditions of the form (1.3) for meromorphically convex functions of order α and for functions in a family that are convex

The author focused the research on the various optimization models adequate for the (MOF)JSSP, both conventional and non-conventional: Petri nets, waiting systems, general

• Exploitation: pull as much as possible the Pareto optimal arms of relevant scalarization functions.. Multi-armed bandits for

It is used to detect the tumor of brain[18].They proposed a Statistical Evaluation of Brain Tumor in Different MRI Protocols using support vector machine and classifies

Dataset used for training Unsupervised machine learning technique is neither classified nor labeled and it this methodology learn how a model can recognize a function

had suggested machine learning based automatic segmentation and hybrid feature analysis for Diabetic Retinopathy classification using fundus Image (2020).. This study used