ADVERTISEMENT

Scientists Are Using Machine Learning To Better Predict Epilepsy

There are 2 aspects of this research that are worth highlighting: (1) we showed that micro-structural extra-hippocampal abnormalities are consistent enough across medial temporal lobe epilepsy (TLE) patients that they can be used to predict TLE, and (2) we obtained regularization values for the models trained on this sparse data in an unusual but effective manner.

Our input data consisted of 3 different diffusion imaging modalities: mean diffusivity (MD), fractional anisotropy (FA), and mean kurtosis (MK). Predictive models trained with MK proved to be the most accurate: .82 vs .68 (FA) and .51 (MD). Also, the highest coefficients of these linear models were located within the inferior medial aspect of the temporal lobes. These locations have complex fiber anatomy with many crossings.

ADVERTISEMENT

Diffusion kurtosis imaging (DKI) is more apt than diffusion tensor imaging (DTI) at capturing fiber crossings due to the presence of non-Gaussian water diffusion. This likely explains why MK, which is calculated from DKI, had higher accuracy than the other 2 modalities, which are calculated from DTI. Finally, if these micro-structural abnormalities can be further identified by future work, then it may be possible to identify the different TLE phenotypes. This is of clinical importance since treatment varies by phenotype.

Linear support vector machine (SVM) models are often used to predict an outcome from input data. In this case, we predict TLE for each subject from the diffusion image (MK, MD, or FA). One of the challenges of using SVM is to find the optimal regularization parameter to prevent overfitting.

One option is to sub-divide the train set into a sub-train and sub-test set on each run of the cross-validation, find the regularization value which performs best on the sub-test with an algorithm such as grid-search and use this best regularization value to train the model on the train set and predict on the test set. However, this is computationally intensive, and further dividing the train set leads to increased variability. The regularization values found among the different folds of the cross-validation are usually different by orders of magnitudes. This is especially true for many life-science problems with a small number of subjects and high dimensionality.

Another option is to assume a distribution of values for the regularization parameter, instead of finding a single value. A model is fitted with the train set for each regularization value in the distribution. The final prediction is calculated as a weighted average of the different modelsā€™ predictions on the test set.

ADVERTISEMENT

The weight (importance) assigned to each model in the prediction average is determined by 2 competing factors: the weight is (1) proportional to how well the associated model fits the train data, and (2) inversely proportional to the distance between the regularization value and the minimum required regularization value (CĪ¼) to have no classification error on the train data. The larger the regularization value, the better the model fits the data, but the greater the distance from CĪ¼.Ā The first factor ensures that we give more weight to models that at least fit the data, the second factor ensures that we donā€™t give too high a weight to models that overfit the data.

This method showed superior performance than using a single regularization value. Averaging over the predictions removed noise. It also helped reduce overfitting because overfit models tend to create very complex, or wiggly, boundaries between the classes in the feature space. By averaging over the models, a new boundary is created that is smoother and simpler. Also worth noting, even though it is not in the paper, is that this method calculated relatively consistent regularization values. The values were within a factor of 2 to 3 between cross-validation runs.

This study,Ā Using machine learning to classify temporal lobe epilepsy based on diffusion MRI was recently published by John Del Gaizo and Leonardo Bonilha in the journalĀ Brain and Behavior.

Comments

READ THIS NEXT

Investigating The Effects Of PM2.5 On Energy Metabolic Reprogramming

During the last two decades, air pollution has been regarded as one of the toughest challenges facing human beings. As […]

Researchers Use Gene Editing To Remove Fatal Blood Disorder From Embryo

It would be an understatement to say that there are many diseases that humans are facing every day. Some of […]

An Analysis Of The Correlation Between Population And Crime Rates

Social scientists have tried to explain geographic concentrations of crime for more than a century. Two macro-sociological explanations are often […]

The Ingredient Of Chili Peppers Changes The Gene Expression Pattern Of Cells

Capsaicin (trans-8-methyl-N-vanillyl-6-nonenamide) is the compound in chili peppers (Capsicum annuum) responsible for their “hot” taste. Capsaicin binds as a ligand […]

Fighting With Your Own Reflection: Can Animals Perceive The Differences Between Natural And Fake Opponents?

While standing in front of a mirror, it is easy for humans to recognize that the image in the mirror […]

Wireless Sweat Monitoring Could Make The Self-Tracking Of Chronic Diseases Easier

Traditional checks for chronic diseases, such as diabetes, often involve analysis via blood samples and require trained medical staff. However, […]

Developments In Piezoelectric Nanogenerators

Acceleration sensors play a crucial and essential role in many fields, such as large mechanical structure testing, national defense, transportation, […]

Science Trends is a popular source of science news and education around the world. We cover everything from solar power cell technology to climate change to cancer research. We help hundreds of thousands of people every month learn about the world we live in and the latest scientific breakthroughs. Want to know more?