US10304006B2 - Method for integrating and fusing heterogeneous data types to perform predictive analysis - Google Patents
Method for integrating and fusing heterogeneous data types to perform predictive analysis Download PDFInfo
- Publication number
- US10304006B2 US10304006B2 US14/183,018 US201414183018A US10304006B2 US 10304006 B2 US10304006 B2 US 10304006B2 US 201414183018 A US201414183018 A US 201414183018A US 10304006 B2 US10304006 B2 US 10304006B2
- Authority
- US
- United States
- Prior art keywords
- data
- kernel
- training
- function
- disease
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
- 238000000034 method Methods 0.000 title claims abstract description 53
- 238000004458 analytical method Methods 0.000 title claims description 17
- 238000012549 training Methods 0.000 claims abstract description 82
- 208000037265 diseases, disorders, signs and symptoms Diseases 0.000 claims abstract description 46
- 201000010099 disease Diseases 0.000 claims abstract description 45
- 238000004364 calculation method Methods 0.000 claims abstract description 13
- 230000006870 function Effects 0.000 claims description 105
- 208000024827 Alzheimer disease Diseases 0.000 claims description 29
- 208000010877 cognitive disease Diseases 0.000 claims description 20
- 230000001149 cognitive effect Effects 0.000 claims description 20
- 208000027061 mild cognitive impairment Diseases 0.000 claims description 17
- 238000012360 testing method Methods 0.000 claims description 16
- 238000012706 support-vector machine Methods 0.000 claims description 11
- 230000004190 glucose uptake Effects 0.000 claims description 10
- 230000002123 temporal effect Effects 0.000 claims description 10
- 210000004556 brain Anatomy 0.000 claims description 7
- 230000002146 bilateral effect Effects 0.000 claims description 4
- 230000006996 mental state Effects 0.000 claims description 4
- 239000000523 sample Substances 0.000 description 26
- 238000004422 calculation algorithm Methods 0.000 description 24
- 239000013598 vector Substances 0.000 description 20
- 238000013459 approach Methods 0.000 description 15
- 238000007635 classification algorithm Methods 0.000 description 14
- 230000008569 process Effects 0.000 description 10
- 102100034112 Alkyldihydroxyacetonephosphate synthase, peroxisomal Human genes 0.000 description 8
- 101000799143 Homo sapiens Alkyldihydroxyacetonephosphate synthase, peroxisomal Proteins 0.000 description 8
- 238000000848 angular dependent Auger electron spectroscopy Methods 0.000 description 8
- 239000002131 composite material Substances 0.000 description 8
- 230000004153 glucose metabolism Effects 0.000 description 8
- 238000010586 diagram Methods 0.000 description 7
- 238000005457 optimization Methods 0.000 description 7
- 238000006243 chemical reaction Methods 0.000 description 6
- 210000001519 tissue Anatomy 0.000 description 6
- 230000004927 fusion Effects 0.000 description 5
- 238000013507 mapping Methods 0.000 description 5
- 230000007246 mechanism Effects 0.000 description 5
- 238000003860 storage Methods 0.000 description 5
- 238000003745 diagnosis Methods 0.000 description 4
- 230000008901 benefit Effects 0.000 description 3
- 230000001747 exhibiting effect Effects 0.000 description 3
- 230000002068 genetic effect Effects 0.000 description 3
- 238000000926 separation method Methods 0.000 description 3
- 208000007333 Brain Concussion Diseases 0.000 description 2
- 208000004051 Chronic Traumatic Encephalopathy Diseases 0.000 description 2
- 208000028698 Cognitive impairment Diseases 0.000 description 2
- 206010019196 Head injury Diseases 0.000 description 2
- 208000030886 Traumatic Brain injury Diseases 0.000 description 2
- 238000009825 accumulation Methods 0.000 description 2
- 230000004075 alteration Effects 0.000 description 2
- 230000015556 catabolic process Effects 0.000 description 2
- 210000001175 cerebrospinal fluid Anatomy 0.000 description 2
- HVYWMOMLDIMFJA-DPAQBDIFSA-N cholesterol Chemical compound C1C=C2C[C@@H](O)CC[C@]2(C)[C@@H]2[C@@H]1[C@@H]1CC[C@H]([C@H](C)CCCC(C)C)[C@@]1(C)CC2 HVYWMOMLDIMFJA-DPAQBDIFSA-N 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 230000000295 complement effect Effects 0.000 description 2
- 238000010276 construction Methods 0.000 description 2
- 208000017004 dementia pugilistica Diseases 0.000 description 2
- 238000002405 diagnostic procedure Methods 0.000 description 2
- 230000014509 gene expression Effects 0.000 description 2
- 230000006872 improvement Effects 0.000 description 2
- 238000005259 measurement Methods 0.000 description 2
- 230000003340 mental effect Effects 0.000 description 2
- 239000000203 mixture Substances 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000002610 neuroimaging Methods 0.000 description 2
- 210000002569 neuron Anatomy 0.000 description 2
- 239000013610 patient sample Substances 0.000 description 2
- 208000028173 post-traumatic stress disease Diseases 0.000 description 2
- 230000009529 traumatic brain injury Effects 0.000 description 2
- 102100027831 14-3-3 protein theta Human genes 0.000 description 1
- 101150037123 APOE gene Proteins 0.000 description 1
- 102000013455 Amyloid beta-Peptides Human genes 0.000 description 1
- 108010090849 Amyloid beta-Peptides Proteins 0.000 description 1
- 102000007592 Apolipoproteins Human genes 0.000 description 1
- 108010071619 Apolipoproteins Proteins 0.000 description 1
- 208000024172 Cardiovascular disease Diseases 0.000 description 1
- 206010010254 Concussion Diseases 0.000 description 1
- WQZGKKKJIJFFOK-GASJEMHNSA-N Glucose Natural products OC[C@H]1OC(O)[C@H](O)[C@@H](O)[C@@H]1O WQZGKKKJIJFFOK-GASJEMHNSA-N 0.000 description 1
- 238000005481 NMR spectroscopy Methods 0.000 description 1
- 201000007201 aphasia Diseases 0.000 description 1
- 230000003542 behavioural effect Effects 0.000 description 1
- 238000004166 bioassay Methods 0.000 description 1
- 239000000090 biomarker Substances 0.000 description 1
- 235000012000 cholesterol Nutrition 0.000 description 1
- 230000006999 cognitive decline Effects 0.000 description 1
- 230000009514 concussion Effects 0.000 description 1
- 238000002790 cross-validation Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000012774 diagnostic algorithm Methods 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 238000009472 formulation Methods 0.000 description 1
- 239000008103 glucose Substances 0.000 description 1
- 235000009200 high fat diet Nutrition 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 238000002955 isolation Methods 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 239000003550 marker Substances 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 239000002207 metabolite Substances 0.000 description 1
- 238000000491 multivariate analysis Methods 0.000 description 1
- 210000000653 nervous system Anatomy 0.000 description 1
- 238000002600 positron emission tomography Methods 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 108090000623 proteins and genes Proteins 0.000 description 1
- 102000004169 proteins and genes Human genes 0.000 description 1
- 230000008430 psychophysiology Effects 0.000 description 1
- 238000011084 recovery Methods 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 208000024891 symptom Diseases 0.000 description 1
- 210000000225 synapse Anatomy 0.000 description 1
- 230000000946 synaptic effect Effects 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
-
- G06F19/00—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
- G06N20/10—Machine learning using kernel methods, e.g. support vector machines [SVM]
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/20—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
Definitions
- the technical field relates generally to classification and prediction algorithms for use with a dataset composed of information from heterogeneous data types.
- Kernel methods are a class of algorithms used for pattern analysis. Kernel methods can be used to map input data into a high dimensional or infinite dimensional inner-product, or Hilbert feature space, where each dimension corresponds to a feature. In the feature space, a variety of methods may be used to find relationships within the input data.
- the system can comprise a memory storing a training data set including a first plurality of training samples of a first data type and a corresponding second plurality of training samples of a second data type, the first data type including discrete data and the second data type including continuous data, one or more processors coupled to the memory, a disease prediction component executable by the one or more processors and configured to: receive patient data including a first input sample of the first data type and a second input sample of the second data type, provide the first input sample and the first plurality of training samples to a first kernel function of a multiple kernel decision function, provide the second input sample and the second plurality of training samples to a second kernel function of the multiple kernel decision function, perform at least one calculation using the multiple kernel decision function to produce at least one result, and determine a probability of whether the patient data indicates that the patient will develop the disease based on the at least one result of the multiple kernel decision function.
- the disease prediction component is further configured to perform the at least one calculation using the multiple kernel function at least in part by calculating at least one of a first set of weighting factors and a second set of weighting factors, the first set of weighting factors including a weight for each training sample in the training data set, the second set of weighting factors including a weight for each data type.
- the disease prediction system is for predicting the onset of Alzheimer's disease in individuals who express mild cognitive impairment.
- One or more further aspects of the present disclosure involve embodiments directed to a method for predicting the onset of a disease.
- the method can comprise receiving patient data including a first input sample of a first data type and a second input sample of a second data type, the first data type including discrete data and the second data type including continuous data, receiving a training data set including a first plurality of training samples of the first data type and a corresponding second plurality of training samples of the second data type, providing the first input sample and the first plurality of training samples to a first kernel function of a multiple kernel decision function, providing the second input sample and the second plurality of training samples to a second kernel function of the multiple kernel decision function, performing at least one calculation using the multiple kernel decision function to produce at least one result, and determining a probability of whether the patient data indicates that the patient will develop the disease based on the at least one result of the multiple kernel decision function.
- performing the at least one calculation using the multiple kernel function includes calculating at least one of a first set of weighting factors and a second set of weighting factors, the first set of weighting factors including a weight for each training sample in the training data set, the second set of weighting factors including a weight for each data type.
- One or more further aspects of the present disclosure are directed to a method for predicting the probability of an event comprising receiving a plurality of input samples, each input sample of the plurality of input samples having a data type of a plurality of data types, the data types including quantitative data, discrete data, continuous data, and categorical data, receiving a training data set including a plurality of training samples, each of the plurality of training samples having a data type of the plurality of data types, providing each input sample and each respective training sample with the same data type of the plurality of training samples to a corresponding kernel function of a multiple kernel decision function having a plurality of kernel functions, performing at least one calculation using the multiple kernel decision function to produce at least one result, and determining a probability of an event based on the at least one result of the multiple kernel decision function.
- At least one of the first kernel function and the second kernel function is a radial basis function.
- the multiple kernel decision function is consistent with a Kernel Fisher Discriminant Analysis classifier.
- the multiple kernel decision function is consistent with a Support Vector Machine classifier.
- the discrete data includes results from at least one cognitive assessment exam and the continuous data includes results from at least one rate of glucose uptake by tissue.
- the at least one cognitive assessment exam is selected from the list consisting of: an Alzheimer's Disease Assessment Score, a Mini Mental State Exam, a Boston Naming Test, and an American National Adult Reading Test.
- the at least one rate of glucose uptake by tissue is retrieved from one or more voxels of the brain selected from the list consisting of: left angular gyrus, right angular gyrus, left temporal, right temporal, and bilateral cingulum posterior.
- FIG. 1 is a diagram of a single kernel classification algorithm in accordance with one example
- FIG. 2 is a diagram of a multiple kernel classification algorithm in accordance with another example
- FIG. 3 is a graph illustrating prediction performance using several classification algorithms in accordance with one example
- FIG. 4 is a second graph prediction performance using classification algorithms in accordance with another example
- FIG. 5 is a pair of graphs illustrating at least one benefit of using a classification algorithm in accordance with one example
- FIG. 6 is a flow diagram of one example method of disease prediction.
- FIG. 7 is a block diagram of one example of a computer system upon which various aspects of the present embodiments may be implemented.
- the method may include receiving a plurality of input samples, where each input sample corresponds to a data type.
- the data types may include quantitative data, discrete data, continuous data, and categorical data.
- the method may further include receiving a training data set that includes a plurality of training samples, where each of the plurality of training samples corresponds to a data type.
- Each input sample and each respective training sample with the same data type of the plurality of training samples may correspond to a kernel function of a multiple kernel decision function having a plurality of kernel functions.
- At least one calculation may be performed using the multiple kernel decision function to produce at least one result, and the probability of the event may be determined based on the at least one result.
- references to “or” may be construed as inclusive so that any terms described using “or” may indicate any of a single, more than one, and all of the described terms.
- the term usage in the incorporated reference is supplementary to that of this document; for irreconcilable inconsistencies, the term usage in this document controls.
- titles or subtitles may be used in the specification for the convenience of a reader, which shall have no influence on the scope of the present invention.
- classification and prediction algorithms may be optimized for use with one or more datasets comprising information from a wide range of disparate sources.
- Such heterogeneous data sets may contain a mixture of different types of quantitative, discrete, continuous, and categorical data, and may further include one or more signals collected from multiple sensors sampled at different time scales.
- the predictive and diagnostic algorithms built from these kinds of datasets may be referred to as heterogeneous data fusion algorithms. These algorithms are capable of being optimized to exploit both the unique and complementary information each type of data provides.
- a type of heterogeneous data fusion algorithm includes a Multiple Kernel Learning (MKL) classification algorithm, also referred to as a multiple kernel decision function, for purposes of predicting and diagnosing.
- MKL Multiple Kernel Learning
- the MKL algorithm may be used to predict or determine the probability of an event based on training data and sample data constructed from sets of heterogeneous data types of data.
- the MKL algorithm may be used for diagnostic purposes, such as for diagnostic patients suffering from one or more symptoms or possessing certain characteristics.
- the methodology is an expansion of single-kernel machine learning classification algorithms that may be used to develop “class” or “group” discriminating functions.
- a classification algorithm Given a set of data points describing the characteristics of several distinct groups, a classification algorithm will learn a mathematical “rule” for separating the groups.
- a kernel-based classifier first transforms the data from raw data space into a higher dimensional space, where it may be more linearly separable. Through optimization, a decision function may be identified that separates the classes or groups in the new “feature” space. New unlabeled data points may then be classified based on where they are positioned relative to the hyperplane.
- the MKL approach creates a method for creating and maximizing diagnostic and predictive power from sets of heterogeneous data.
- the underlying premise is that using one or more combinations of different data types allows for greater diagnostic and predictive capability than by considering each data type in isolation.
- This type of analysis is useful in problems where many different types of data are involved.
- the different types of data may include quantitative data, discrete data, continuous data, and categorical data.
- discrete data may result from either a finite or a countable infinity of possible options for the values present in a given discrete data set. Therefore, discrete data may represent data that can only take certain values, and may be a count of something. Further, discrete data may include numerical, ordinal, or qualitative (nominal) kinds of data.
- continuous data may be measured, and may be capable of taking any numerical value.
- the numerical values may be subdivided into finer increments, depending upon the desired precision.
- Quantitative data may refer to measurable data, and may include discrete or continuous types of data.
- Categorical data may result from placing individual values into groups or categories. Categorical data may be derived from quantitative or qualitative types of data.
- An example of an area where the MKL approach may be useful is the medical field, where certain types of diseases or conditions may be the result of very diverse and different types of inputs, including continuous data, such as physical data that may be collected from tests conducted on a patient's body, and discrete data, such as cognitive data that may be collected from tests completed by the patient using their mental capacity. Other types of data may also be included that relates to the patient, such as demographic data, and data related to medical, psychiatric, and family history.
- MKL approach may be useful in the field of engineering, where predictive algorithms may be useful in determining locations of natural resources, failure points in a structure, weather patterns, and natural disasters such as earthquakes, tornadoes, and tsunamis.
- Rescue recovery operations, containment exercises (e.g., oil spills), process manufacturing, construction methods, and election prediction are also areas that may benefit from the MKL approach to one or more problems of interest.
- An especially useful application for the MKL approach is for predicting whether a population with certain characteristics will develop a certain type of condition and can be diagnosed with a certain type of condition.
- individuals with a history of multiple concussions and other forms of head injury may develop Chronic Traumatic Encephalopathy (CTE).
- soldiers returning from war or combat conditions may suffer from at least one of Post-Traumatic Stress Disorder (PTSD) and Mild Traumatic Brain Injury (MTBI).
- PTSD Post-Traumatic Stress Disorder
- MTBI Mild Traumatic Brain Injury
- the MKL approach is capable of mathematically “homogenizing” one or more different types of data, weighting them, and then combining their weighted contributions to effectively increase the overall separation of classes. This allows for greater accuracy in predicting and diagnosing these disorders.
- the functionality is tied to the ability to optimize combinations of different types of data.
- the MKL algorithm may be applied to the problem of predicting whether an individual exhibiting or expressing Mild Cognitive Impairment (MCI) will progress to be diagnosed with Alzheimer's Disease (AD).
- MCI Mild Cognitive Impairment
- one or more subsets of data may be used to successfully predict conversion of MCI to AD using an MKL classification algorithm.
- the MKL algorithm may use the combination of discrete data, such as the results from at least one cognitive assessment exam, with continuous data, such as the rate of glucose uptake by tissue (e.g., FDG-PET) to provide a probability that a patient with MCI will develop AD.
- AD Alzheimer's disease
- data sources may include discrete, continuous, quantitative, and categorical data types.
- results from neuroimaging, psychophysiology, biological assays, psychosocial history, medical history, and gene expression may all be used to form a complete diagnosis.
- Useful data may include results collected from brain metabolite concentrations derived from Nuclear Magnetic Resonance (NMR) spectroscopy, MRI images, fluorodeoxyglucose-positron emission tomography (FDG-PET) glucose levels, cerebrospinal fluid (CSF) biomarkers, cognitive assessment(s), genetic profile information, and demographic assessments.
- NMR Nuclear Magnetic Resonance
- FDG-PET fluorodeoxyglucose-positron emission tomography
- CSF cerebrospinal fluid
- cognitive assessment(s) genetic profile information
- demographic assessments may convey different diagnostic information, all the data may be related or linked to a certain condition, such as the onset of a disease. These heterogeneous sets of data may be integrated to form cross
- the heterogeneous sets of data may be used to predict the onset of AD.
- the heterogeneous sets of data may be used to diagnose AD.
- the actual diagnosis of AD may be made by an individual's Primary Care Physician (PCP) using one or more of the previously mentioned types of data.
- PCP Primary Care Physician
- a diagnosis may include medical and family history, psychiatric history, the results from cognitive tests such as the Alzheimer's Disease Assessment Score, physical and neurologic exams, a history of cognitive and behavioral changes, and results from MRI testing that identifies brain changes and may be capable of eliminating other causes of cognitive decline.
- the causes of AD are believed to come from many sources, including: the accumulation of the protein beta-amyloid outside the nerve cells, the accumulation of the protein tau inside neurons, family history, and within the nervous system, the inability or failure of information to transfer at the synapse.
- One or more risk factors may be associated with AD. These may include family history, which may be the result of genetic and/or environmental factors, the presence or level of Apolipoprotein E- ⁇ 4, which is one of three forms of the APOE gene and provides the blueprint for a protein that carries cholesterol in the bloodstream, Mild Cognitive Impairment (MCI), cardiovascular disease, physical/mental inactivity, high fat diets, and head trauma and/or Traumatic Brain Injury (TBI).
- MCI Mild Cognitive Impairment
- MCI Mild Cognitive Impairment
- cardiovascular disease the physical/mental inactivity
- high fat diets high fat diets
- TBI Traumatic Brain Injury
- an MKL approach may be used to determine a probability of whether an individual exhibiting one or more of these risk factors, such as MCI, will develop AD.
- an MKL approach may be used to diagnose an individual exhibiting one or more of these risk factors and/or other diagnostic data to determine if they actually have AD.
- Kernels are functions that provide a mapping between different vector spaces.
- the key to the kernel-based classification algorithms is their ability to represent the data in a higher dimensional space.
- kernels may perform an implicit dot product between two vectors in the higher dimensional space without actually having to go into that space. Datasets with non-linear class boundaries in raw data space may become linearly separable when they have been transformed appropriately.
- explicitly mapping the data to the feature space may be computationally intensive. Instead, a kernel function may be employed to perform an implicit mapping of the data to the feature space.
- the kernel function is an operation that can be applied to the data vectors that is equivalent to computing their inner product after they have been embedded in the feature space without actually having to map them first.
- Kernel evaluations can therefore be used in place of every inner product operation required by optimization algorithms designed to identify class boundaries in feature space.
- a different kernel may be chosen for each modality of data.
- Commonly used kernel functions include those represented below by Equation 2 and Equation 3.
- Equation 2 describes a linear kernel function which computes the dot product between two data vectors.
- Equation 3 represents the Gaussian Radial Basis Function, which performs a non-linear mapping of the data into a higher dimensional feature space. The user may choose a at their own discretion, or may perform an optimization to find the best value.
- Linear: K ( ⁇ right arrow over (x) ⁇ , ⁇ right arrow over (y) ⁇ ) ⁇ right arrow over (x) ⁇ T ⁇ right arrow over (y) ⁇ (2)
- Gaussian Radial Basis Function
- Kernel-based classification algorithms differ primarily in their criterion used to optimize the separation of classes in feature space.
- One such method is the Support Vector Machine (SVM) classifier, which identifies a linear hyperplane that maximizes the margin between the classes in kernel space.
- Another method is the Kernel Fisher Discriminant Analysis (KFDA), which approaches the optimization by finding a boundary that maximizes the ratio of between-class scatter and within-class scatter in kernel space. Both of these methods scale well to the multiple-kernel case.
- the value of the output of the decision function corresponds to its class membership, based on a threshold value optimized during training.
- the optimization criteria for the separation of classes using KFDA is the maximization of the ratio of the between-class scatter and within-class scatter, where “scatter” is analogous to multi-dimensional “variance.”
- Assessment of the ratio of the between-class scatter and within-class scatter requires computation of the means and covariances of the class data in feature space.
- the sample means of the positive and negative classes are computed by the equations 5 and 6, respectively, as shown below:
- the scatter-ratio function is defined in terms of these means (i.e., equations 5 and 6) and covariances (i.e., equations 7 and 8).
- the scatter-ratio maximization equation is shown below in equation 9 in terms of w and the kernel transform K:
- ⁇ is a small positive regularization parameter
- I is the identity matrix.
- the class-separability optimization function may be reformulated in terms of kernels, where all inner product computations are replaced with their corresponding kernel transforms. The result is shown in Equation 10 below, where w is redefined as:
- Equation 11 the decision function may be formulated in terms of kernel computations, as shown below in Equation 11:
- FIG. 1 is a diagram of a single kernel classification algorithm 100 in accordance with one example that classifies an input unclassified data vector 102 .
- a feature mapping relationship 104 between the input unclassified data vector 102 is implicitly created by representing the input unclassified data vector relative to each training sample of a set of training samples in kernel space.
- the values of each comparison in kernel space from the implicit representation 106 are weighted and summed consistent with the set of weights 108 in the decision function 110 .
- the output of the decision function may be compared to a pre-determined threshold to identify the class membership of the input unclassified data vector 102 .
- N is the number of kernels used to create the composite kernel K c
- the N-dimensional vector ⁇ right arrow over ( ⁇ ) ⁇ contains the weights on each sub-kernel, determined through the optimization routine.
- Equation 12 a vector of training sample weights, ⁇ right arrow over ( ⁇ ) ⁇ , is also specified.
- the expressions shown in Equation 12 may also include a scalar offset.
- FIG. 2 is a diagram of a multiple kernel classification algorithm 200 that illustrates how this process may be implemented for a dataset composed of two heterogeneous data types.
- Each data type is transformed into its own kernel space relative to the corresponding data in each training sample as illustrated by implicit representation 208 .
- Composite kernels 212 are formed based on a weighted combination consistent with the data type weights 210 . For a given training sample ⁇ right arrow over (x) ⁇ i , the composite kernel
- An MKL-KFDA algorithm was applied to the problem of predicting whether an individual having Mild Cognitive Impairment (MCI) would progress to having Alzheimer's Disease (AD).
- the dataset utilized in this test case was collected as part of the Alzheimer's Disease Neuroimaging Initiative (ADNI) project, and was constructed from a diverse array of data collected from hundreds of participants, including demographic data, cognitive exam scores, MRI images and voxel volumes, FDG-PET glucose metabolism levels, and genotypes.
- Table 2 describes the data used in the analysis.
- the data set combines both discrete (cognitive exams scores) and continuous (FDG-PET) variables, collected from different sources.
- Four cognitive exam scores (explained further below) and FDG-PET glucose metabolism levels (which is linked to synaptic activity) from five brain voxels were used to create a nine-dimensional feature vector.
- each set of features was represented by a separate kernel. Therefore, the cognitive exam scores represented one kernel and the FDG-PET results represented a second kernel.
- ADAS Alzheimer's Disease Assessment Score evaluates cognitive impairment in the assessment of Alzheimer's disease MMSE Mini Mental State Exam - samples functions Score including arithmetic, memory, and orientation; tests for cognitive impairment BNT Boston Naming Test - patients with anomic Score aphasia often have greater difficulties with the naming both difficult and low frequency objects, as well as easy and high frequency objects ANART American National Adult Reading Test - Score estimates levels of premorbid intelligence 2
- FDG-PET Glucose metabolism, left angular gyrus voxel Left Angular Gyrus
- FDG-PET Glucose metabolism, right angular gyrus voxel Right Angular Gyrus
- FDG-PET Glucose metabolism, left temporal voxel Left Temporal FDG-PET: Glucose metabolism, right temporal voxel Right Temporal FDG-PET: Glucose metabolism
- LDA Linear Discriminant Analysis
- FIG. 3 shows a graphical comparison of MKL-KFDA against LDA, single kernel SVM, and single kernel KFDA algorithms. All nine features were concatenated into a single vector to train the algorithm. As illustrated in FIG. 3 , the MKL-KFDA outperforms the other three types of classifiers.
- each feature type was used to train each of the LDA, single-kernel SVM, and single-kernel KFDA algorithms. These algorithms were then compared against the MKL-KFDA algorithm. The results of the comparison are shown graphically in FIG. 4 .
- Each feature type offers unique information to the prediction problem, and the MKL method is capable of increasing performance results by using the combination of these features. In other words, the method by which the features are fused determines the strength of their combined efficacy.
- bPCC balanced percent correct classifications
- the “% Agree” refers to the proportion of subjects who were given identical diagnoses by all of the univariate predictor variables.
- the “% Disagree” refers to the proportion for which the individual predictor variable generated conflicting diagnoses.
- the ADAS and BNT univariate classification algorithms agreed on the subjects' diagnoses (conversion/non-conversion to AD) 62% of the time.
- the results indicate that the pairwise and three-way agreement between the variables is low, which means that similar univariate bPCC values do not necessarily signify similar information.
- the methods and systems disclosed herein may be used to predict the probability that a patient may contract a disease.
- the methods and systems disclosed herein may be used to predict the probability that a patient will develop Alzheimer's disease.
- the disease prediction processes may be executed by a microprocessor-based computer system, such as the computer system 700 described below with reference to FIG. 7 .
- FIG. 6 illustrates one example disease prediction test process 600 that may be performed by a computer system (e.g., executed by processor 706 of computer system 700 ).
- the disease prediction process 600 includes the acts of receiving patient data 602 , optionally receiving training data 604 , providing first data to a first kernel function 606 , providing second data to a second kernel function 608 , computing a decision function result 610 , and determining a probability 612 .
- the system receives data from a patient.
- the patient data may include one or more patient samples including one or more data types.
- the patient samples may include discrete data and continuous data.
- the discrete data may include results from at least one cognitive assessment exam and the continuous data may include at least one rate of glucose uptake by tissue.
- the system receives a training data set.
- the training data set may include a first plurality of training samples of a first data type and a corresponding second plurality of training samples of a second data type.
- the first data type includes discrete data and the second data type includes continuous data.
- the discrete data includes results from at least one cognitive assessment exam and the continuous data includes results from at least one rate of glucose uptake by tissue.
- this example includes first and second data types, multiple data types are within the scope of the systems and methods disclosed herein.
- other types of data may be included in the algorithm, such as results from an MRI, demographic assessment data, and genetic profile data.
- the algorithm may be used without the inclusion of the cognitive assessment exam data and/or the at least one rate of glucose uptake by tissue.
- optional act 604 of receiving the training data set is only performed once and the received training data is stored in memory (e.g., memory device 710 of computer system 700 ). In this embodiment, option act 604 may not be repeated for subsequent iterations of the disease prediction process 600 .
- the system respectively provides a first data set to a first kernel function and a second data set to a second kernel function.
- Providing the first data set to the first kernel function may include providing patient data and training data of the first type to the first kernel function.
- Providing the second data set to the second kernel function may include providing patient data and training data of the second type to the second kernel function.
- computing the decision function result may include computing one or more composite kernels (e.g., composite kernels 212 ) consistent with a set of data type weights (e.g., data type weights 210 ).
- the computed one or more composite kernels may be combined consistent with a set of sample weights (e.g., sample weights 214 ) to generate the decision function result.
- the threshold values may be determined by an individual, such as a researcher or physician, or may be determined by the system. In certain instances, the threshold value may be specific to the application of interest. For example, the threshold value for one disease may be different than another disease.
- a disease diagnostic process may includes the acts of receiving patient data, optionally receiving training data, providing data to one or more kernel functions, computing a decision function result, and determining a probability, where the probability is linked to diagnosing a patient.
- the patient may be diagnosed with a disease or condition.
- aspects and functions described herein in accordance with the present disclosure may be implemented as hardware, software, firmware or any combination thereof. Aspects in accordance with the present disclosure may be implemented within methods, acts, systems, system elements and components using a variety of hardware, software or firmware configurations. Furthermore, aspects in accordance with the present disclosure may be implemented as specially-programmed hardware and/or software.
- FIG. 7 there is illustrated a block diagram of one example of computing components forming a system 700 which may be configured to implement one or more aspects disclosed herein.
- the system 700 may be communicatively coupled to a PCU or included within a PCU and configured to perform an MKL algorithm as described above.
- the system 700 may include for example a general-purpose computing platform such as those based on Intel PENTIUM-type processor, Motorola PowerPC, Sun UltraSPARC, Texas Instruments-DSP, Hewlett-Packard PA-RISC processors, or any other type of processor.
- System 700 may include specially-programmed, special-purpose hardware, for example, an application-specific integrated circuit (ASIC).
- ASIC application-specific integrated circuit
- Various aspects of the present disclosure may be implemented as specialized software executing on the system 700 such as that shown in FIG. 7 .
- the system 700 may include a processor/ASIC 706 connected to one or more memory devices 710 , such as a disk drive, memory, flash memory or other device for storing data. Memory 710 may be used for storing programs and data during operation of the system 700 .
- Components of the computer system 700 may be coupled by an interconnection mechanism 708 , which may include one or more buses (e.g., between components that are integrated within a same machine) and/or a network (e.g., between components that reside on separate machines).
- the interconnection mechanism 708 enables communications (e.g., data, instructions) to be exchanged between components of the system 700 . Further, in some embodiments the interconnection mechanism 708 may be disconnected during servicing of a PDU.
- the system 700 also includes one or more input devices 704 , which may include for example, a keyboard or a touch screen. An input device may be used for example to configure the measurement system or to provide input parameters.
- the system 700 includes one or more output devices 702 , which may include for example a display.
- the computer system 700 may contain one or more interfaces (not shown) that may connect the computer system 700 to a communication network, in addition or as an alternative to the interconnection mechanism 708 .
- the system 700 may include a storage system 712 , which may include a computer readable and/or writeable nonvolatile medium in which signals may be stored to provide a program to be executed by the processor or to provide information stored on or in the medium to be processed by the program.
- the medium may, for example, be a disk or flash memory and in some examples may include RAM or other non-volatile memory such as EEPROM.
- the processor may cause data to be read from the nonvolatile medium into another memory 710 that allows for faster access to the information by the processor/ASIC than does the medium.
- This memory 710 may be a volatile, random access memory such as a dynamic random access memory (DRAM) or static memory (SRAM).
- DRAM dynamic random access memory
- SRAM static memory
- the processor 706 may manipulate the data within the integrated circuit memory 710 and then copy the data to the storage 712 after processing is completed.
- a variety of mechanisms are known for managing data movement between storage 712 and the integrated circuit memory element 710 , and the disclosure is not limited thereto. The disclosure is not limited to a particular memory system 710 or a storage system 712 .
- the system 700 may include a general-purpose computer platform that is programmable using a high-level computer programming language.
- the system 700 may also be implemented using specially programmed, special purpose hardware, e.g. an ASIC.
- the system 700 may include a processor 706 , which may be a commercially available processor such as the well-known Pentium class processor available from the Intel Corporation. Many other processors are available.
- the processor 706 may execute an operating system which may be, for example, a Windows operating system available from the Microsoft Corporation, MAC OS System X available from Apple Computer, the Solaris Operating System available from Sun Microsystems, or UNIX and/or LINUX available from various sources. Many other operating systems may be used.
- the processor and operating system together may form a computer platform for which application programs in high-level programming languages may be written. It should be understood that the disclosure is not limited to a particular computer system platform, processor, operating system, or network. Also, it should be apparent to those skilled in the art that the present disclosure is not limited to a specific programming language or computer system. Further, it should be appreciated that other appropriate programming languages and other appropriate computer systems could also be used.
Landscapes
- Engineering & Computer Science (AREA)
- Medical Informatics (AREA)
- Software Systems (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Data Mining & Analysis (AREA)
- Public Health (AREA)
- Biomedical Technology (AREA)
- General Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Engineering & Computer Science (AREA)
- Databases & Information Systems (AREA)
- Pathology (AREA)
- Epidemiology (AREA)
- General Health & Medical Sciences (AREA)
- Primary Health Care (AREA)
- Measuring And Recording Apparatus For Diagnosis (AREA)
- Medical Treatment And Welfare Office Work (AREA)
Abstract
Description
K({right arrow over (x)},{right arrow over (y)})=Φ({right arrow over (x)}),Φ({right arrow over (y)})=Φ({right arrow over (x)})TΦ({right arrow over (y)}), (1)
where x and y are data vectors in the original data space, and Φ is a function that can embed the data in the higher dimensional feature space. The kernel function is an operation that can be applied to the data vectors that is equivalent to computing their inner product after they have been embedded in the feature space without actually having to map them first. Kernel evaluations can therefore be used in place of every inner product operation required by optimization algorithms designed to identify class boundaries in feature space. When applied to MKL, a different kernel may be chosen for each modality of data. Commonly used kernel functions include those represented below by
Linear: K({right arrow over (x)},{right arrow over (y)})={right arrow over (x)} T {right arrow over (y)} (2)
Gaussian Radial Basis Function:
Classifications in Feature Space
h(z)=w TϕK(z)+b, (4)
where w is a vector of feature weights, and b is a constant bias term.
Sample covariances of the positive and negative classes may be expressed as equations 7 and 8, respectively, as shown below:
In equation 9, λ is a small positive regularization parameter and I is the identity matrix. The class-separability optimization function may be reformulated in terms of kernels, where all inner product computations are replaced with their corresponding kernel transforms. The result is shown in
As a result, the decision function may be formulated in terms of kernel computations, as shown below in Equation 11:
It is appreciated that the optimal weight vector, {right arrow over (α)}*, can be solved using one or more analytical methods.
where N is the number of kernels used to create the composite kernel Kc, and the N-dimensional vector {right arrow over (β)} contains the weights on each sub-kernel, determined through the optimization routine. Similar to the single kernel case, a vector of training sample weights, {right arrow over (α)}, is also specified. In certain embodiments, the expressions shown in
Application of MKL-KFDA to Heterogeneous Data
is formed. This process is repeated until the unclassified sample vector has been compared to all m training samples in a similar fashion. Finally, the output of the
TABLE 1 |
Description of the subject population used in the analysis |
Number of MCI | 198 |
individuals at baseline | |
Number of MCI | 126 |
individuals who did not | |
convert to AD | |
Number of MCI | 72 |
individuals who did | |
convert to AD |
Breakdown of converters | Month 6 | |
Month 24 | Month 48 |
5 | 46 | 18 | 3 | |
TABLE 2 |
Description of features used in the analysis |
| Feature | Description | |
1 | ADAS | Alzheimer's Disease Assessment Score - | |
Score | evaluates cognitive impairment in the assessment | ||
of Alzheimer's disease | |||
MMSE | Mini Mental State Exam - samples functions | ||
Score | including arithmetic, memory, and orientation; | ||
tests for cognitive impairment | |||
BNT | Boston Naming Test - patients with anomic | ||
Score | aphasia often have greater difficulties with the | ||
naming both difficult and low frequency objects, | |||
as well as easy and high frequency objects | |||
ANART | American National Adult Reading Test - | ||
Score | estimates levels of |
||
2 | FDG-PET: | Glucose metabolism, left angular gyrus voxel | |
Left | |||
Angular | |||
Gyrus | |||
FDG-PET: | Glucose metabolism, right angular gyrus voxel | ||
Right | |||
Angular | |||
Gyrus | |||
FDG-PET: | Glucose metabolism, left temporal voxel | ||
Left | |||
Temporal | |||
FDG-PET: | Glucose metabolism, right temporal voxel | ||
Right | |||
Temporal | |||
FDG-PET: | Glucose metabolism, bilateral cingulum posterior | ||
Cingulum | voxel | ||
Posterior | |||
TABLE 3 |
Univariate classification results |
Cognitive Scores | FDG-PET Voxel |
ADAS | MMSE | BNT | ANART | L. | R. | L. | R. | Cing. | |
Score | Score | Score | Score | Ang | Ang | Temp | Temp | Post | |
Standard | 61.8% | 58.1% | 57.4% | 44.1% | 60.2% | 57.5% | 58.1% | 55.3% | 60.7% |
Linear | |||||||||
Classifiers | |||||||||
Single | 66.3% | 57.4% | 56.6% | 58.0% | 61.7% | 61.2% | 63.2% | 55.9% | 61.9% |
Kernel | |||||||||
SVM | |||||||||
Single | 66.4% | 58.8% | 59.7% | 59.5% | 61.8% | 61.9% | 63.4% | 57.9% | 62.7% |
Kernel | |||||||||
KFDA | |||||||||
TABLE 4 |
Pairwise and three-way univariate predictor analysis |
Univariate Predictor Variables | % Agree | % Disagree | ||
ADAS + BNT | 62% | 38% | ||
BNT + FDG-PET | 49% | 51% | ||
ADAS + FDG-PET | 61% | 39% | ||
ADAS + BNT + FDG-PET | 43% | 57% | ||
Claims (18)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/183,018 US10304006B2 (en) | 2013-02-15 | 2014-02-18 | Method for integrating and fusing heterogeneous data types to perform predictive analysis |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201361765530P | 2013-02-15 | 2013-02-15 | |
US14/183,018 US10304006B2 (en) | 2013-02-15 | 2014-02-18 | Method for integrating and fusing heterogeneous data types to perform predictive analysis |
Publications (2)
Publication Number | Publication Date |
---|---|
US20140236872A1 US20140236872A1 (en) | 2014-08-21 |
US10304006B2 true US10304006B2 (en) | 2019-05-28 |
Family
ID=51352023
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/183,018 Active 2036-08-23 US10304006B2 (en) | 2013-02-15 | 2014-02-18 | Method for integrating and fusing heterogeneous data types to perform predictive analysis |
Country Status (1)
Country | Link |
---|---|
US (1) | US10304006B2 (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20210082575A1 (en) * | 2019-09-18 | 2021-03-18 | Cerner Innovation, Inc. | Computerized decision support tool for post-acute care patients |
US11537374B1 (en) | 2021-06-03 | 2022-12-27 | Oracle International Corporation | System and method for hot method call graph analysis |
US11687621B2 (en) | 2021-03-29 | 2023-06-27 | International Business Machines Corporation | Multi-modal fusion techniques considering inter-modality correlations and computer model uncertainty |
Families Citing this family (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10491663B1 (en) * | 2013-10-28 | 2019-11-26 | Amazon Technologies, Inc. | Heterogeneous computations on homogeneous input data |
KR101758055B1 (en) * | 2014-10-27 | 2017-07-14 | 삼성에스디에스 주식회사 | Method and apparatus for analyzing patient's constitutional peculiarity |
US10679143B2 (en) | 2016-07-01 | 2020-06-09 | International Business Machines Corporation | Multi-layer information fusing for prediction |
JP2019525343A (en) * | 2016-08-08 | 2019-09-05 | コーニンクレッカ フィリップス エヌ ヴェKoninklijke Philips N.V. | Optimization of electronic clinical decision support equipment based on hospital demographics |
CN110211697B (en) * | 2019-05-29 | 2023-06-09 | 安徽医科大学第一附属医院 | Electronic data analysis method of gastroenterology based on SVM model |
CN111598156B (en) * | 2020-05-14 | 2024-07-19 | 北京工业大学 | PM based on multi-source heterogeneous data fusion2.5Prediction method |
CN112017776B (en) * | 2020-10-27 | 2021-01-15 | 平安科技(深圳)有限公司 | Disease prediction method based on dynamic graph and medical knowledge map and related equipment |
CN112348083B (en) * | 2020-11-06 | 2024-09-17 | 湖北玥研科技有限公司 | Image classification method and device |
CN113159371B (en) * | 2021-01-27 | 2022-05-20 | 南京航空航天大学 | Unknown target feature modeling and demand prediction method based on cross-modal data fusion |
CN115005822A (en) * | 2022-05-13 | 2022-09-06 | 中国科学院计算技术研究所 | A method and system for ADHD intelligent decision support based on heterogeneous graph |
CN115600111B (en) * | 2022-11-07 | 2024-12-20 | 宁波吉利汽车研究开发有限公司 | Resource prediction model training method, cloud resource prediction method and device |
Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO1998022885A1 (en) | 1996-11-20 | 1998-05-28 | Jannarone Robert J | Multi-kernel neural network concurrent learning, monitoring, and forecasting system |
US6216119B1 (en) | 1997-11-19 | 2001-04-10 | Netuitive, Inc. | Multi-kernel neural network concurrent learning, monitoring, and forecasting system |
US6569403B1 (en) * | 1995-09-01 | 2003-05-27 | Miicro, Incorporated | Determining effects of external stimuli on the brain using PET |
US20040103001A1 (en) * | 2002-11-26 | 2004-05-27 | Mazar Scott Thomas | System and method for automatic diagnosis of patient health |
US20040265919A1 (en) * | 2003-05-22 | 2004-12-30 | Hugo Vanderstichele | Method for the prediction, diagnosis and differential diagnosis of Alzheimer's disease |
US20090043795A1 (en) * | 2007-08-08 | 2009-02-12 | Expanse Networks, Inc. | Side Effects Prediction Using Co-associating Bioattributes |
US20110257025A1 (en) * | 2008-04-11 | 2011-10-20 | Houtman Rene | Method for tumor classification |
US20130097108A1 (en) | 2011-10-13 | 2013-04-18 | Nec Laboratories America, Inc. | Two-Stage Multiple Kernel Learning Method |
US8566268B2 (en) | 2010-10-08 | 2013-10-22 | International Business Machines Corporation | System and method for composite distance metric leveraging multiple expert judgments |
US8595155B2 (en) | 2010-03-23 | 2013-11-26 | International Business Machines Corporation | Kernel regression system, method, and program |
US9087294B2 (en) * | 2012-01-11 | 2015-07-21 | International Business Machines Corporation | Prediction method, prediction system and program |
US9211294B2 (en) * | 2007-06-19 | 2015-12-15 | Wista Laboratories Ltd. | Phenothiazine compounds for treating mild cognitive impairment |
-
2014
- 2014-02-18 US US14/183,018 patent/US10304006B2/en active Active
Patent Citations (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6569403B1 (en) * | 1995-09-01 | 2003-05-27 | Miicro, Incorporated | Determining effects of external stimuli on the brain using PET |
WO1998022885A1 (en) | 1996-11-20 | 1998-05-28 | Jannarone Robert J | Multi-kernel neural network concurrent learning, monitoring, and forecasting system |
US6216119B1 (en) | 1997-11-19 | 2001-04-10 | Netuitive, Inc. | Multi-kernel neural network concurrent learning, monitoring, and forecasting system |
US6647377B2 (en) | 1997-11-19 | 2003-11-11 | Netuitive, Inc. | Multi-kernel neural network concurrent learning, monitoring, and forecasting system |
US20040103001A1 (en) * | 2002-11-26 | 2004-05-27 | Mazar Scott Thomas | System and method for automatic diagnosis of patient health |
US20040265919A1 (en) * | 2003-05-22 | 2004-12-30 | Hugo Vanderstichele | Method for the prediction, diagnosis and differential diagnosis of Alzheimer's disease |
US9211294B2 (en) * | 2007-06-19 | 2015-12-15 | Wista Laboratories Ltd. | Phenothiazine compounds for treating mild cognitive impairment |
US20090043795A1 (en) * | 2007-08-08 | 2009-02-12 | Expanse Networks, Inc. | Side Effects Prediction Using Co-associating Bioattributes |
US20110257025A1 (en) * | 2008-04-11 | 2011-10-20 | Houtman Rene | Method for tumor classification |
US8595155B2 (en) | 2010-03-23 | 2013-11-26 | International Business Machines Corporation | Kernel regression system, method, and program |
US8566268B2 (en) | 2010-10-08 | 2013-10-22 | International Business Machines Corporation | System and method for composite distance metric leveraging multiple expert judgments |
US20130097108A1 (en) | 2011-10-13 | 2013-04-18 | Nec Laboratories America, Inc. | Two-Stage Multiple Kernel Learning Method |
US9087294B2 (en) * | 2012-01-11 | 2015-07-21 | International Business Machines Corporation | Prediction method, prediction system and program |
Non-Patent Citations (10)
Title |
---|
"Prediction of Alzheimer's Disease Using the CSF AB42/AB40 Ratio in Patients with Mild Cognitive Impairment" Hansson et al, Dement Geriatr Cogn Disord 2007;23:316-320 DOI: 10.1159/000100926 Accepted: Jan. 18, 2007 Published online: Mar. 19, 2007. * |
Cortes, C. and V. Vapnik (1996). "Support-Vector Networks." Machine Learning 20(3): 273-297. |
Gonen, M., E. Alpaydm (2011). "Multiple Kernel Learning Algorithms." Journal of Machine Learning Research. 12: 2211-2268. |
Hinrichs, C. V. Singh, et al. (2011). "Predictive markers for AD in a multi-modality framework: An analysis of MCI progression in the ADNI population." NeuroImage 55(2): 574-589. |
Mika, S., G. Ratsch, et al. (1999). Fisher discriminant analysis with kernels. Neural Networks for Signal Processing IX, 1999. Proceedings of the 1999 IEEE Signal Processing Society Workshop. |
Mika, S., G. Ratsch, et al. (2003). "Constructing descriptive and discriminative nonlinear features: Rayleigh coefficients in kernel feature spaces." Pattern Analysis and Machine Intelligence, IEEE Transactions on 25(5): 623-628. |
Shawne-Taylor, J. and N. Cristianini (2004). "Properties of Kernels," Kernel Methods for Pattern Analysis, Cambridge University Press, pp. 47-59. |
Ye, J., K. Chen, et al. (2008). Heterogeneous Data Fusion for Alzheimer's Disease Study. Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. Las Vegas, Nevada, USA, ACM: 1025-1033. |
Ye, J., S. Ji, et al. (2008). "Multi-class Discriminant Kernel Learning via Convex Programming." J. Mach. Learn. Res. 9: 719-758. |
Zien, A., C.S. Ong, (2007). "Multiclass Multiple Kernel Learning." Proceedings of the 24th International Conference on Machine Learning. Corvallis, OR, USA. |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20210082575A1 (en) * | 2019-09-18 | 2021-03-18 | Cerner Innovation, Inc. | Computerized decision support tool for post-acute care patients |
US11687621B2 (en) | 2021-03-29 | 2023-06-27 | International Business Machines Corporation | Multi-modal fusion techniques considering inter-modality correlations and computer model uncertainty |
US11537374B1 (en) | 2021-06-03 | 2022-12-27 | Oracle International Corporation | System and method for hot method call graph analysis |
Also Published As
Publication number | Publication date |
---|---|
US20140236872A1 (en) | 2014-08-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10304006B2 (en) | Method for integrating and fusing heterogeneous data types to perform predictive analysis | |
Nogay et al. | Machine learning (ML) for the diagnosis of autism spectrum disorder (ASD) using brain imaging | |
Arya et al. | A systematic review on machine learning and deep learning techniques in the effective diagnosis of Alzheimer’s disease | |
Pisner et al. | Support vector machine | |
Sudharsan et al. | Alzheimer's disease prediction using machine learning techniques and principal component analysis (PCA) | |
Amini et al. | Diagnosis of Alzheimer’s disease severity with fMRI images using robust multitask feature extraction method and convolutional neural network (CNN) | |
Dimitriadis et al. | How random is the random forest? Random forest algorithm on the service of structural imaging biomarkers for Alzheimer's disease: from Alzheimer's disease neuroimaging initiative (ADNI) database | |
Nsugbe | Toward a self-supervised architecture for semen quality prediction using environmental and lifestyle factors | |
Dubey et al. | Analysis of sampling techniques for imbalanced data: An n= 648 ADNI study | |
Toshkhujaev et al. | Classification of Alzheimer’s disease and mild cognitive impairment based on cortical and subcortical features from MRI T1 brain images utilizing four different types of datasets | |
Jiang et al. | Correlation-aware sparse and low-rank constrained multi-task learning for longitudinal analysis of Alzheimer's disease | |
Jia et al. | Deep learning and multimodal feature fusion for the aided diagnosis of Alzheimer's disease | |
Han et al. | A novel convolutional variation of broad learning system for Alzheimer’s disease diagnosis by using MRI images | |
AlMohimeed et al. | Explainable artificial intelligence of multi-level stacking ensemble for detection of Alzheimer’s disease based on particle swarm optimization and the sub-scores of cognitive biomarkers | |
KP et al. | Alzheimer's classification using dynamic ensemble of classifiers selection algorithms: A performance analysis | |
Humayun et al. | Superlative feature selection based image classification using deep learning in medical imaging | |
Wang et al. | Exploring automated machine learning for cognitive outcome prediction from multimodal brain imaging using streamline | |
Huang et al. | Computational optimization and statistical methods for big data analytics: Applications in neuroimaging | |
Majumder et al. | A Bayesian group lasso classification for ADNI volumetrics data | |
Nguyen et al. | Knockout: A simple way to handle missing inputs | |
Kanwal et al. | Alzheimer's disease classification using abnormal hippocampal functional connectivity and machine learning | |
Larsen | Data-assisted differential diagnosis of dementia by deep neural networks | |
Kawaguchi et al. | Ensemble clustering multiblock sparse multivariable analysis for multimodal brain imaging | |
Mousavian | Machine Learning Methods for Depression Detection Using SMRI and RS-FMRI Images | |
RS et al. | Alzheimer’s disease (AD) classification using MRI: A deep ensemble model with modified local pattern feature set |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: THE CHARLES STARK DRAPER LABORATORY, INC., MASSACH Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KESHAVA, NIRMAL;MARIANO, LAURA JANE;SIGNING DATES FROM 20140303 TO 20140311;REEL/FRAME:032441/0115 |
|
AS | Assignment |
Owner name: US ARMY, SECRETARY OF THE ARMY, MARYLAND Free format text: CONFIRMATORY LICENSE;ASSIGNOR:CHARLES STARK DRAPER LABORATORY;REEL/FRAME:033258/0694 Effective date: 20140325 |
|
AS | Assignment |
Owner name: US ARMY, SECRETARY OF THE ARMY, MARYLAND Free format text: CONFIRMATORY LICENSE;ASSIGNOR:CHARLES STARK DRAPER LABORATORY;REEL/FRAME:036910/0701 Effective date: 20140325 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT RECEIVED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2551); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY Year of fee payment: 4 |