Diversity in Independent Component and Vector Analyses: Identifiability, algorithms, and applications in medical imaging
Links to Fileshttps://ieeexplore.ieee.org/document/6784026/authors#authors
MetadataShow full item record
Type of Work16 pages
journal articles postprints
Citation of Original PublicationTulay Adali , Matthew Anderson, Geng-Shen Fu, Diversity in Independent Component and Vector Analyses: Identifiability, algorithms, and applications in medical imaging , IEEE Signal Processing Magazine ( Volume: 31 , Issue: 3 , May 2014) , DOI: 10.1109/MSP.2014.2300511
RightsThis item is likely protected under Title 17 of the U.S. Copyright Law. Unless on a Creative Commons license, for uses protected by Copyright Law, contact the copyright holder or the author.
© 2014 IEEE
Subjectshigher order statistics
independent component analysis
maximum likelihood estimation
signal processing algorithms
medical image processing
Machine Learning for Signal Processing Lab
Starting with a simple generative model and the assumption of statistical independence of the underlying components, independent component analysis (ICA) decomposes a given set of observations by making use of the diversity in the data, typically in terms of statistical properties of the signal. Most of the ICA algorithms introduced to date have considered one of the two types of diversity: non-Gaussianity?i.e., higher-order statistics (HOS)?or, sample dependence. A recent generalization of ICA, independent vector analysis (IVA), generalizes ICA to multiple data sets and adds the use of one more diversity, dependence across multiple data sets for achieving an independent decomposition, jointly across multiple data sets. Finally, both ICA and IVA, when implemented in the complex domain, enjoy the addition of yet another type of diversity, noncircularity of the sources?underlying components. Mutual information rate provides a unifying framework such that all these statistical properties?types of diversity?can be jointly taken into account for achieving the independent decomposition. Most of the ICA methods developed to date can be cast as special cases under this umbrella, as well as the more recently developed IVA methods. In addition, this formulation allows us to make use of maximum likelihood theory to study large sample properties of the estimator, derive the Cramer-Rao lower bound(CRLB) and determine the conditions for the identifiability of the ICA and IVA models. In this overview article, we first present ICA, and then its generalization to multiple data sets, IVA, both using mutual information rate, present conditions for the identifiability of the given linear mixing model and derive the performance bounds. We address how various methods fall under this umbrella and give examples of performance for a few sample algorithms compared with the performance bound. We then discuss the importance of approaching the performance bound depending on the goal, and use medical image analysis as the motivating example.