INSTITUTIONAL DIGITAL REPOSITORY

Biomedical signal and image synthesis using deep learning

Show simple item record

dc.contributor.author Sikka, A.
dc.date.accessioned 2020-12-31T11:08:00Z
dc.date.available 2020-12-31T11:08:00Z
dc.date.issued 2020-12-31
dc.identifier.uri http://localhost:8080/xmlui/handle/123456789/1710
dc.description.abstract Accurate modeling and characterization of biomedical data is imperative for comprehensive understanding and reliable medical diagnosis. However, developing automated solutions to various biomedical challenges is a complex task as biomedical data is inherently high-dimensional with large variability and non-linear and non-stationary dynamics. Consequently, machine learning algorithms that offer data-driven model-free approach have been used extensively for a multitude of tasks such as synthesis, classification and segmentation. However, conventional methods that use hand-engineered features (based on prior knowledge) do not generalize well due to heterogeneity inherent to biomedical data. In recent years, deep learning models have consistently outperformed previous state-of-the-art machine learning techniques in several fields, with computer vision being one of the most prominent domains. With their ability to exploit compositional structure to find complex patterns in high dimensional and heterogeneous data, deep learning has emerged as a critical tool for analyzing biomedical data. With growing availability of public datasets, several efforts have been directed towards synthesizing biomedical signals and images based on phenomenologic or datadriven models to address diverse range of challenges such as data augmentation, signal encoding and transfer of knowledge across modalities or domains. Inspired by their overwhelming success in the field of computer vision, this thesis investigates the feasibility of using deep learning algorithms for biomedical signal and image synthesis. Specifically, we addressed the following challenges: (a) Synthesizing functional Positron Emission Tomography (PET) scans from structural Magnentic Resonance (MR) images to facilitate accurate diagnosis of Alzheimer’s disease with incomplete multi-modal data (b) Capturing complex temporal patterns underlying Electroencephalography (EEG) microstate sequences for better characterization of neuro-psychiatric disorders and (c) Learning a non-linear transfer function to map EEG signals to corresponding blood oxygenation level dependent (BOLD) signals to investigate the relationship between brain’s electrophysiological and hemodynamic measures. Generic image-to-image translation, signal encoding and sequence-to-sequence models were tailored to address these specific tasks by incorporating domain specific knowledge. Extensive evaluation of our proposed techniques against state-of-the-art algorithms demonstrates their ability to handle complex, multi-modal, multi-variate biomedical data towards providing better understanding and clinical decision support. Finally, we utilized visualization and interpretation strategies to overcome the limited interpretability of the black-box deep learning models. These tools were most effective in identifying key features and meaningful representations that proved valuable in gaining novel insights from the data. en_US
dc.language.iso en_US en_US
dc.title Biomedical signal and image synthesis using deep learning en_US
dc.type Thesis en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search DSpace


Advanced Search

Browse

My Account