Jakob Wasza | 2019 |
Dominik Neumann | 2019 |
Peter Fischer | 2019 |
Chris Schwemmer | 2019 |
Florin Ghesu | 2019 |
Felix Lugauer-Breunig | 2018 |
Vincent Christlein | 2018 |
Tanja Kurzendorfer | 2018 |
Matthias Weidler | 2018 |
Markus Mayer | 2018 |
Peter Fürsattel | 2018 |
Oliver Taubmann | 2018 |
Thomas Köhler | 2017 |
Mathias Unberath | 2017 |
Johannes Jordan | 2017 |
Christoph Forman | 2017 |
Martin Kraus | 2016 |
Florian Hönig | 2016 |
Bharath Navalpakkam | 2016 |
Simon Schöll | 2016 |
Firas Mualla | 2016 |
Martin Berger | 2016 |
Jan Paulus | 2016 |
Sven Haase | 2016 |
Yan Xia | 2016 |
Ulf Jensen | 2016 |
Juan Rafael Orozco-Arroyave | 2015 |
Robert Grimm | 2015 |
Michal Cachovan | 2015 |
Michael Manhart | 2014 |
Benjamin Keck | 2014 |
Sebastian Bauer | 2014 |
Davide Piccini | 2014 |
Arne Militzer | 2014 |
Jana Hutter | 2014 |
Kerstin Müller | 2014 |
Eva Rothgang | 2014 |
Philipp Ritt | 2013 |
Zhicong Yu | 2013 |
Stefan Wenhardt | 2013 |
Ahmad Osman | 2013 |
Konrad Sickel | 2013 |
Christian Riess | 2012 |
Korbinian Riedhammer | 2012 |
Dirk Kolb | 2012 |
Alexander Brost | 2012 |
Ahmed El-Rafei | 2012 |
Johannes Feulner | 2012 |
Christoph Schmalz | 2012 |
Volker Daum | 2011 |
Christian Schaller | 2011 |
Michael Balda | 2011 |
Andreas Fieselmann | 2011 |
Martin Spiegel | 2011 |
Christoph Gütter | 2011 |
Christopher Rohkohl | 2010 |
Johannes Zeintl | 2010 |
Kurt Höller | 2010 |
Holger Scherl | 2010 |
Florian Jäger | 2010 |
Michael Wels | 2010 |
Björn Eskofier | 2010 |
Martin Raab | 2010 |
Anja Borsdorf | 2009 |
Jingfeng Han | 2009 |
Dieter Hahn | 2009 |
Marcus Prümmer | 2009 |
Andreas Maier | 2009 |
Frank Dennerlein | 2008 |
Stefan Hoppe | 2008 |
|
|||
Real-time Respiratory Motion Analysis Using GPU-accelerated Range Imaging
Abstract:
Respiratory motion analysis and management are crucial issues for a plurality of medical applications. Of particular scientific concern are methods that allow to analyze the patient’s breathing in a non-invasive and real-time manner that does not involve ionizing radiation exposure. For this purpose, range imaging technologies, that allow to dynamically acquire three-dimensional body surface data, have been proposed over the last years. A particular challenge with such methods is a fully automatic investigation and assessment of the body surface data, as well as computation times that comply with real-time constraints. This dissertation is concerned with the application of range imaging principles for real-time automatic respiratory motion analysis. The focus is on the development of efficient methods for data preprocessing and fusion as well as machine learning and surface registration techniques. A particular emphasis of this thesis is the design of the proposed algorithms for GPU architectures to enable real-time computation. The first part of this thesis covers the general challenges and requirements for respiratory motion analysis in diagnostic and therapeutic applications. Furthermore, the range imaging technologies that are relevant for this thesis are introduced and the suitability of GPU architectures for accelerating the computation of several tasks inherent to range imaging based respiratory motion analysis are investigated. The second part of this work is concerned with pre-processing and fusion techniques for range data. To account for the low signal-to-noise ratio that is common with range data, this work proposes a processing pipeline that allows to reconstruct the ideal data with an error trueness less than 1.0 mm at run-times of 2 ms. For fusing range image data in a multi-camera setup, as it is required for the simultaneous acquisition of frontal and lateral body surface, this thesis proposes a novel framework that enables the computation of a 180◦ coverage body surface model consisting of more than 3.0 × 105 points with a computation time of less than 5 ms. The third part of this work is concerned with patient-specific respiratory motion models. The thesis proposes machine learning techniques to generate a continuous motion model that features the ability to automatically differentiate between thoracic and abdominal breathing as well as to quantitatively analyze the patient’s respiration magnitude. By using purposely developed surface registration schemes, these models are then brought in congruence with body surface data acquired by range imaging sensors. This allows for respiratory motion compensated patient positioning that reduces the alignment error observed with conventional approaches by a factor of 3 to less than 4.0 mm. Further, this approach allows to automatically derive a multi-dimensional respiration surrogate that yields a correlation coefficient greater than 0.97 compared to commonly employed invasive or semi-automatic approaches and that can be computed in 20 ms. The fourth part concludes this thesis with a summary of the presented methods and results, as well as an outlook regarding future research directions and challenges towards clinical translation. |
|
|||
Robust Personalization of Cardiac Computational Models
Abstract:
Heart failure (HF) is a major cause of morbidity and mortality in the Western world, yet early diagnosis and treatment remain a major challenge. As computational cardiac models are becoming more mature, they are slowly evolving into clinical tools to better stratify HF patients, predict risk and perform treatment planning. A critical prerequisite, however, is their ability to precisely capture an individual patient’s physiology. The process of fitting a model to patient data is called personalization, which is the overarching topic of this thesis. An image-based, multi-physics 3D whole-heart model is employed in this work. It consists of several components covering anatomy, electrophysiology, biomechanics and hemodynamics. Building upon state-of-the-art personalization techniques, the first goal was to develop an automated pipeline for personalizing all components of the model in a streamlined and reproducible fashion, based on routinely acquired clinical data. Evaluation was performed on a multi-clinic cohort consisting of 113 patients, the largest cohort in any comparable study to date. The goodness of fit between per- sonalized models and ground-truth clinical data was mostly below clinical variability, while a full personalization was finalized within only few hours. This showcases the ability of the proposed pipeline to extract advanced biophysical parameters robustly and efficiently. Designing such personalization algorithms is a tedious, model- and data-specific process. The second goal was to investigate whether artificial intelligence (AI) con- cepts can be used to learn this task, inspired by how humans manually perform it. A self-taught artificial agent based on reinforcement learning (RL) is proposed, which first learns how the model behaves, then computes an optimal strategy for person- alization. The algorithm is model-independent; applying it to a new model requires only adjusting few hyper-parameters. The obtained results for two different mod- els suggested that equivalent, if not better goodness of fit than standard methods could be achieved, while being more robust and with faster convergence rate. AI ap- proaches could thus make personalization algorithms generalizable and self-adaptable to any patient and any model. Due to limited data, uncertainty in the clinical measurements, parameter non- identifiability, and modeling assumptions, various combinations of parameter val- ues may exist that yield the same quality of fit. The third goal of this work was uncertainty quantification (UQ) of the estimated parameters and to ascertain the uniqueness of the found solution. A stochastic method based on Bayesian inference and fast surrogate models is proposed, which estimates the posterior of the model, taking into account uncertainties due to measurement noise. Experiments on the biomechanics model showed that not only goodness of fits equivalent to the standard methods could be achieved, but also the non-uniqueness of the problem could be demonstrated and uncertainty estimates reported, crucial information for subsequent clinical assessments of the personalized models. |
|
|||
Respiratory Motion Compensation in X-Ray Fluoroscopy
Abstract:
Fluoroscopy is a common imaging modality in medicine for guidance of minimally invasive interventions due to its high temporal and spatial resolution and the good visibility of interventional devices and bones. To counteract its lack of 3-D information and soft-tissue contrast, the X-ray images can be enhanced with overlays. This constitutes a medical application of augmented reality technology. Most commonly, the overlays are static. Due to inevitable respiratory and cardiac motion of the patient when imaging chest or abdomen, the images and the overlays frequently are inconsistent. In this thesis, two methods for compensating this involuntary motion are presented. |
|
|||
3-D Imaging of Coronary Vessels Using C-arm CT
Abstract:
Cardiovascular disease has become the number one cause of death worldwide. For the diagnosis and therapy of coronary artery disease, interventional C-arm-based fluoroscopy is animaging method of choice. |
|
|||
Artificial Intelligence for Medical Image Understanding
Abstract:
Robust and fast detection and segmentation of anatomical structures in medical image data represents an important component of medical image analysis technologies. Current solutions for this problem are nowadays based on machine learning techniques that exploit large annotated image databases in order to learn the appearance of the captured anatomy. These solutions are subject to several limitations. This includes the use of suboptimal image feature engineering methods and most importantly the use of computationally suboptimal search-schemes for anatomy parsing, e.g., exhaustive hypotheses scanning. In particular, these techniques do not effectively address cases of incomplete data, i.e., scans acquired with a partial field-of-view. To address these challenges, we introduce in this thesis marginal space deep learning, a framework for medical image parsing which exploits the automated feature design of deep learning models and an efficient object parametrization scheme in hierarchical marginal spaces. To support the efficient evaluation of solution hypotheses under complex transformations, such as rotation and anisotropic scaling, we propose a novel cascaded network architecture, called sparse adaptive neural network. Experiments on detecting and segmenting the aortic root in 2891 3D ultrasound volumes from 869 patients, demonstrate a high level of robustness with an accuracy increase of 30-50% against the state-of-the-art. Despite these facts, using a scanning routine to explore large parameter subspaces results in a high computational complexity, with false-positive predictions and limited scalability to high resolution volumetric data. To deal with these limitations, we propose a novel paradigm for medical image parsing, based on principles of cognitive modeling and behavior learning. The anatomy detection problem is reformulated as a behavior learning task for an intelligent artificial agent. Using deep reinforcement learning, agents are taught how to search for an anatomical structure. This resumes to learning to navigate optimal search trajectories through the image space that converge to the locations of the sought anatomical structures. To support the effective parsing of high resolution volumetric data, we apply elements from scale-space theory and enhance our framework to support the learning of multi-scale search strategies through the scale-space representation of medical images. Finally, to enable the accurate recognition whether certain anatomical landmarks are missing from the field-of-view, we exploit prior knowledge about the anatomy and ensure the spatial coherence of the agents by using statistical shape modeling and robust estimation theory. Comprehensive experiments demonstrate a high level of accuracy, compared to state-of-the-art solutions, without failures of clinical significance. In particular, our method achieves 0% false positive and 0% false negative rates at detecting whether anatomical structures are captured in the field-of-view (excluding border cases). The dataset contains 5043 3D computed tomography volumes from over 2000 patients, totaling over 2,500,000 image slices. A significant increase in accuracy compared to reference solutions is achieved on additional 2D ultrasound and 2D/3D magnetic resonance datasets, containing up to 1000 images. Most importantly, this paradigm improves the detection-speed of the previous solutions by 2-3 orders of magnitude, achieving unmatched real-time performance on high resolution volumetric scans. |
|
|||
Iterative Reconstruction Methods for Accelerating Quantitative Abdominal MRI
Abstract:
|
|
|||
Handwriting with Focus on Writer Identification and Writer Retrieval
Abstract:
In the course of the mass digitization of historical as well as contemporary sources, an individual examination by means of historical or forensic experts is no longer feasible. A solution could be an automatic handwriting analysis that determines or suggests script attributes, such as the writer or the date of a document. In this work, several novel techniques based on machine learning are presented to obtain these attributes from a single document image. The focus lies on writer recognition for which a novel pipeline is developed, which identifies the correct writer of a given sample in over 99 % of all tested contemporary datasets, numbering between 150 and 310 writers each, with four to five samples per writer. In a large historical dataset, consisting of 720 writers and five samples per writer, an identification rate of close to 90 % is achieved. Robust local descriptors play a major role in the success of this pipeline. Shape- and histogram-based descriptors prove to be very effective. Furthermore, novel deep-learningbased features are developed using deep convolutional neural networks, which are trained with writer information from the training set. While these features achieve very good results in contemporary data, they lack distinctiveness in the evaluated historical dataset. Therefore, a novel feature learning technique is presented that solves this by learning robust writer-independent script features in an unsupervised manner. The computation of a global descriptor from the local descriptors is the next step. For this encoding procedure, various techniques from the speech and computer vision community are investigated and thoroughly evaluated. It is important to counter several effects, such as feature correlation and the over-counting of local descriptors. Overall, methods based on aggregating first order statistics of residuals are the most effective approaches. Common writer recognition methods use the global descriptors directly for comparison. In contrast, exemplar classifiers are introduced in this thesis allowing sample-individual similarities to be learned, which are shown to be very effective for an improved writer recognition. This writer recognition pipeline is adapted to other tasks related to digital paleography. Medieval papal charters are automatically dated up to an error range of 17 years. Furthermore, an adapted pipeline is among the best to classify medieval Latin manuscripts into twelve different script types. This information can then be used for a pre-sorting of documents or as a preprocessing step for handwritten text recognition. It turns out that counteracting different illumination and contrast effects is an important factor for deep-learning-based approaches. The observation that script has similar tubal structures to blood vessels is exploited for an improved text block segmentation in historical data by means of a well-known medical filtering technique. This work sets new recognition standards for several tasks, allowing the automatic document analysis of large corpora with low error rates. These methods are also applicable to other fields, such as forensics or paleography, to determine writers, script types or other metadata of contemporary or historical documents. |
|
|||
Fully Automatic Segmentation of Anatomy and Scar from LGE-MRI
Abstract:
The leading cause of death worldwide are cardiovascular diseases. In addition, the number of patients suffering from heart failure is rising. The underlying cause of heart failure is often a myocardial infarction. For diagnosis in clinical routine, cardiac magnetic resonance imaging is used, as it provides information about morphology, blood flow, perfusion, and tissue characterization. In more detail, the analysis of the tissue viability is very important for diagnosis, procedure planning, and guidance, i.e., for implantation of a bi-ventricular pacemaker. The clinical gold standard for the viability assessment is 2-D late gadolinium enhanced magnetic resonance imaging (LGE-MRI). In the last years, the imaging quality continuously improved and LGE-MRI was extended to a 3-D whole heart scan. This scan guarantees an accurate quantification of the myocardium to the extent of myocardial scarring. The main challenge arises in the accurate segmentation and analysis of such images. In this work, novel methods for the segmentation of the LGE-MRI data sets, both 2-D and 3-D, are proposed. One important goal is the direct segmentation of the LGE-MRI and the independence of an anatomical scan to avoid errors from the anatomical scan contour propagation. For the 2-D LGE-MRI segmentation, the short axis stack of the left ventricle (LV) is used. First, the blood pool is detected and a rough outline is maintained by a morphological active contours without edges approach. Afterwards, the endocardial and epicardial boundary is estimated by either a filter or learning based method in combination with a minimal cost path search in polar space. For the endocardial contour refinement, an additional scar exclusion step is added. For the 3-D LGE-MRI, the LV is detected within the whole heart scan. In the next step, the short axis view is estimated using principal component analysis. For the endocardial and epicardial boundary estimation also a filter based or learning based approach can be applied in combination with dynamic programming in polar space. Furthermore, because of the high resolution also the papillary muscles are segmented. In addition to the fully automatic LV segmentation approaches, a generic semi- automatic method based on Hermite radial basis function interpolation is introduced in combination with a smart brush. Effective interactions with less number of equations accelerate the performance and therefore, a real-time and an intuitive, interactive segmentation of 3-D objects is supported effectively. After the segmentation of the left ventricle’s myocardium, the scar tissue is quantified. In this thesis, three approaches are investigated. The full-width-at-half-max algorithm and the x-standard deviation methods are implemented in a fully automatic manner. Furthermore, a texture based scar classification algorithm is introduced. Subsequently, the scar tissue can be visualized, either in 3-D as a surface mesh or in 2-D projected onto the 16 segment bull’s eye plot of the American Heart Association. However, for precise procedure planning and guidance, the information about the scar transmurality is very important. Hence, a novel scar layer visualization is introduced. Therefore, the scar tissue is divided into three layers depending on the location of the scar within the myocardium. With this novel visualization, an easy distinction between endocardial, mid-myocardial, or epicardial scar is possible. The scar layers can also be visualized in 3-D as surface meshes or in 2-D projected onto the 16 segment bull’s eye plot. |
|
|||
Automated Glaucoma Detection with Optical Coherence Tomography
Abstract:
The number of patients suffering from the glaucoma disease will increase in the future. A further automation of parts of the diagnostic routine is inevitable to use limited examination times more efficiently. Optical coherence tomography (OCT) technology has become a widespread tool for glaucoma diagnosis, and data collections in the clinics have been built up in recent years that now allow for data mining and pattern recognition approaches to be applied to the diagnostic challenge. A complete pattern recognition pipeline to automatically discriminate glaucomatous from normal eyes with OCT data is proposed, implemented and evaluated. A data collection of 1024 Spectralis HRA+OCT circular scans around the optic nerve head from 565 subjects build the basis for this work. The data collection is labeled with 4 diagnoses: 453 healthy (H), 179 ocular hypertension (OHT), 168 preperimetric glaucoma (PPG), and 224 perimetric glaucoma (PG) eyes. In a first step, 6 retinal layer boundaries are automatically segmented by edge detection and the minimization of a custom energy functional, which was established in preceeding work by the author. The segmentation algorithm is evaluated on a subset consisting of 120 scans. The automatically segmented layer boundaries are compared to a gold standard (GS) created from manual corrections to the automated results by 5 observers. The mean absolute difference of the automated segmentation to the GS for the outer nerve fiber layer boundary is 2.84mum. The other layers have less or almost no segmentation error. No significant correlation between the segmentation error and scans of bad quality or glaucomatous eyes could be found for any layer boundary. The difference of the automated segmentation to the GS is not much worse than the single observer’s manual correction difference to the GS. In a second step, the thickness profiles generated by the segmentation are used in a classification system: In total, 762 features are generated, including novel ratio and principal component analysis features. “Forward selection and backward elimination” selects the best performing features with respect to the classwise averaged classification rate (CR) on the training data. The segmentations of the complete dataset were manually corrected so that the classification experiments could either be run on manually corrected or purely automated segmentations. Three classifiers were compared. The support vector machine classifier (SVM) performed best in a 10-fold cross-validation and differentiated non-glaucomatous (H and OHT) from glaucomatous (PPG and PG) eyes with a CR of 0.859 on manually corrected data. The classification system adapts to the less reliable purely automated segmentations by choosing features of a more global scale. Training with manually corrected and testing with purely automated data and vice versa shows that it is of advance to use manually corrected data for training, no matter what the type of test data is. The distance of the feature vectors to the SVM decision boundary is used as a basis for a novel glaucoma probability score based on OCT data, the OCT-GPS. |
|
|||
Accurate Measurements with Off-the-Shelf Range Cameras
Abstract:
Distance cameras have gained large popularity in the last years. With more than 24 million Microsoft Kinect units sold and the proliferation of 3-D sensors for biometric authentication, these cameras have reached the mass market. Distance cameras capture an image in which each pixel encodes the distance to its corresponding point in the scene. This opens up new application possibilities which are difficult or even impossible to implement with normal gray-level or color cameras. These new applications are particular useful if they can be implemented with low-cost consumer 3-D cameras. However, this is problematic as these sensors have only limited accuracy compared to professional measurement systems and are thus not yet sufficient for many applications. In this thesis, several aspects that affect the accuracy of time-of-flight and structured light cameras are discussed.
The calibration of cameras, i. e. the calculation of an exact camera model, is of major importance. The estimation of these models requires point correspondences between the scene and the camera image. Whenever high accuracy is required, it is recommended to use calibration patterns such as checkerboards. This thesis introduces two methods, which find checkerboards more reliably and accurately than existing algorithms. The evaluation of the measurement errors of distance cameras requires reference values that are considerably more accurate than those of the camera. This thesis presents a method that allows using a terrestrial laser scanner to acquire such data. However, before the reference data can be used for error analysis, it is necessary to transform the measurements of both sensors into a common coordinate system. For this purpose, an automatic method was developed that reliably calculates the unknown transformation based on a single calibration scene. The accuracy of this approach is confirmed in several experiments and clearly exceeds the accuracy of the competing state-of-the-art method. In addition, it is possible to generate reference distance images with this method, which can subsequently be used for the evaluation of distance cameras.
Time-of-flight (ToF) cameras have some error sources that are characteristic for this measurement principle. In order to be able to better compensate for these errors, it is first necessary to investigate the nature of these errors. This thesis also presents a comprehensive, standardized evaluation of the systematic errors of ToF cameras. For this purpose, six experiments are defined, which are subsequently carried out with eight ToF cameras. The evaluation of these experiments shows that the characteristic errors are differently pronounced with the investigated cameras, but nonetheless, can be observed even with the most recent models. Finally, a new calibration method for structured-light sensors is proposed, as well as an algorithm which refines parametric models of objects. The calibration method allows calculating a complete sensor model based on two or more images, even if the projected pattern is unknown. This is particularly necessary if the sensor does not use a regular projector, but emits the pattern with a diffractive optical element. This part of the thesis also presents a novel refinement method for parametric models that uses the developed camera model. The evaluation results show that the proposed method computes more accurate model parameters than state-of-the-art fitting algorithms.
|
|
|||
Dynamic Cardiac Chamber Imaging in C-arm Computed Tomography
Abstract:
Cardiovascular diseases, i.e. disorders pertaining to the heart and blood vessels, are a major cause of mortality in developed countries. Many of these disorders, such as stenoses and some cases of valvular dysfunction, can be diagnosed and treated minimally invasively in percutaneous, catheter-based interventions. Navigation of the catheters as well as assessment and guidance of these procedures rely on interventional X-ray projection imaging performed using an angiographic C-arm device.
From rotational angiography acquisitions, during which the C-arm rotates on a circular trajectory around the patient, volumetric images can be reconstructed similarly to conventional computed tomography (CT). A three-dimensional representation of the beating heart allowing for a comprehensive functional analysis during the intervention would be useful for clinicians. However, due to the slow rotational speed of the C-arm and the resulting inconsistency of the raw data, imaging dynamic objects is challenging. More precisely, only small, substantially undersampled subsets of the data, which correspond to the same cardiac phases, are approximately consistent. This causes severe undersampling artifacts in the images unless sophisticated reconstruction algorithms are employed. The goal of this thesis is to develop and evaluate such methods in order to improve the quality of dynamic imaging of cardiac chambers in C-arm CT.
One of the two approaches that is investigated in this work aims to mitigate raw data inconsistencies by compensating for the heart motion. It relies on a non-rigid motion estimate obtained from a preliminary reconstruction by means of image registration. We develop a pipeline for artifact reduction and denoising of these preliminary images that increases the robustness of motion estimation and thus removes artificial motion patterns in the final images. We also propose an iterative scheme alternating motion estimation and compensation combined with spatio-temporal smoothing to further improve both image quality and accuracy of motion estimation. Furthermore, we design an open-source tool for comparing motion-compensated reconstruction methods in terms of edge sharpness.
The other approach formulates reconstruction as an optimization problem and introduces prior models of the image appearance in order to find a suitable solution. In particular, sparsity-based regularization as suggested by compressed sensing theory proves beneficial. We investigate and compare temporal regularizers, which yield considerable image quality improvements. In a task-based evaluation concerned with functional analysis of the left ventricle, we study how spatio-temporally regularized reconstruction, carried out with a state-of-the-art proximal algorithm, degrades when the number of projection views is reduced. Finally, we devise a correction scheme that enables dynamic reconstruction of a volume of interest in order to reduce computational effort.
Compared to one another, the approaches exhibit differences with regard to the appearance of the reconstructed images in general and the cardiac motion in particular. A straightforward combination of the methods yields a trade-off between these properties. All in all, both the hybrid and the individual approaches are able to reconstruct dynamic cardiac images with good quality in light of the challenges of rotational angiography.
|
|
|||
Multi-Frame Super-Resolution Reconstruction with Applications to Medical Imaging
Abstract:
The optical resolution of a digital camera is one of its most crucial parameters with broad relevance for consumer electronics, surveillance systems, remote sensing, or medical imaging. However, resolution is physically limited by the optics and sen- sor characteristics. In addition, practical and economic reasons often stipulate the use of out-dated or low-cost hardware. Super-resolution is a class of retrospec- tive techniques that aims at high-resolution imagery by means of software. Multi- frame algorithms approach this task by fusing multiple low-resolution frames to reconstruct high-resolution images. This work covers novel super-resolution methods along with new applications in medical imaging. The first contribution of this thesis concerns computational methods to super- resolve image data of a single modality. The emphasis lies on motion-based algo- rithms that are derived from a Bayesian statistics perspective, where subpixel mo- tion of low-resolution frames is exploited to reconstruct a high-resolution image. More specifically, we introduce a confidence-aware Bayesian observation model to account for outliers in the image formation, e. g. invalid pixels. In addition, we propose an adaptive prior for sparse regularization to model natural images ap- propriately. We then develop a robust optimization algorithm for super-resolution using this model that features a fully automatic selection of latent hyperparam- eters. The proposed approach is capable of meeting the requirements regarding robustness of super-resolution in real-world systems including challenging con- ditions ranging from inaccurate motion estimation to space variant noise. For in- stance, in case of inaccurate motion estimation, the proposed method improves the peak-signal-to-noise ratio (PSNR) by 0.7 decibel (dB) over the state-of-the-art. The second contribution concerns super-resolution of multiple modalities in the area of hybrid imaging. We introduce novel multi-sensor super-resolution techniques and investigate two complementary problem statements. For super- resolution in the presence of a guidance modality, we introduce a reconstruction algorithm that exploits guidance data for motion estimation, feature driven adap- tive regularization, and outlier detection to reliably super-resolve a second modal- ity. For super-resolution in the absence of guidance data, we generalize this ap- proach to a reconstruction algorithm that jointly super-resolves multiple modali- ties. These multi-sensor methodologies boost accuracy and robustness compared to their single-sensor counterparts. The proposed techniques are widely appli- cable for resolution enhancement in a variety of multi-sensor vision applications including color-, multispectral- and range imaging. For instance in color imag- ing as a classical application, joint super-resolution of color channels improves the PSNR by 1.5 dB compared to conventional channel-wise processing. The third contribution transfers super-resolution to workflows in healthcare. As one use case in ophthalmology, we address retinal video imaging to gain spatio- temporal measurements on the human eye background non-invasively. In order to enhance the diagnostic usability of current digital cameras, we introduce a frame- work to gain high-resolution retinal images from low-resolution video data by exploiting natural eye movements. This framework enhances the mean sensitiv- ity of automatic blood vessel segmentation by 10 % when using super-resolution for image preprocessing. As a second application in image-guided surgery, we investigate hybrid range imaging. To overcome resolution limitations of current range sensor technologies, we propose multi-sensor super-resolution based on domain-specific system calibrations and employ high-resolution color images to steer range super-resolution. In ex-vivo experiments for minimally invasive and open surgery procedures using Time-of-Flight (ToF) sensors, this technique im- proves the reliability of surface and depth discontinuity measurements compared to raw range data by more than 24 % and 68 %, respectively. |
|
|||
Signal Processing for Interventional X-ray-based Coronary Angiography
Abstract:
Rotational angiography using C-arm scanners enables intra-operative 3-D imaging that has proved beneficial for diagnostic assessment and interventional guidance. Despite previous efforts, rotational angiography was not yet successfully established in clinical practice for coronary artery imaging but remains subject of intensive academic research. 3-D reconstruction of the coronary vasculature is impeded by severe lateral truncation of the thorax, as well as substantial intra-scan respiratory and cardiac motion. Reliable and fully automated solutions to all of the aforementioned problems are required to pave the way for clinical application of rotational angiography and, hence, sustainably change the state-of-care. |
|
|||
Interactive Analysis of Multispectral and Hyperspectral Image Data
Abstract:
A multispectral or hyperspectral sensor captures images of high spectral resolution by dividing the light spectrum into many narrow bands. With the advent of affordable and flexible sensors, the modality is constantly widening its range of applications. This necessitates novel tools that allow general and intuitive analysis of the image data. In this work, a software framework is presented that bundles interactive visualization techniques with powerful analysis capabilities and is accessible through efficient computation and an intuitive user interface. Towards this goal, several algorithmic solutions to open problems are presented in the fields of edge detection, clustering, supervised segmentation and visualization of hyperspectral images. |
|
|||
Iterative Reconstruction Methods to Reduce Respiratory Motion Artifacts in Cartesian Coronary MRI
Abstract:
Cardiovascular diseases and coronary artery disease (CAD) in particular are the leading cause of death in most developed countries worldwide. Although CAD progresses slowly over several years, it often remains unnoticed and may lead to myocardial infarction in a sudden event. For this reason, there is a strong clinical need for the development of non-invasive and radiation-free screening methods allowing for an early diagnosis of these diseases. In this context, magnetic resonance imaging (MRI) represents a promising imaging modality. However, the slow acquisition process and the consequent susceptibility to artifacts due to cardiac and respiratory motion are major challenges for coronary MRI and have so far hindered its routine application in clinical examinations. Commonly, respiratory motion is addressed during free-breathing acquisitions by gating the scan to a consistent respiratory phase in end-expiration with a navigator monitoring the patient's diaphragm. Acceptance rates below 40% lead to a prolonged total acquisition time that is also not predictable in advance. |
|
|||
Motion Correction and Signal Enhancement in Optical Coherence Tomography
Abstract:
Optical Coherence Tomography (OCT) is a non-invasive optical imaging modality with micron scale resolution and the ability to generate 2D and 3D images of the human retina. OCT has found widespread use in ophthalmology. However, motion artifacts induced by the scanning nature restrict the ability to have reliable quantification of OCT images. Furthermore, OCT suffers from speckle noise and signal quality issues. |
|
|||
Automatic Assessment of Prosody in Second Language Learning
Abstract:
The present thesis studies methods for automatically assessing the prosody of non-native speakers for the purpose of computer-assisted pronunciation training. We study the detection of word accent errors, and the general assessment of the appropriateness of a speaker’s rhythm. We propose a flexible, generic approach that is (a) very successful on these tasks, (b) competitive to other state-of-the-art result, and at the same time (c) flexible and easily adapted to new tasks. |
|
|||
MR-Based Attenuation Correction for PET/MR Hybrid Imaging
Abstract:
The recent and successful integration of positron emission tomography (PET) and magnetic resonance imaging (MRI) modalities in one device has gained wide attention. This new hybrid imaging modality now makes it possible to image the functional metabolism from PET in conjunction with MRI with its excellent soft tissue contrast. Besides providing specific anatomical detail, MRI also eliminates any ionizing radiation from eg. computed tomography (CT) examinations that is otherwise performed in standard PET/CT hybrid imaging systems. However, an unsolved problem is the question of how to correct for the PET attenuation in an PET/MR system. In this respect, the knowledge of the spatial distribution of linear attenuation coefficients (LAC) of the patient at the PET energy level of 511 keV is required. In standalone PET systems, transmission scans using radioactive sources were used for PET attenuation correction (AC) and if needed were scaled to the PET photon energy level. While in PET/CT systems, the CT information was scaled to PET energies for the same purpose. However, in PET/MR hybrid imaging systems, this approach is not feasible as MR and CT measure aspects of proton and electron densities respectively. Therefore alternate approaches to extract attenuation information have to be pursued. One such approach is to use MR information to estimate the distribution of attenuation coefficients within the imaging subject. This is done by using a simple limited class segmentation procedure to delineate air, soft tissue, fat and lung classes and subsequent assignment of their respective attenuation coefficients at PET energy of 511 keV. This way of generating attenuation maps (μ-maps) is however far from ideal as the most attenuating medium such as cortical bone is ignored. They are instead replaced by the attenuation coefficient of a soft tissue. While this approximation has been widely accepted for PET quantification in whole-body research, it has severe underestimation effects for brain studies. In this thesis, we propose an improved MR-based μ-map generation approach. We demonstrate that dedicated MR sequences such as ultrashort echo time sequences (UTE) are useful for the purpose of attenuation correction. From a multitude of MR images, we generate μ-maps that include cortical bone and contain continuous Hounsfield units (HU) akin to a patient CT. These are then compared against segmentation based approaches. The efficacy of continuous valued μ-maps towards PET quantification is analyzed against different μ-maps such as patient CT, segmented patient CT with bone and segmented patient CT without bone. Results indicate that the proposed MR-based μ-maps provide a less than 5% error in PET quantification than any segmentation based μ-maps for brain studies. |
|
|||
Cell Culture Monitoring with Novel Bright Field Miniature Microscopy Prototypes
Abstract:
Cell cultures are monitored to develop new drugs, to find efficient ways to produce vaccines and to perform toxicity tests. The cells are cultivated in an incubator and the monitoring steps such as the acquisition of images and the counting of cells are often done outside. As part of a research project, novel bright field miniature microscopy prototypes were developed. These prototypes were designed to work inside the incubator, and hence, they need to be very small. In this thesis, image processing methods for these systems (at different development stages) are presented. These methods made the systems usable for cell monitoring in an incubator. This is a main contribution of the thesis. Our analyses of the system and its components helped to improve the development of the systems. A calibration procedure and algorithms for adjusting the illumination and the focus position of these systems are introduced. Moreover, the proposed preprocessing steps such as illumination correction and contrast enhance ment improved the image quality. An image processing library and a cell monitoring software u sing the library were developed. An algorithm for counting cells in images of the prototype system was included in the image processing library. Features for viability determination were investigated and also included in the library. Another main contribution is related to all bright field microscopes. They have the following effect in common: Focusing of very thin (phase) objects differs from focusing of objects that are thicker and less transparent for light. This effect is investigated in detail, explained, and the calculation of different useful focus positions for phase objects is derived. The optical focus position can be used for applications such as phase retrieval. Slightly defocused cell images with a maximum in contrast at small details can be useful for applications such as cell segmentation or cell analysis. Strongly defocused cell images with a maximum in contrast for the cell borders can be used for applications such as cell detection. |
|
|||
Automatic Unstained Cell Detection in Bright Field Microscopy
Abstract:
Bright field microscopy is preferred over other microscopic imaging modalities whenever ease of implementation and minimization of expenditure are main concerns. This simplicity in hardware comes at the cost of image quality yielding images of low contrast. While staining can be employed to improve the contrast, it may complicate the experimental setup and cause undesired side effects on the cells. In this thesis, we tackle the problem of automatic cell detection in bright field images of unstained cells. The research was done in context of the interdisciplinary research project COSIR. COSIR aimed at developing a novel microscopic hardware having the following feature: the device can be placed in an incubator so that cells can be cultivated and observed in a controlled environment. In order to cope with design difficulties and manufacturing costs, the bright field technique was chosen for implementing the hardware. The contributions of this work are briefly outlined in the text which follows. An automatic cell detection pipeline was developed based on supervised learning. It employs Scale Invariant Feature Transform (SIFT) keypoints, random forests, and agglomerative hierarchical clustering (AHC) in order to reliably detect cells. A keypoint classifier is first used to classify keypoints into cell and background. An intensity profile is extracted between each two nearby cell keypoints and a profile classifier is then utilized to classify the two keypoints whether they belong to the same cell (inner profile) or to different cells (cross profile). This two-classifiers approach was used in the literature. The proposed method, however, compares to the state-of-the-art as follows: 1) It yields high detection accuracy (at least 14% improvement compared to baseline bright field methods) in a fully-automatic manner with short runtime on the low-contrast bright field images. 2) Adaptation of standard features in literature from being pixel-based to adopting a keypoint-based extraction scheme: this scheme is sparse, scale-invariant, orientation-invariant, and feature parameters can be tailored in a meaningful way based on a relevant keypoint scale and orientation. 3) The pipeline is highly invariant with respect to illumination artifacts, noise, scale and orientation changes. 4) The probabilistic output of the profile classifier is used as input for an AHC step which improves detection accuracy. A novel linkage method was proposed which incorporates the information of SIFT keypoints into the linkage. This method was proved to be combinatorial, and thus, it can be computed efficiently in a recursive manner. Due to the substantial difference in contrast and visual appearance between suspended and adherent cells, the above-mentioned pipeline attains higher accuracy in separate learning of suspended and adherent cells compared to joint learning. Separate learning refers to the situation when training and testing are done either only on suspended cells or only on adherent cells. On the other hand, joint learning refers to training the algorithm to detect cells in images which contain both suspended and adherent cells. Since these two types of cells coexist in cell cultures with shades of gray between the two terminal cases, it is of practical importance to improve joint learning accuracy. We showed that this can be achieved using two types of phasebased features: 1) physical light phase obtained by solving the transport of intensity equation, 2) monogenic local phase obtained from a low-passed axial derivative image. In addition to the supervised cell detection discussed so far, a cell detection approach based on unsupervised learning was proposed. Technically speaking, supervised learning was utilized in this approach as well. However, instead of training the profile classifier using manually-labeled ground truth, a self-labeling algorithm was proposed with which ground truth labels can be automatically generated from cells and keypoints in the input image itself. The algorithm learns from extreme cases and applies the learned model on the intermediate ones. SIFT keypoints were successfully employed for unsupervised structure-of-interest measurements in cell images such as mean structure size and dominant curvature direction. Based on these measurements, it was possible to define the notion of extreme cases in a way which is independent from image resolution and cell type. |
|
|||
Motion-Corrected Reconstruction in Cone-Beam Computed Tomography for Knees under Weight-Bearing Condition
Abstract:
Medical imaging plays an important role in diagnosis and grading of knee conditions such as osteoarthritis. In current clinical practice, 2-D radiography is regularly applied under weight-bearing conditions, which is known to improve diagnostic accuracy. However, 2-D images cannot fully cover the complexity of a knee joint, whereas current 3-D imaging modalities are inherently limited to a supine, unloaded patient position. Recently, cone-beam computed tomography (CBCT) scanners for 3-D weight-bearing imaging have been developed. Their specialized acquisition trajectory poses several challenges for image reconstruction. Patient motion caused by standing or squatting positions can substantially deteriorate image quality, such that the motion has to be corrected during reconstruction. Initial work on motion correction is based on fiducial markers, yet, the approach prolonged image acquisition and required a large amount of manual interaction. The goal of this thesis was to further develop innovative motion correction methods for weight-bearing imaging of knees. |
|
|||
Evaluation Methods for Stereopsis Performance
Abstract:
Stereopsis is one mechanism of visual depth perception, which gains 3D information from the displaced images of both eyes. Depth is encoded by disparity, the offset between the corresponding projections of one point in both retinas. Players in ball sports, who are adapted to highly competitive environments, can be assumed to develop improved performances in stereopsis, as they are required and thus trained constantly to rapidly and accurately estimate the distance of the ball. However, literature provides controversial results on the impact of stereopsis on sports such as baseball or soccer. The standard method to quantify stereopsis is to evaluate near static stereo acuity only, which denotes a subject’s minimum perceivable disparity from a near distance with stationary visual targets. These standard methods fail to reveal potential contributions of further components of stereopsis such as recognition speed, distance stereo acuity, and dynamic stereopsis, which were identified by literature to be important to describe the performance of stereopsis in sports. Therefore, this thesis contributes to the literature by introducing the Stereo Vision Performance (StereoViPer) test, which combines distance stereo acuity and response time analyses for static and dynamic stereopsis by using 3D stereo displays. |
|
|||
Hybrid RGB/Time-of-Flight Sensors in Minimally Invasive Surgery
Abstract:
Nowadays, minimally invasive surgery is an essential part of medical interventions. In a typical clinical workflow, procedures are planned preoperatively with 3-dimensional (3-D) computed tomography (CT) data and guided intraoperatively by 2-dimensional (2-D) video data. However, accurate preoperative data acquired for diagnose and operation planning is often not feasible to deliver valid information for orientation and decisions within the intervention due to issues like organ movements and deformations. Therefore, innovative interventional tools are required to aid the surgeon and improve safety and speed for minimally invasive procedures. Augmenting 2-D color information with 3-D range data allows to use an additional dimension for developing novel surgical assistance systems. Here, Time-of-Flight (ToF) is a promising low-cost and real-time capable technique that exploits reflected near-infrared light to estimate the radial distances of points in a dense manner. This thesis covers the entire implementation pipeline of this new technology into a clinical setup, starting from calibration to data preprocessing up to medical applications. |
|
|||
Region-of-Interest Imaging with C-arm Computed Tomography
Abstract:
C-arm based flat-detector computed tomography (FDCT) is a promising approach for neurovascular diagnosis and intervention since it facilitates proper analysis of surgical implants and intra-procedural guidance. In the majority of endovascular treatments, intra-procedural updates of the imaged object often are restricted to a small diagnostic region of interest (ROI). Such targeted ROI is often the region of intervention that contains device/vessel specific information such as stent expansion or arterial wall apposition. Following the principle of as low as reasonably achievable (ALARA), it is highly desirable to reduce unnecessary peripheral doses outside an ROI by using physical X-ray collimation, leading to substantial reduction of patient dose. However, such a technique gives rise to severely truncated projections from which conventional reconstruction algorithms generally yield images with strong truncation artifacts. |
|
|||
Design Considerations and Application Examples for Embedded Classification Systems
Abstract:
Wearable athlete support systems are a popular technology for performance enhancement in sports. The complex signal and data analysis on these systems is often tackled with pattern recognition techniques like classification. The implementation of classification algorithms on mobile hardware is called embedded classification. Thereby, technical challenges arise from the restricted computational power, battery capacity and size of such mobile systems. |
|
|||
Analysis of speech of people with Parkinson's disease
Abstract:
The analysis of speech of people with Parkinson's disease is an interesting and highly relevant topic that has been addressed in the research community during several years. |
|
|||
Reconstruction Techniques for Dynamic Radial MRI
Abstract:
Today, magnetic resonance imaging (MRI) is an essential clinical imaging modality and routinely used for orthopedic, neurological, cardiovascular, and oncological diagnosis. The relatively long scan times lead to two limitations in oncological MRI. Firstly, in dynamic contrast-enhanced MRI (DCE-MRI), spatial and temporal resolution have to be traded off against each other. Secondly, conventional acquisition techniques are highly susceptible to motion artifacts. As an example, in DCE-MRI of the liver, the imaging volume spans the whole abdomen and the scan must take place within a breath-hold to avoid respiratory motion. Dynamic imaging is achieved by performing multiple breath-hold scans before and after the injection of contrast agent. In practice, this requires patient cooperation, exact timing of the contrast agent injection, and limits the temporal resolution to about 10 seconds. This thesis addresses both challenges by combining a radial k-space sampling technique with advanced reconstruction algorithms for higher temporal resolution and improved respiratory motion management. |
|
|||
Motion Corrected Quantitative Imaging in Multimodal Emission Tomography
Abstract:
Nuclear medicine has been using single photon emission computed tomography (SPECT) for several decades in order to diagnose and enable the treatment of patients for various clinical applications. Traditionally, routine SPECT has been used for qualitative image interpretation based diagnosis which was not backed up with data on quantitative assessment of the encountered disease. However, latest research and development introduced a novel and yet unexplored feature of quantitative measurements to clinical practice. With the introduction of new quantitative reconstruction techniques many technological questions have to be answered. This thesis presents novel methods for enhancing quantitative iterative SPECT reconstruction by means of runtime and motion correction. These methods are evaluated with clinical practice and protocols in mind. |
|
|||
Dynamic Interventional Perfusion Imaging: Reconstruction Algorithms and Clinical Evaluation
Abstract:
Acute ischaemic stroke is a major cause for death and disabilities with increasing prevalence in aging societies. Novel interventional stroke treatment procedures have the potential to improve the clinical outcome of certain stroke-affected patients. Certainly, prompt diagnosis and treatment are required. Brain perfusion imaging with computed tomography (CT) or magnetic resonance imaging (MRI) is a routine method for stroke diagnosis. However, in the interventional room usually only CT imaging with flat detector C-arm systems is available, which do not support dynamic perfusion imaging yet. Enabling flat detector CT perfusion (FD-CTP) in clinical practice could support optimized stroke management. By stroke diagnosis in the interventional room precious time until the start of treatment could be saved. |
|
|||
High Performance Iterative X-Ray CT with Application in 3-D Mammography and Interventional C-arm Imaging Systems
Abstract:
Medical image reconstruction is a key component for a broad range of medical imaging technologies. For classical Computed Tomography systems the amount of measured signals per second increased exponentially over the last four decades, whereas the computational complexity of the majority of utilized algorithms has not changed significantly. |
|
|||
Rigid and Non-Rigid Surface Registration for Range Imaging Applications in Medicine
Abstract:
The introduction of low-cost range imaging technologies that are capable of acquiring the three-dimensional geometry of an observed scene in an accurate, dense, and dynamic manner holds great potential for manifold applications in health care. Over the past few years, the use of range imaging modalities has been proposed for guidance in computer-assisted procedures, monitoring of interventional workspaces for safe robot-human interaction and workflow analysis, touch-less user interaction in sterile environments, and for application in early diagnosis and elderly care, among others. This thesis is concerned with the application of range imaging technologies in computer-assisted and image-guided interventions, where the geometric alignment of range imaging data to a given reference shape – either also acquired with range imaging technology or extracted from tomographic planning data – poses a fundamental challenge. In particular, we propose methods for both rigid and non-rigid surface registration that are tailored to cope with the specific properties of range imaging data. |
|
|||
Respiratory Self-Navigation for Whole-Heart Coronary Magnetic Resonance Imaging
Abstract:
As the average life span of the world population increases, cardiovascular diseases firmly establish themselves as the most frequent cause of death in many of the developed countries. Coronary artery disease (CAD) is responsible for more than half of these cases and there is, hence, a strong need for a non-invasive and radiation-free test that could be reliably adopted for its assessment in clinical routine. Although coronary magnetic resonance imaging (MRI) has always been regarded with high expectations, it is still not considered for clinical assessment of CAD. This is mainly due to several limitations of current coronary MRI examinations. The complex anatomy of the coronary arteries requires extensive scout-scanning to precisely plan the actual data acquisition. The current speed limitations of the MRI scanners and the contribution of cardiac and respiratory motion do not allow the high resolution acquisitions to be performed within the fraction of a single heartbeat. Consequently, data acquisition must be split into multiple heartbeats and usually performed during free-breathing. At the same time, gating with respect to a consistent respiratory position is applied using an interleaved navigated scan which monitors the position of the subject's diaphragm. Major improvements in standard navigator-gated free-breathing coronary MRI have been achieved in recent years, but a number of important intrinsic limitations, such as the prolonged and unknown acquisition times, the non-linearity of the motion compensation, and the complexity of the examination setup have so far hindered the clinical usage of this technique. In contrast, a technique known as self-navigation, which performs motion detection and correction solely based on imaging data of the heart, promises a priori knowledge of the duration of the acquisition with improved accuracy of the motion compensation and requires minimal expertise for the planning of the examination. In this work, a novel acquisition and motion correction strategy for free-breathing selfnavigated whole-heart coronary MRA was introduced, analyzed and implemented to be entirely integrated in a clinical MR scanner. The proposed acquisition method consists of a novel interleaved 3D radial trajectory, mathematically constructed on the basis of a spiral phyllotaxis pattern, which intrinsically minimizes the eddy currents artifacts of the balanced steady state free-precessing acquisition, while ensuring a complete and uniform coverage of k-space. The self-navigated respiratory motion detection is performed on imaging readouts oriented along the superior-inferior axes and is based on a method for the isolation and automatic segmentation of the bright signal of the blood pool. Motion detection of the segmented blood pool is then performed using a cross-correlation technique. This fully automated respiratory selfnavigated method offers an easy and robust solution for coronary MR imaging that can also be integrated into a regular clinical routine examination. The technique was tested in volunteers, compared to the standard navigator-gating approach, and, for the first time to the author's knowledge, allowed self-navigation to be positively applied to a large patient study in an advanced clinical setting. |
|
|||
Boosting Methods for Automatic Segmentation of Focal Liver Lesions
Abstract:
Over the past decades, huge progress has been made in treatme nt of cancer, decreasing fatality rates despite a growing number of cases. Technical achievements had a big share in this development. With modern image acquisition techniques, most types of tum ors can be made visible. Automatic processing of these images to support diagnosis a nd therapy, on the other hand, is still very basic. Marking lesions for volume measurement s, intervention planning or tracking over time requires a lot of manual interaction, whi ch is both tedious and error prone. The work at hand therefore aims at providing tools for the aut omatic segmentation of liver lesions. A system is presented that receives a contras t enhanced CT image of the liver as input and, after several preprocessing steps, decides fo r each image voxel inside the liver whether it belongs to a tumor or not. That way, tumors are not o nly detected in the image but also precisely delineated in three dimensions. For the d ecision step, which is the main target of this thesis, we adopted the recently proposed Prob abilistic Boosting Tree. In an offline learning phase, this classifier is trained using a num ber of example images. After training, it can process new and previously unseen images. Such automatic segmentation systems are particularly valu able when it comes to moni- toring tumors of a patient over a longer period of time. There fore, we propose a method for learning a prior model to improve segmentation accuracy for such follow-up examinations. It is learned from a number of series of CT images, where each se ries contains images of one patient. Two different ways of incorporating the model i nto the segmentation system are investigated. When acquiring an image of a patient, the sy stem can use the model to calculate a patient specific lesion prior from images of the s ame patient acquired earlier and thus guide the segmentation in the current image. The validity of this approach is shown in a set of experiments on clinical images. When comparing the points of 90% sensitivity in these experiment s, incorporating the prior im- proved the precision of the segmentation from 82.7% to 91.9% . This corresponds to a reduction of the number of false positive voxels per true pos itive voxel by 57.8%. Finally, we address the issue of long processing times of cla ssification based segmen- tation systems. During training, the Probabilistic Boostin g Tree builds up a hierarchy of AdaBoost classifiers. In order to speed up classification duri ng application phase, we mod- ify this hierarchy so that simpler and thus faster AdaBoost cl assifiers are used in higher levels. To this end, we introduce a cost term into AdaBoost tra ining that trades off dis- criminative power and computational complexity during fea ture selection. That way the optimization process can be guided to build less complex cla ssifiers for higher levels of the tree and more complex and thus stronger ones for deeper level s. Results of an experimental evaluation on clinical images are presented, which show tha t this mechanism can reduce the overall cost during application phase by up to 76% withou t degrading classification ac- curacy. It is also shown that this mechanism could be used to o ptimize arbitrary secondary conditions during AdaBoost training. |
|
|||
Accelerated Non-contrast-enhanced Morphological and Functional Magnetic Resonance Angiography
Abstract:
Cardiovascular diseases such as stroke, stenosis, peripheral or renal artery disease require accurate angiographic visualization techniques both for diagnosis and treatment planning. Beside the morphological imaging, the in-vivo acquisition of blood flow information gained increasing clinical importance in recent years. Non-contrast-enhanced Magnetic Resonance Angiography (nceMRA) provides techniques for both fields. For morphological imaging, Time of Flight (TOF) and magnetization-prepared balanced Steady State Free Precession (mp-bSSFP) offer non-invasive, ionizing-radiation free and user independent alternatives to clinically established methods such as Digital Subtraction Angiography, Computed Tomography or Ultrasound. In the field of functional imaging, unique novel possibilities are given with three-directional velocity fields, acquired simultaneously to the morphological information using Phase Contrast Imaging (PCI). But the wider clinical use of nceMRA is still hampered by long acquisition times. Thus, accelerating nceMRA is a problem of high relevance and with great potential clinical impact. In this thesis, acceleration strategies based on k -space sampling below the Nyquist limit and adapted reconstruction techniques, combining parallel MRI (pMRI) methods with Compressed Sensing (CS), are developed for both types of nceMRA methods. This includes contributions to all relevant parts of the reconstruction algorithms, the sampling strategy, the regularization technique and the optimization method. For morphological imaging, a novel analytical pattern combining aspects of pMRI and CS, called the MICCS pattern, is proposed in combination with an adapted Split Bregman algorithm. This allows for a reduction in the acquisition time for peripheral TOF imaging of the entire lower vasculature from over 30 minutes to less than 8 minutes. Further acceleration is achieved for 3-D free-breathing renal angiography using mp-bSSFP, where the entire volume can be acquired in less than 1 minute instead of over 8 minutes. In addition, organ based evaluations including the vessel sharpness at important positions show the diagnostic usability and the increased accuracy over clinically established acceleration methods. For PCI, advances are achieved with a dedicated novel sampling strategy, called I-VT sampling, including interleaved variations for all dimensions. Furthermore, two novel regularization techniques for PCI are developed in this thesis. First, a novel temporally masked and weighted strategy focusing on enhanced temporal fidelity, referred to as TMW, is developed. This fully automatic approach uses dynamic and static vessel masks to guide the influence specifically to the static areas. Second, the low rank and sparse decomposition model, is extended to PCI, combined with adapted sparsity assumptions and the unconstrained Split Bregman algorithm. These methods are successfully applied to the carotid bifurcation, a region with a huge demand of significant acceleration as well high spatial and temporal accuracy of the flow values. But all algorithmic contributions exploit inherent properties of the acquisition technique, and thus can be applied for further applications. In summary, the main contribution of this thesis is significant acceleration of nceMRA achieved with novel sampling, regularization and optimization elements. |
|
|||
3-D Imaging of the Heart Chambers with C-arm CT
Abstract:
Nowadays, angiography is the gold standard for the visualization of the morphology of the cardiac vasculature and cardiac chambers in the interventional suite. Up to now, high resolution 2-D X-ray images are acquired with a C-arm system in standard views and the diagnosis of the cardiologist is based on the observations in the planar X-ray images. No dynamic analysis of the cardiac chambers can be performed in 3-D. In the last years, cardiac imaging in 3-D using a C-arm system becomes of more and more interest in the interventional catheter laboratory. Furthermore, the analysis of the 3-D motion would provide valuable information with respect to functional cardiac imaging. However, cardiac motion is a challenging problem in 3-D imaging, which leads to severe imaging artifacts in the 3-D image. Therefore, the main research goal of this thesis was the visualization and extraction of dynamic and functional parameters of the cardiac chambers in 3-D using an interventional angiographic C-arm system. In this thesis, two different approaches for cardiac chamber motion-compensated reconstruction have been developed and evaluated. The first technique addresses the visualization of the left ventricle. Therefore, a whole framework for left ventricular tomographic reconstruction and wall motion analysis has been developed. Dynamic surface models are generated from the 2-D X-ray images acquired during a short scan of a C-arm scanner using the 2-D bloodpool information. The acquisition time is about 5 s and the patients have normal sinus rhythm. Due to the acquisition time of about 5 s of the C-arm, no valuable retrospective ECG-gated reconstructions are possible. The dynamic surface LV model comprises a sparse motion vector field on the surface, which can be used for functional wall motion analysis. Furthermore, applying various interpolation schemes, dense motion vector fields can be generated for a tomographic motion-compensated reconstruction. In this thesis, linear interpolation methods and spline-based methods have been compared. The combination of the wall motion analysis and the motion-compensated reconstruction is of great value to the diagnostic of pathological regions in cardiac interventions. The second motion-compensated reconstruction approach uses volume-based motion estimation algorithms for the reconstruction of two - left atrium and left ventricle - to four heart chambers. A longer C-arm acquisition and contrast protocol allows for the generation of initial images at various heart phases. However, the initial image quality is not sufficient for motion estimation. Therefore, different pre-processing techniques, e.g., bilateral filtering or iterative reconstruction techniques, to improve the image quality were tested in combination with different motion estimation techniques. Overall, the results of this thesis highly demonstrate the feasibility of dynamic and functional cardiac chamber imaging using data from an interventional angiographic C-arm system for clinical applications. |
|
|||
Magnetic Resonance Imaging for Percutaneous Interventions
Abstract:
The fundamental motivation for all percutaneous interventions is to improve patient care by reducing the invasiveness of the procedure. An increasing number of percutaneous interventions from biopsies, targeted drug delivery to thermal ablations are performed under magnetic resonance (MR) guidance. Its excellent soft-tissue contrast and multiplanar imaging capabilities make MRI an attractive alternative to computed tomography or ultrasound for real-time image-guided needle placement, in particular for targets requiring a highly angulated approach and non-axial scan planes. MRI further provides the unique ability to monitor spatial temperature changes in real-time. The research efforts of this dissertation were focused on improving and simplifying the workflow of MR-guided percutaneous procedures by introducing novel imagebased methods without the need for any additional equipment. For safe and efficient MR-guided percutaneous needle placement, a set of methods was developed that allows the user to: 1) plan an entire procedure, 2) directly apply this plan to skin entry site localization without further imaging, and 3) place a needle under real-time MR guidance with automatic image plane alignment along a planned trajectory with preference to the principal patient axes. Methods for enhanced MR thermometry visualization and treatment monitoring were also developed to support an effective thermal treatment facilitating the ablation of tumor tissue without damaging adjacent healthy structures. To allow for an extensive in-vitro and in-vivo validation, the proposed methods for both needle guidance and MR thermometry were implemented in an integrated prototype. The clinical applicability was demonstrated for a wide range of MR-guided percutaneous interventions emphasizing the relevance and impact of the conducted research. |
|
|||
Automatic Classification of Cerebral Gliomas by Means of Quantitative Emission Tomography and Multimodal Imaging
Abstract:
Cerebral gliomas represent a common type of cancer of the human brain with many tumor grades which express a huge diversity in growth characteristics and have a highly varying malignancy. The optimal treatment for a cerebral glioma is only ensured if the underlying tumor grade is known. One very common grading scheme is the World Health Organization (WHO) Classification of tumors of the central nervous system, which differentiates four grades. The de facto standard of grading a glioma is based on bioptic samples which are obtained in invasive interventions. These interventions pose significant risks for the patients and add more time delays between an initial evidence of the tumor, usually found by X-ray computed tomography (CT) or magnetic resonance imaging (MRI) and the initiation of a treatment. On the other side, versatile imaging modalities like CT, MRI and from the field of nuclear medicine, positron emission tomography (PET) cover various aspects of the morphology and physiology of a tumor. The information gained from medical imaging thus can indicate the grade of a cerebral glioma without any invasive intervention. The multimodal imaging often results in a high complexity that makes if difficult to diagnose the malignancy solely based on the visual interpretation of medical images. In this thesis, we present approaches for an extensive pattern recognition pipeline for the grading of cerebral gliomas based on tomographic datasets from MRI, CT, and PET. More specifically, we use gadolinium contrast-enhanced T1-weighted MRI, T2-weighted fluid attenuated inversion recovery MRI, diffusion-weighted MRI, non contrast-enhanced low-dose X-ray CT, and dynamic (multiple acquired time frames) [18F]-Fluor-Ethyl-Tyrosine (FET) PET. Our setup includes image preprocessing, feature extraction and calculation, feature normalization, and finally fully automatic classification. We propose the imaging modalities and the classifiers which performed best for our patient population and show that inter-dataset normalization as a preprocessing step helps to improve the classification rate for cerebral gliomas. As the PET is acquired over a lengthy time period which can lead to substantial patient motion, we present a retrospective motion correction technique based on image registration, which improves the image quality of the PET data. The presented approaches underline that diagnostic statements can be gained from highly complex, multimodal image data in an automated fashion. We can differentiate not only low- and high-grade tumors, but also aid in distinguishing between the four WHO grades within some limitations. |
|
|||
C-arm Computed Tomography with Extended Axial Field-of-View
Abstract:
C-arm computed tomography (CT) is an innovative imaging technique in the interventional room that enables a C-arm system to generate 3D images like a CT system. Clinical re- ports demonstrate that this technique can help reduce treatment-related complications and may improve interventional efficacy and safety. However, currently, C-arm CT is only capable of imaging axially-short object, because it employs a single circular data acqui- sition geometry. This shortcoming can be a problem in some intraoperative cases when imaging a long object, e.g., the entire spine, is crucial. A new technique, C-arm CT for axially-long objects, namely extended-volume C-arm CT, has to be developed. This thesis aims at achieving this development. In particular, this thesis designs and analyzes data acquisition geometries as well as develops and implements reconstruction algorithms for extended-volume C-arm CT.
The thesis consists of three parts. In the first part, we studied three data acquisition geometries and invented two thereof. For these geometries, we investigated their feasi- bility on a C-arm system and analyzed their possibility for efficient, theoretically-exact and -stable (TES) reconstruction algorithms. We observed that the reverse helical trajec- tory is a good start for real data test and the novel ellipse-line-ellipse trajectory is a good candidate for efficient TES image reconstruction. In the second part, we developed and im- plemented geometry-specific reconstruction algorithms. For the reverse helix, we designed three Feldkamp-Davis-Kress (FDK)-type reconstruction methods. Among the three meth- ods, the Fusion-RFDK and Fusion-HFDK methods are preferred as they are more practical and produce acceptable images for extended-volume C-arm CT. For the ellipse-line-ellipse trajectory, we established an efficient TES reconstruction scheme, which makes proficient use of the geometry of this trajectory. In the third part, we conducted the first experiment for extended-volume C-arm CT on a laboratorial Artis zeego system. In this experiment, cone-beam data were reliably acquired using the reverse helical trajectory and 3D images were successfully reconstructed by the Fusion-RFDK method. The consistency among the- oretical understanding, simulation results and achieved image quality from a real system strongly demonstrate feasibility of extended-volume C-arm CT in the interventional room.
|
|
|||
Ansichtenauswahl für die 3-D-Rekonstruktion statischer Szenen
Abstract:
The problem of 3-D reconstruction is one of the main topics in computer vision. If certain imaging parameters can be modified to improve the 3-D reconstruction result, the question how to select this parameters belongs to a domain called active vision. The active parameters in our case are the focal length of the camera, which can be controlled by a zoom lens, and the pose, i. e. translation and rotation of the camera. The camera is mounted on a robot, so the position of the camera can be controlled. Usually, active vision for 3-D reconstruction means either scene exploration or most accurate estimation of the 3-D structure of an object. Of course, there are approaches trying to find a trade-off of the two aspects. This thesis focuses only on the aspect of high accurate estimates. For this purpose feature points are extracted from the images to estimate their 3-D coordinates. Here two different approaches are developed and evaluated: a geometric approach for stereo camera systems and a probabilistic approach. The geometric approach considers only stereo camera systems, i. e. systems which consist of exactly two cameras. The influence of the active parameters (translation, rotation and focal length) are evaluated and, if possible, analytically proven. The probabilistic approach determines the next best view, to increase the accuracy of the current estimate. Therefore it is necessary to describe the problem of 3-D reconstruction as a state estimation problem. The state estimation is solved by the extended Kalman filter. So, it is possible to improve the current state estimate of the 3-D coordinates by additional observations. This thesis derives a modification of the Kalman filter,which allows to reduce the calculation complexity drastically. For this modification only some simple assumptions are necessary, but it is discussed, why these assumptions are meaningful in the application of 3-D reconstruction. This modification is exact, i. e. there is no approximation required. A 3-D point to be reconstructed may be invisible, e. g. it is occluded by the object itself or its projection is outside of the field of view of the camera. Therefore, the next best view planning has to consider whether the point is visible or not from a certain view. We will show how the probability of visibility of a 3-D point can be calculated and further, how we can integrate the visibility issue into the closedform optimization criteria for the next best view planning. Another aspect of next best view planning is to move the camera to the desired position. Therefore, the question is, which position is reachable by the used robot device the camera is mounted on. In former publications, this aspect is either ignored or it is assumed the camera canmove only on a (part-)sphere around the object. But this thesis describes the reachable workspace by the Denavit-Hartenberg matrix. This allows to consider the complete workspace of the used robot in the next best view planning, without any unnecessary limitation to a (part-)sphere. |
|
|||
Automated Evaluation of Three Dimensional Ultrasonic Datasets
Abstract:
Non-destructive testing has become necessary to ensure the quality of materials and components either in-service or at the production stage. This requires the use of a rapid, robust and reliable testing technique. As a main testing technique, the ultrasound technology has unique abilities to assess the discontinuity location, size and shape. Such information play a vital role in the acceptance criteria which are based on safety and quality requirements of manufactured components. Consequently, an extensive usage of the ultrasound technique is perceived especially in the inspection of large scale composites manufactured in the aerospace industry. Signicant technical advances have contributed into optimizing the ultrasound acquisition techniques such as the sampling phased array technique. However, acquisition systems need to be complemented with an automated data analysis procedure to avoid the time consuming manual interpretation of all produced data. Such a complement would accelerate the inspection process and improve its reliability. The objective of this thesis is to propose an analysis chain dedicated to automatically process the 3D ultrasound volumes obtained using the sampling phased array technique. First, a detailed study of the speckle noise aecting the ultrasound data was conducted, as speckle reduces the quality of ultrasound data. Afterward, an analysis chain was developed, composed of a segmentation procedure followed by a classication procedure. The proposed segmentation methodology is adapted for ultrasound 3D data and has the objective to detect all potential defects inside the input volume. While the detection of defects is vital, one main diculty is the high amount of false alarms which are detected by the segmentation procedure. The correct distinction of false alarms is necessary to reduce the rejection ratio of safe parts. This has to be done without risking missing true defects. Therefore, there is a need for a powerful classier which can eciently distinguish true defects from false alarms. This is achieved using a specic classication approach based on data fusion theory. The chain was tested on several ultrasound volumetric measures of Carbon Fiber Reinforced Polymers components. Experimental results of the chain revealed high accuracy, reliability in detecting, characterizing and classifying defects. |
|
|||
Computerized Automatic Modeling of Medical Prostheses
Abstract:
In this thesis we study artificial intelligence methods, rule-based expert systems in particular, for the task of automatically designing customized medical prostheses. Here, the term design denotes the shaping or modeling of the prostheses and not their functional design. The challenge of the task at hand lies in designing prostheses that fit perfectly to the anatomy of the patient, and in many cases have to support additional functionality. Hence, each prosthesis has to be unique. Therefore, medical prostheses are usually designed starting with a template of the patient’s anatomy, e. g. acquired using CT data or scanned and digitized molds. In this thesis we assume that the template data is given as a triangle mesh in 3-D. To address the challenge of automatically designing medical prostheses, we develop an expert system framework consisting of an expert system shell, a knowledge base and a feature detection unit. The framework is integrated into an existing modeling software. In the following, we denote the complete system as Expert System for Automatic Modeling (ESAM). The architecture of ESAM is generic and can be used for all kinds of design tasks. The specialization for the application in mind can be achieved by providing the necessary design rules and by adjusting the feature detection algorithms. Our expert system specializes in monitoring and controlling a CAD software. Thus, it defines the parameters of the CAD tools, executes the tools and monitors the results by constantly analyzing the current shape. As part of the expert system we develop a knowledge representation language to structure and store the expert knowledge. The language is easy to understand, flexible and can be extended as required. The knowledge base is created in interaction with experts of the field. In addition, we study methods to extend, improve and maintain the knowledge base. We evaluate two methods for rule creation and rule critic. On the one hand, we apply genetic programming as a rule learning technique. On the other hand, we use a heuristic method based on data generated by ESAM. For the latter, we develop a tool that generates statistics about rule performance, rule relationships and user interaction. The thereby gained knowledge is integrated into the knowledge base, resulting in a higher performance of ESAM, e.g. the completion rate increased by about 30 %. We apply two types of feature detection methods for the detection of surface features on the given templates. The first method analyzes the surface of the given template for peaks, depressions, ridges and combinations of these generic features. The generality of the detected features allows a simple adjustment to different anatomies. The second method uses registration in order to copy features from a labeled template to an unlabeled one. As a first step, it applies clustering techniques to identify a suitable set of representative templates. In the second step, these templates are labeled by a domain expert. Subsequently, the labels can be transferred based on the result of an ICP registration. Our experiments show that the second approach results in a higher quality of the detected features, e. g. mean deviation is reduced from approximately 3.8 mm by about 30 % to approximately 2.6 mm. ESAM is verified using the example of customized in-the-ear hearing aid design. An industry partner provides the domain knowledge necessary to create the knowledge base as well as the possibility to verify the system in a real production environment. In order to ensure the quality of the designed and manufactured in-the-ear hearing aids, the system is verified while running in a semi-automatic mode. The semi-automatic mode allows a modeling expert to monitor and correct the system if necessary. During the verification and practical usage of ESAM thousands of customized in-the-ear hearing aid shells are manufactured. It could be shown that compared to the manual approach the design consistency improves by about 10% and the design time is reduced by about 30 %. The overall acceptance rate of an expert system rule is 76 %. In addition to that, ESAM provides a framework, which guides the modeler through the complex design process, and thereby reduces the amount of design errors and avoids unnecessary process steps. As a consequence of these positive evaluation results our industry partner continues to apply ESAM on its production floor. |
|
|||
Physics-based and Statistical Features for Image Forensics
Abstract:
The objective of blind image forensics is to determine whether an image is authentic or captured with a particular device. In contrast to other security-related fields, like watermarking, it is assumed that no supporting pattern has been embedded into the image. Thus, the only available cues for blind image forensics are either a) based on inconsistencies in expected (general) scene and camera properties or b) artifacts from particular image processing operations that were performed as part of the manipulation.
In this work, we focus on the detection of image manipulations. The contributions can be grouped in two categories: techniques that exploit the statistics of forgery artifacts and methods that identify inconsistencies in high-level scene information. The two categories complement each other. The statistical approaches can be applied to the majority of digital images in batch processing. If a particular, single image should be investigated, high-level features can be used for a detailed manual investigation. Besides providing an additional, complementary testing step for an image, high-level features are also more resilient to intentional disguise of the manipulation operation.
Hence, the first part of this thesis focuses on methods for the detection of statistical artifacts introduced by the manipulation process. We propose improvements to the detection of so-called copy-move forgeries. We also develop a unified, extensively evaluated pipeline for copy-move forgery detection. To benchmark different detection features within this pipeline, we create a novel framework for the controlled creation of semi-realistic forgeries. Furthermore, if the image under investigation is stored in the JPEG format, we develop an effective scheme to expose inconsistencies in the JPEG coefficients.
The second part of this work aims at the verification of scene properties. Within this class of methods, we propose a preprocessing approach to assess the consistency of the illumination conditions in the scene. This algorithm makes existing work applicable to a broader range of images. The main contribution in this part is a demonstration of how illuminant color estimation can be exploited as a forensic cue. In the course of developing this method, we extensively study color constancy algorithms, which is the classical research field for estimating the color of the illumination. In this context, we investigate extensions of classical color constancy algorithms to the new field of non-uniform illumination. As part of this analysis, we create a new, highly accurate ground truth dataset and propose a new algorithm for multi-illuminant estimation based on conditional random fields.
|
|
|||
Interactive Approaches to Video Lecture Assessment
Abstract:
A growing number of universities and other educational institutions record videos of regularly scheduled classes and lectures to provide students with additional resources for their study. However, the video alone is not necessarily the same than a carefully prepared educational video. The main issue is that they are typically not post-processed in an editorial sense. That is, the videos often contain longer periods of silence or inactivity, unnecessary repetitions, spontaneous interaction with students, or even corrections of prior false statements or mistakes. Furthermore, there is often no summary or table of contents of the video, unlike with educational videos that supplement a certain curriculum and are well scripted and edited. Thus, the plain recording of a lecture is a good start but far from a good e-learning resource. |
|
|||
Efficient and Trainable Detection and Classification of Radio Signals
Abstract:
The strong need for robust and efficient radio signal detection and classification can be found both in civil and non-civil applications. Most of the current state-of-the-art systems are based on a separated version of detection and classification. Within this work, a new integrated approach is presented. The proposed system combines the insights into communications intelligence, cognitive radio and modern pattern recognition ideas. |
|
|||
Image Processing for Fluoroscopy Guided Atrial Fibrillation Ablation Procedures
Abstract:
Atrial fibrillation is a common heart arrhythmia and is associated with an increased risk of stroke. The current state-of-the-art treatment option is the minimally invasive catheter ablation. During such procedures, the four pulmonary veins attached to the left atrium are electrically isolated. New methods to guide these procedures are presented in this work. |
|
|||
Diffusion Tensor Imaging Analysis of the Visual Pathway with Application to Glaucoma
Abstract:
Glaucoma is an optic neuropathy affecting the entire visual system. The world-wide prevalence of glaucoma is estimated to be 60.5 million people. The visual disorder caused by glaucoma can reach complete blindness if untreated. Various treatment approaches exist that can largely prevent the visual disability and limit the vision loss due to glaucoma if the disease is diagnosed in its early phases. Nevertheless, the slow progression of the disease along with the lack of clear symptoms results in the late identification of glaucoma. Moreover, the pathophysiology of glaucoma, and its biological foundation and factors are not yet fully determined or understood. Therefore, novel directions are essential for improving the diagnostic flow and the understanding of the glaucoma mechanism. |
|
|||
Machine Learning Methods in Computed Tomography Image Analysis
Abstract:
Lymph nodes have high clinical relevance because they are often affected by cancer, and also play an important role in all kinds of infections and inflammations in general. Lymph nodes are commonly examined using computed tomography (CT). |
|
|||
Robust Single Shot Structured Light
Abstract:
In this thesis a new robust approach for Single-Shot Structured Light 3D scanning is developed. As the name implies, this measurement principle requires only one image of an object, illuminated with a suitable pattern, to reconstruct the shape and distance of the object. This technique has several advantages. It can be used to record 3D video with a moving sensor or of a moving scene. Since the required hardware is very simple, the sensor can also be easily miniaturized. Single-Shot Structured Light, thus, has the potential to be the basis of a versatile and inexpensive 3D scanner. |
|
|||
Model-Constrained Non-Rigid Registration in Medicine
Abstract:
The aim of image registration is to compute a mapping from one image's frame of reference to another's, such that both images are well aligned. Even when the mapping is assumed to be rigid (only rotation and translation) this can be a quite challenging task to accomplish between different image modalities. Noise and other imaging artifacts like bias fields in magnetic resonance (MR) imaging or streak artifacts in computed tomography (CT) can pose additional problems. In non-rigid image registration these problems are further compounded by the additional degrees of freedom in the transform. |
|
|||
Time-of-Flight - A New Modality for Radiotherapy
Abstract:
In this work, one of the first approaches utilizing so-called Time-of-Flight cameras for medical applications is presented. Using Time-of-Flight cameras it is feasible to acquire a 3-D model in real-time with a single sensor. Several systems for managing motion within radiotherapy are presented. There are five major contributions in this work: A method to verify internal tumor movement with an external respiratory signal on-line, the application of a novel technology to medical image processing and the introduction of three novel systems, one to measure respiratory motion and two other to position patients. The algorithm to correlate external and internal motion is an image-based synchronization procedure that automatically labels pre-treatment fluoroscopic images with corresponding 4-D CT phases. It is designed as an optimization process and finds the optimal mapping between both sequences by maximizing the image similarity between the corresponding pairs while preserving a temporal coherency. It is both evaluated at synthetic and patient data and an average of 93% correctly labeled frames could be achieved. The Time-of-Flight based respiratory motion system enables the simultaneously measurement of different regions. We evaluate the system using a novel body phantom. Tests showed, that the system signal and the ground truth signal of the phantom have a reliable correlation of more than 80% for amplitudes greater 5 mm. The correlation of both systems is independent (always more than 80%) of the respiratory frequency. Furthermore, the measured signals were compared with a well-established external gating system, the Anzai belt. These experiments were performed on human persons. We could show a correlation of about 88% of our system and the Anzai system. The first positioning system is able to position a C-arm like device with respect to the patient. Therefore, a Time-of-Flight camera acquires the whole body of the patient and segments it into meaningful anatomical regions, like head, thorax, abdomen, legs. The system computes 3-D bounding boxes of the anatomical regions and computes the isocenter of the boxes. Using this information, the C-arm system can automatically position itself and perform a scan. The system is evaluated using a body phantom and an accuracy within the patient table accuracy of 1 cm could be shown. The second system deals with surface-based positioning of a patient with respect to a priorly acquired surface of the same patient. Such systems are necessary, e.g. in radiotherapy or multi-modal imaging. The method uses an Iterative-Closest-Point algorithm, tailored to Time-of-Flight cameras. It is evaluated using a body phantom and obtains an overall accuracy of 0.74 mm +/ 0.37 mm for translations in all three room directions within 10 mm. |
|
|||
Quantitative Computed Tomography
Abstract:
Computed Tomography (CT) is a wide-spread medical imaging modality. Traditional CT yields information on a patient's anatomy in form of slice images or volume data. Hounsfield Units (HU) are used to quantify the imaged tissue properties. Due to the polychromatic nature of X-rays in CT, the HU values for a specific tissue depend on its density and composition but also on CT system parameters and settings and the surrounding materials. The main objective of Quantitative CT (QCT) is measuring characteristic physical tissue or material properties quantitatively. These characteristics can, for instance, be density of contrast agents or local X-ray attenuation. Quantitative measurements enable specific medical applications such as perfusion diagnostic or attenuation correction for Positron Emission Tomography (PET). |
|
|||
Interventional Perfusion Imaging Using C-arm Computed Tomography: Algorithms and Clinical Evaluation
Abstract:
A stroke is a medical emergency which requires immediate diagnosis and treatment. For several years, image-based stroke diagnosis has been assisted using perfusion computed tomography (CT) and perfusion magnetic resonance imaging (MRI). A contrast agent bolus is injected and time-resolved imaging, at typically one frame per second, is used to measure the contrast agent flow. However, these two modalities are not accessible in the interventional suite where catheter-guided stroke treatment actually takes place. Thus, interventional perfusion imaging, which could lead to optimized stroke management, is currently not available. |
|
|||
Patient-Specific Cerebral Vessel Segmentation with Application in Hemodynamic Simulation
Abstract:
Cerebral 3-D rotational angiography has become the state-of-the-art imaging modality in modern angio suites for diagnosis and treatment planning of cerebrovascular diseases, e. g. intracranial aneurysms. Among other reasons, it is believed that the incidence of aneurysms is due to the local prevalent hemodynamic pattern. To study such a hemodynamic behavior, the 3-D vessel geometry has to be extracted from 3-D DSA data. Since 3-D DSA data may be influenced by beam hardening, inhomogeneous contrast agent distribution, patient movement or the applied reconstruction kernel, this thesis describes a novel vessel segmentation framework seamlessly combining 2-D and 3-D vessel information to overcome the aforementioned factors of influence. The main purpose of this framework is to validate 3-D segmentation results based on 2-D information and to increase the accuracy of 3-D vessel geometries by incorporating additional 2-D vessel information into the 3-D segmentation process. Three major algorithmic contributions are given within this framework: (1) a classification-based summation algorithm of 2-D DSA series such that 2-D vessel segmentation becomes feasible, (2) a 3-D ellipsoid-based vessel segmentation method which allows for local adaptations driven by 2-D vessel segmentations and (3) a mesh size evaluation study investigating the influence of different mesh type elements and resolutions w. r. t. hemodynamic simulation results. Moreover, this work is chamfered by a simulation study which evaluates the impact of different vessel geometries on the simulation result. The vessel geometries are computed by different segmentation techniques working on the same patient dataset. The evaluation of each framework component revealed high accuracy and algorithmic stability to be applied in a clinical environment. |
|
|||
Statistical Intensity Prior Models with Applications in Multimodal Image Registration
Abstract:
Deriving algorithms that automatically align images being acquired from different sources (multimodal image registration) is a fundamental problem that is of importance to several active research areas in image analysis, computer vision, and medical imaging. In particular, the accurate estimation of deformations in multimodal image data perpetually engages researchers while playing an essential role in several clinical applications that are designed to improve available healthcare. Since the field of medical image analysis has been rapidly growing for the past two decades, the abundance of clinical information that is available to medical experts inspires more automatic processing of medical images. |
|
|||
Motion Estimation and Compensation for Interventional Cardiovascular Image Reconstruction
Abstract:
The minimal invasive interventional treatment of cardiac diseases is of high importance in the modern society. Catheter-based procedures are becoming increasingly complex and novel tools for planning and guiding the interventions are required. In recent years intraprocedural 3-D imaging has found its way into the clinics. Based on 2-D X-ray images from C-arm systems a 3-D image with high spatial resolution can be computed. Cardiac vessels are small and moving fast and thus pose a problem to standard reconstruction algorithms. In this thesis, the issues of existing approaches are investigated and novel algorithms are developed that mitigate todays problems in terms of image quality, runtime and assumptions on the cardiac motion. One major contribution is the development of an optimized ECG-gated reconstruction algorithm compensating for non-periodic motion. A cost function inspired from iterative reconstruction algorithms is used to assess the reconstruction quality of an analytic reconstruction algorithm. This key concept is utilized to derive a motion estimation algorithm. The efficient and compact problem formulation allows for the first time the application of ECG-gating in case of non-periodic motion patterns which cannot be reconstructed with previous methods. This significant finding is incorporated into a novel B-spline based motion estimation algorithm which can cope with flexible 3-D motions over time and uses all the projection data. It again takes advantage of an analytic reconstruction algorithm to arrive at a highly efficient, well parallelizable and stable algorithm. In the evaluation it is shown that the developed algorithms allow the reconstruction of clinically challenging cases at high image quality in under 10 minutes. Therefore it combines the desirable properties of reconstruction algorithms in the interventional environment which no other algorithm provided before. |
|
|||
Optimizing Application Driven Multimodality Spatio-Temporal Emission Imaging
Abstract:
Single Photon Emission Computed Tomography (SPECT) is a widely used nuclear medicine imaging technique with many applications in diagnosis and therapy. With the introduction of hybrid imaging systems, integrating a SPECT and a Computed Tomography (CT) system in one gantry, diagnostic accuracy of nuclear procedures has been improved. Current imaging protocols in clinical practice take between 15 and 45 minutes and Filtered Backprojection (FBP) is still widely used to reconstruct nuclear images. Routine clinical diagnosis is based on reconstructed image intensities which do not represent the true absolute activity concentration of the target object, due to various effects inherent to SPECT image formation. |
|
|||
Novel Techniques for Spatial Orientation in Natural Orifice Translumenal Endoscopic Surgery (NOTES)
Abstract:
With a novel approach abdominal surgery can be performed without skin incisions. The natural orifices provide the entry point with a following incision in stomach, colon, vagina or bladder. “Natural Orifice Translumenal Endoscopic Surgery” (NOTES) is assumed to offer significant benefits to patients such as less pain and reduced traumata as well as reduced collateral damages, faster recovery, and better cosmesis. Particular improvement can be reached even for obesity and burn injury patients and children. But the potential advantages of this new technology can only be exploited through safe and standardized operation methods. Several barriers identified for the clinical practicability in flexible intra-abdominal endoscopy can be solved with computer-assisted surgical systems. In order to assist the surgeon during the intervention and to enhance his visual perception, some of these systems are able to additionally provide 3-D information of the intervention site, for others 3-D information is even mandatory. |
|
|||
Evaluation moderner Hardwarearchitekturen zur schnellen CT Rekonstruktion aus Kegelstrahlprojektionen
Abstract:
|
|
|||
Normalization of Magnetic Resonance Images and its Application to the Diagnosis of the Scoliotic Spine
Abstract:
Due to its excellent soft tissue contrast and novel innovative acquisition sequences, Magnetic Resonance Imaging has become one of the most popular imaging modalities in health care. However, associated acquisition artifacts can significantly reduce image quality. Consequently, this imperfections can disturb the assessment of the acquired images. In the worst case, they may even lead to false decisions by the physician. Moreover, they can negatively influence an automatic processing of the data, e.g., image segmentation or registration. The most commonly observed artifacts are intensity inhomogeneities and a missing sequence-dependent general intensity scale. |
|
|||
Probabilistic Modeling for Segmentation in Magnetic Resonance Images of the Human Brain
Abstract:
This thesis deals with the fully automatic generation of semantic annotations for medical imaging data by means of medical image segmentation and labeling. In particular, we focus on the segmentation of the human brain and related structures from magnetic resonance imaging (MRI) data. We present three novel probabilistic methods from the field of database-guided knowledge-based medical image segmentation. We apply each of our methods to one of three MRI segmentation scenarios: 1) 3-D MRI brain tissue classification and intensity non-uniformity correction, 2) pediatric brain cancer segmentation in multi-spectral 3-D MRI, and 3) 3-D MRI anatomical brain structure segmentation. All the newly developed methods make use of domain knowledge encoded by probabilistic boosting-trees (PBT), which is a recent machine learning technique. For all the methods we present uniform probabilistic formalisms that group the methods into the broader context of probabilistic modeling for the purpose of image segmentation. We show by comparison with other methods from the literature that in all the scenarios our newly developed algorithms in most cases give more accurate results and have a lower computational cost. Evaluation on publicly available benchmarking data sets ensures reliable comparability of our results to those of other current and future methods. We also document the participation of one of our methods in the ongoing online caudate segmentation challenge (www.cause07.org), where we rank among the top five methods for this particular segmentation scenario. |
|
|||
Application of Pattern Recognition Methods in Biomechanics (external: University of Calgary, Canada)
Abstract:
Biomechanical studies often attempt to identify differences between groups. Several scientific methods are available for identifying such differences. Traditional methods often focus on the analysis of single variables and do not take into account high-dimensional dependencies. Moreover, the analysis procedures are often biased by the expectations of the researcher. Pattern recognition based methods provide data driven analysis often conducted simultaneously in multiple dimensions. Such algorithms have recently been applied for biomechanical analysis tasks. However, the use of pattern recognition algorithms is still not well understood in the biomechanical community. Therefore, the contribution of this thesis was to add further understanding of tools from pattern recognition to biomechanical tasks of group differentiation. |
|
|||
Praktikable Ansätze für mehrsprachige und nicht-muttersprachliche Spracherkennung
Abstract:
|
|
|||
Adaptive Filtering for Noise Reduction in X-Ray Computed Tomography
Abstract:
The projection data measured in computed tomography (CT) and, consequently, the slices reconstructed from these data are noisy. This thesis investigates methods for structure preserving noise reduction in reconstructed CT datasets. The goal is to improve the signal-to-noise ratio without increasing the radiation dose or loss of spatial resolution. Due to the close relation between noise and radiation dose, this improvement at the same time opens up a possibility for dose reduction. Two different original approaches, which automatically adapt themselves to the non-stationary and non-isotropic noise in CT, were developed, implemented and evaluated. The first part of the thesis concentrates on wavelet based noise reduction methods. They are based on the idea of using reconstructions from two disjoint subsets of projections as input to the noise reduction algorithm. Correlation analysis between the wavelet coefficients of the input images and noise estimation in the wavelet domain is used for differentiating between structures and noise. In the second part, an original approach based on noise propagation through the reconstruction algorithm is presented. A new method for estimating the local noise variance and correlation in the image from the noise estimates of the measured data is proposed. Based on the additional information about the image noise, an adaptive bilateral filter is introduced. The proposed methods are all evaluated with respect to the obtained noise reduction rate, but also in terms of their ability to preserve structures. A contrast dependent resolution analysis is performed to estimate the dose reduction potential of the different methods. The achieved noise reduction of about 60% can lead to dose reduction rates between 40% to 80%, depending on the clinical task. |
|
|||
One-to-one Edge Based Registration and Segmentation Based Validations in Hybrid Imaging
Abstract:
During the past decade, image registration has become an essential tool for medical treatment in clinics, by finding the spatial mapping between two images, observing the changes of anatomical structure and merging the information from different modalities. On the other hand, the matching of appropriately selected features is becoming more and more important for the further improvement of registration methods, as well as for the qualitative validation of registration. The purpose of this thesis is to solve the following two problems: How to integrate feature detection into a non-rigid registration framework, so that a high quality spatial mapping can be achieved? How to systematically measure the quality of multi-modal registration by automatically segmenting the corresponding features? For the first problem, we develop a general approach based on the Mumford-Shah model for simultaneously detecting the edge features of two images and jointly estimating a consistent set of transformations to match them. The entire variational model is realized in a multi-scale framework of the finite element approximation. The optimization process is guided by an EM type algorithm and an adaptive generalized gradient flow to guarantee a fast and smooth relaxation. This one-to-one edge matching is a general registration method, which has been successfully adapted to solve image registration problems in several medical applications, for example mapping inter-subject MR data, or alignment of retina images from different cameras. For the second problem, we propose a new method validating the hybrid functional and morphological image fusion, especially for the SPECT/CT modality. It focuses on measuring the deviation between the corresponding anatomical structures. Two kinds of anatomical structures are investigated as validationmarkers: (1) the hot spot in a functional image and its counterpart in the morphological image (2) the kidneys in both modalities. A series of special methods are developed to segment these structures in both modalities with minimum user interaction. Accuracy of the validation methods have been confirmed by experiments with real clinical data-sets. The inaccuracies of hot spot based validation for neck regions are reported to be 0.7189±0.6298mm in X-direction, 0.9250 ± 0.4535mm in Y -direction and 0.9544 ± 0.6981mm in Z-direction. While the inaccuracies of kidneys based validation for abdomen regions are 1.3979±0.8401 mm in X-direction, 1.9992 ± 1.3920 mm in Y -direction and 2.7823 ± 2.0672 mm in Z-direction. In the end, we also discuss a new interpolation based method to effectively improve the SPECT/CT fusion and present preliminary results. |
|
|||
Statistical Medical Image Registration with Applications in Epilepsy Diagnosis and Shape-Based Segmentation
Abstract:
The advances in scanner technologies over the past years and a growing number of modalities in medical imaging result in an increased amount of patient data. The physicians are faced with an overwhelming amount of information when comparing different scans. Therefore, automatic image processing algorithms are necessary to facilitate everyday clinical workflows. The present work focuses on automatic, statistical image registration approaches and applications in epilepsy diagnosis and shape-based segmentation. Registration algorithms based on image intensity statistics are currently state-of-the-art to automatically compute an alignment between multi-modal images. The parameters, however, are sensitive to the input data. In the present work, we study the mutual influences of these parameters on the intensity statistics and present datadriven estimation schemes to optimize them with respect to the input images. This is necessary to register large sets of images both accurately and reliably. The presented evaluation results, which are based on a database with an established gold standard, confirm that individually optimized parameters lead to improved results compared to standard settings found in literature. Besides spatial accuracy, the reduction of the computation time for the registration is equally important. In this thesis, we present an approach to reduce the search space for the optimization of a rigid registration transform by a nonlinear projection scheme, which is closely related to the concept of marginalization of random variables. Within each projection, a disjoint subset of the transform parameters is optimized with greatly reduced computational complexity. With a good choice of the projection geometry, the search space can be separated into disjoint subsets. In the case of rigid 3-D image registrations, the nonlinear projection onto a cylinder surface allows for an optimization of the rotation around the cylinder axis and a translation along its direction without the need for a reprojection. Sub-volume registration problems are supported by fitting the projection geometry into the overlap domain of the input images. The required objective functions are constrained by systems of linear inequalities and solved by means of constrained, nonlinear optimization techniques. A statistical framework is proposed to measure the accuracy of the registration algorithms with respect to manual segmentation results. The aforementioned concepts of the data-driven density estimators are adopted for the estimation of spatial densities of the segmented labels in order to model the observer reliability. The accuracy of the spatial registration transform is measured between the estimated distributions of the segmented labels in both input images using the Kullback-Leibler divergence. The proposed algorithms are evaluated by a registration of a database of morphological and functional images with an established gold standard based on fiducial marker implants. Applications are presented for the subtraction of single emission computed tomography scans for epilepsy diagnosis, where the intensity distributions are estimated for both the task of the registration and the normalization of the images. Finally, the registration is utilized for shape-based image segmentation to establish a model for the variability within a collective of segmented training shapes. |
|
|||
Cardiac C-Arm Computed Tomography: Motion Estimation and Dynamic Reconstruction
Abstract:
Generating three dimensional images of the heart during interventional procedures is a significant challenge. In addition to real-time fluoroscopy, angiographic C-arm systems can also be used to generate 3-D/4-D CT images on the same system. One protocol for cardiac Computed Tomography (CT) uses electrocardiogram (ECG) triggered multi-sweep scans. A 3-D volume of the heart at a particular cardiac phase is reconstructed by the Feldkamp, Davis and Kress (FDK) algorithm using projection images with retrospective ECG gating. In this thesis we introduce a unified framework for heart motion estimation and dynamic cone-beam reconstruction using motion corrections. Furthermore, theoretical considerations about dynamic filtered backprojection (FBP) as well as dynamic algebraic reconstruction techniques (ART) are presented, discussed and evaluated. Dynamic CT reconstruction allows to improve temporal resolution and image quality using image processing. It is limited by C-arm device hardware like rotation speed. The benefits of motion correction are: (1) increased temporal and spatial resolution by removing cardiac motion which may still exists in the ECG-gated data sets, and (2) increased signal-to-noise ratio (SNR) by using more projection data than is used in standard ECG gated methods. Three signal enhancing reconstruction methods are introduced that make use of all of the acquired projection data to generate a time resolved 3-D reconstruction. The first averages all motion corrected backprojections; the second and third perform a weighted averaging according to: (1) intensity variations and (2) temporal distance to a time resolved and motion corrected reference FDK reconstruction. In a study seven methods are compared: non-gated FDK, ECG-gated FDK, ECG-gated and motion corrected FDK, the three signal enhancing approaches, and temporally aligned and averaged ECG-gated FDK reconstructions. The quality measures used for comparison are spatial resolution and SNR. Additionally new dynamic algebraic reconstruction techniques (ART) are introduced, compared to dynamic Filtered Backprojection (FBP) methods and evaluated. In ART we model the objects motion either using a dynamic projector model or a dynamic grid of the object, defining the spatial sampling of the reconstructed density values. Both methods are compared to each other as well as to dynamic FBP. Spatial and temporal interpolation issues in dynamic ART and FBP and the computational complexity of the algorithms are addressed. The subject-specific motion estimation is performed using standard non-rigid 3-D/3-D and novel 3-D/2-D registration methods that have been specifically developed for the cardiac C-arm CT reconstruction environment. In addition theoretical considerations about fast shift-invariant filtered backprojection methods in dependency of an affine, rayaffine and non-rigid motion model are presented. Evaluation is performed using phantom data and several animal models. We show that data driven and subject-specific motion estimation combined with motion correction can decrease motion-related blurring substantially. Furthermore, SNR can be increased by up to 70% while maintaining spatial resolution at the same level as it is provided by the ECGgated FDK. The presented framework provides excellent image quality for cardiac C-arm CT. The thesis contributes to an improved image quality in cardiac C-arm CT and provides several methods for dynamic FBP and ART reconstruction. |
|
|||
Speech of children with Cleft Lip and Palate: Automatic Assessment
Abstract:
This work investigates the use of automatic speech processing techniques for the automatic assessment of children’s speech disorders. The target group were children with cleft lip and palate (CLP). The speech processing techniques are applied to evaluate the children’s speech intelligibility and their articulation. Another goal of this work is to visualize the kind and degree of the pathology in the children’s speech. Tracking of the children’s therapy progress is also within the reach of the system. |
|
|||
Image Reconstruction from Fan-Beam and Cone-Beam Projections
Abstract:
This thesis addresses the problem of reconstructing static objects in 2D and 3D transmission computed tomography (CT). After reviewing the classical CT reconstruction theory, we discuss and thoroughly evaluate various novel reconstruction methods, two of which are original. Our first original approach is for 2D CT reconstruction from full-scan fan-beam data, i.e., for 2D imaging in the geometry of diagnostic medical CT scanners. Compared to conventional methods, our approach is computationally more efficient and also yields results with an overall reduction of image noise at comparable spatial resolution, as demonstrated in detailed evaluations based on simulated fan-beam data and on data collected with a Siemens Somatom CT scanner. Part two of this thesis discusses the problem of 3D reconstruction in the short-scan circular cone-beam (CB) geometry, i.e., the geometry of medical C-arm systems. We first present a detailed comparative evaluation of innovative methods recently suggested in the literature for reconstruction in this geometry and of the approach applied on many existing systems. This evaluation involves various quantitative and qualitative figures-of-merit to assess image quality. We then derive an original short-scan CB reconstruction method that is based on a novel, theoretically-exact factorization of the 3D reconstruction problem into a set of independent 2D inversion problems, each of which is solved iteratively and yields the object density on a single plane. In contrast to the state-of-the-art methods discussed earlier in this thesis, our factorization approach does not involve any geometric approximations during its derivation and enforces all reconstructed values to be positive; it thus provides quantitatively very accurate results and effectively reduces CB artifacts in the reconstructions, as illustrated in the numerical evaluations based on computer-simulated CB data and also real CB data acquired with a Siemens Axiom Artis C-arm system. |
|
|||
Accurate Cone-Beam Image Reconstruction in C-Arm Computed Tomography
Abstract:
The goal of this thesis is the robust implementation of an accurate cone-beam image reconstruction algorithm such that it is able to process real C-arm data from a circle-plus-arc trajectory. This trajectory is complete and especially well suited for C-arm systems, since it can be performed purely by rotating the C-arm around the patient without the need to move the patient table. We observed two major challenges: i) non-ideal acquisition geometry and ii) data truncation. To account for deviations from the ideal description of the data acquisition geometry, we developed a novel calibration procedure for the circle-plus-arc trajectory. For the second problem, we developed two novel truncation correction methods that approximately but effectively handle data truncation problems. For image reconstruction, we adapted the accurate M-line algorithm. In particular, we applied a novel and numerically stable technique to compute the view dependent derivative with respect to the source trajectory parameter and we developed an efficient way to compute the PI-line backprojection intervals via a polygonal weighting mask. We have chosen the M-line algorithm, since it does not presume an ideal description of the data acquisition geometry. We acquired projection data of a physical phantom of a human thorax on a medical C-arm scanner. Reconstructed images exhibit strong cone-beam artifacts along the bones of the spine when applying the conventional Feldkamp algorithm. These results are compared to those obtained with our implementation of the M-line algorithm. As our ultimate goal, we demonstrate that cone-beam artifacts can be completely eliminated by applying the M-line algorithm to a Tuy complete set of data. |