Ruwan Tennakoon - Academia.edu (original) (raw)
Papers by Ruwan Tennakoon
IEEE Transactions on Pattern Analysis and Machine Intelligence
2019 IEEE International Conference on Image Processing (ICIP)
Classification of 3D shapes into physically meaningful categories is one of the most important ta... more Classification of 3D shapes into physically meaningful categories is one of the most important tasks in understanding the immediate environment. Methods that leverage the recent advancements in deep learning have shown to outperform the traditional approaches. However, performances of those methods have only been analyzed with relatively clean data. Three-dimensional measurement sets (point clouds) produced by 3D scanners are rarely that accurate and often contain noise, outliers or missing points. This paper presents an extensive analysis of the robustness of the state-of-the-art neural network algorithms to realistic data inaccuracies. Our experiments show that the existence of these inaccuracies can significantly affect the performance of the deep learning-based algorithms.
2017 IEEE International Conference on Mechatronics (ICM)
This paper presents a track-before-detect labeled multi-Bernoulli filter tailored for industrial ... more This paper presents a track-before-detect labeled multi-Bernoulli filter tailored for industrial mobile platform safety applications. We derive two application specific separable likelihood functions that capture the geometric shape and colour information of the human targets who are wearing a high visibility vest. These likelihoods are then used in a labeled multi-Bernoulli filter with a novel two step Bayesian update. Preliminary simulation results evaluated using several video sequences show that the proposed solution can successfully track human workers wearing a luminous yellow colour vest in an industrial environment.
arXiv: Computer Vision and Pattern Recognition, 2020
Stereo vision generally involves the computation of pixel correspondences and estimation of dispa... more Stereo vision generally involves the computation of pixel correspondences and estimation of disparities between rectified image pairs. In many applications, including simultaneous localization and mapping (SLAM) and 3D object detection, the disparities are primarily needed to calculate depth values and the accuracy of depth estimation is often more compelling than disparity estimation. The accuracy of disparity estimation, however, does not directly translate to the accuracy of depth estimation, especially for faraway objects. In the context of learning-based stereo systems, this is largely due to biases imposed by the choices of the disparity-based loss function and the training data. Consequently, the learning algorithms often produce unreliable depth estimates of foreground objects, particularly at large distances~($>50$m). To resolve this issue, we first analyze the effect of those biases and then propose a pair of novel depth-based loss functions for foreground and backgroun...
2019 IEEE Intelligent Transportation Systems Conference (ITSC), 2019
With the continuous process of urbanization in many cities around the globe, construction works a... more With the continuous process of urbanization in many cities around the globe, construction works are increasingly interrupting traffic flow in urban roads. At construction sites, usually traffic control personnel are employed to instruct the traffic flow at the times obstruction. Employing manual labour flagmen for traffic control is costly and can potentially expose the workers to safety hazards. Alternative attempts such as using remotely controlled signs or timers are not ideal; the former still requires human input and the latter does not optimize traffic control based on traffic density. To tackle this problem, we propose a fully automated traffic controller solution that uses visual sensors and machine vision technology. This system can replace the human traffic controllers during lane closures on the event of a construction. The implementation of this system eliminates the requirement for conventional traffic controllers, reducing the overall health and safety risks to workers...
2018 21st International Conference on Information Fusion (FUSION), 2018
In this paper we present a novel non-Bayesian filtering method for tracking multiple objects with... more In this paper we present a novel non-Bayesian filtering method for tracking multiple objects with a particular application in time-lapse cell microscopic video sequence. In our method the heat-map of the frame sequence is extracted and represented as a pseudo-probability hypothesis density of the image. The pseudo-probability hypothesis density is used as measurements and fused with a prior Poisson random finite set density. We employed Cauchy-Schwarz divergence for information fusion. The presented algorithm was tested on a publicly available cell microscopic video sequence.
IEEE Transactions on Signal Processing, 2021
This paper presents a new solution for multi-target tracking over a network of sensors with limit... more This paper presents a new solution for multi-target tracking over a network of sensors with limited spatial coverage. The proposed solution is based on the centralized data fusion architecture. The main contribution of the paper is the introduction of a new track-to-track fusion approach in which the posterior distributions of multi-target states, reported by various sensor nodes, are fused in a way that the redundant information are combined and the rest complement each other. The proposed solution is formulated within the labeled random finite set framework in which the fused posterior incorporates all the state and label information provided by multiple sensor nodes. The performance of the proposed method is evaluated via simulation experiments that involve challenging tracking scenarios. The proposed method is implemented using sequential Monte Carlo method and the results confirm its effectiveness.
This paper proposes a novel method in order to detect the presence and obtain voxel-level segment... more This paper proposes a novel method in order to detect the presence and obtain voxel-level segmentation for three fluid lesion types (IRF/SRF/PED) in OCT images provided by the ReTOUCH challenge. The method is based on a deep neural network consisting of encoding and de-coding blocks connected with skip-connections which was trained using a combined cost function comprising of cross-entropy, dice and adversarial loss terms. The segmentation results on a held-out validation set shows that the network architecture and the loss functions used has resulted in improved retinal fluid segmentation.
IEEE Transactions on Intelligent Transportation Systems, 2021
In autonomous vehicles, depth information for the environment surrounding the vehicle is commonly... more In autonomous vehicles, depth information for the environment surrounding the vehicle is commonly extracted using time-of-flight (ToF) sensors such as LiDARs and RADARs. Those sensors have some limitations that may potentially degrade the quality and utility of the depth information to a substantial extent. An alternative solution is depth estimation from stereo pairs. However, stereo matching and depth estimation often fails at ill-posed regions including areas with repetitive patterns or textureless surfaces which are commonly found on planar surfaces. This paper focuses on designing an efficient framework for stereo depth estimation, using deep learning technique, that is robust against the mentioned ill-posed regions. With the observation that disparities of all pixels belonging to planar areas (scene plane) viewed by two rectified stereo images can be described using affine transformations, our proposed method predicts pixel-wise affine transformation parameters based on the de...
Proceedings of the 15th International Conference on Informatics in Control, Automation and Robotics
Condition monitoring of storm-water pipe systems are carried-out regularly using semi-automated p... more Condition monitoring of storm-water pipe systems are carried-out regularly using semi-automated processors. Semi-automated inspection is time consuming, expensive and produces varying and relatively unreliable results due to operators fatigue and novicity. This paper propose an innovative method to automate the stormwater pipe inspection and condition assessment process which employs a computer vision algorithm based on deep-neural network architecture to classify the defect types automatically. With the proposed method, the operator only needs to guide the robot through each pipe and no longer needs to be an expert. The results obtained on a CCTV video dataset of storm-water pipes shows that the deep neural network architectures trained with data augmentation and transfer learning is capable of achieving high accuracies in identifying the defect types.
2016 Sixth International Conference on Image Processing Theory, Tools and Applications (IPTA)
This paper approaches the problem of geometric multi-model fitting as a data segmentation problem... more This paper approaches the problem of geometric multi-model fitting as a data segmentation problem which is solved by a sequence of sampling, model selection and clustering steps. We propose a sampling method that significantly facilitates solving the segmentation problem using the Normalized cut. The sampler is a novel application of Markov-Chain-Monte-Carlo (MCMC) method to sample from a distribution in the parameter space that is obtained by modifying the Least kth Order Statistics cost function. To sample from this distribution effectively, our proposed Markov Chain includes novel long and short jumps to ensure exploration and exploitation of all structures. It also includes fast local optimization steps to target all, even fairly small, putative structures. This leads to a clustering solution through which final model parameters for each segment are obtained. The method competes favorably with the state-of-the-art both in terms of computation power and segmentation accuracy.
ArXiv, 2017
Data preprocessing is a fundamental part of any machine learning application and frequently the m... more Data preprocessing is a fundamental part of any machine learning application and frequently the most time-consuming aspect when developing a machine learning solution. Preprocessing for deep learning is characterized by pipelines that lazily load data and perform data transformation, augmentation, batching and logging. Many of these functions are common across applications but require different arrangements for training, testing or inference. Here we introduce a novel software framework named nuts-flow/ml that encapsulates common preprocessing operations as components, which can be flexibly arranged to rapidly construct efficient preprocessing pipelines for deep learning.
ArXiv, 2020
In this paper, we present a robust spherical harmonics approach for the classification of point c... more In this paper, we present a robust spherical harmonics approach for the classification of point cloud-based objects. Spherical harmonics have been used for classification over the years, with several frameworks existing in the literature. These approaches use variety of spherical harmonics based descriptors to classify objects. We first investigated these frameworks robustness against data augmentation, such as outliers and noise, as it has not been studied before. Then we propose a spherical convolution neural network framework for robust object classification. The proposed framework uses the voxel grid of concentric spheres to learn features over the unit ball. Our proposed model learn features that are less sensitive to data augmentation due to the selected sampling strategy and the designed convolution operation. We tested our proposed model against several types of data augmentation, such as noise and outliers. Our results show that the proposed model outperforms the state of a...
Predicting the presence of a disease in volumetric images is an essential task in medical imaging... more Predicting the presence of a disease in volumetric images is an essential task in medical imaging. The use of state-of-the-art techniques like deep convolutional neural networks (CNN) for such tasks is challenging due to limited supervised training data and high memory usage. This paper presents a weakly supervised solution that can be used in learning deep CNN features for volumetric image classification. In the proposed method, we use extreme value theory to model the feature distribution of the images without a pathology and use it to identify positive instances in an image that contains pathology. The experimental results show that the proposed method can learn classifiers that have similar performance to a fully supervised method and have significantly better performance in comparison with methods that use fixed number of instances from a positive image.
ArXiv, 2020
This paper outlines connections between Monotone Boolean Functions, LP-Type problems and the Maxi... more This paper outlines connections between Monotone Boolean Functions, LP-Type problems and the Maximum Consensus Problem. The latter refers to a particular type of robust fitting characterisation, popular in Computer Vision (MaxCon). Indeed, this is our main motivation but we believe the results of the study of these connections are more widely applicable to LP-type problems (at least 'thresholded versions', as we describe), and perhaps even more widely. We illustrate, with examples from Computer Vision, how the resulting perspectives suggest new algorithms. Indeed, we focus, in the experimental part, on how the Influence (a property of Boolean Functions that takes on a special form if the function is Monotone) can guide a search for the MaxCon solution.
ArXiv, 2020
In this paper we propose a real-time and robust solution to large-scale multiple rotation averagi... more In this paper we propose a real-time and robust solution to large-scale multiple rotation averaging. Until recently, Multiple rotation averaging problem had been solved using conventional iterative optimization algorithms. Such methods employed robust cost functions that were chosen based on assumptions made about the sensor noise and outlier distribution. In practice, these assumptions do not always fit real datasets very well. A recent work showed that the noise distribution could be learnt using a graph neural network. This solution required a second network for outlier detection and removal as the averaging network was sensitive to a poor initialization. In this paper we propose a single-stage graph neural network that can robustly perform rotation averaging in the presence of noise and outliers. Our method uses all observations, suppressing outliers effects through the use of weighted averaging and an attention mechanism within the network design. The result is a network that i...
Scientific Reports
Prostate cancer (PCa) is the second most frequent type of cancer found in men worldwide, with aro... more Prostate cancer (PCa) is the second most frequent type of cancer found in men worldwide, with around one in nine men being diagnosed with PCa within their lifetime. PCa often shows no symptoms in its early stages and its diagnosis techniques are either invasive, resource intensive, or has low efficacy, making widespread early detection onerous. Inspired by the recent success of deep convolutional neural networks (CNN) in computer aided detection (CADe), we propose a new CNN based framework for incidental detection of clinically significant prostate cancer (csPCa) in patients who had a CT scan of the abdomen/pelvis for other reasons. While CT is generally considered insufficient to diagnose PCa due to its inferior soft tissue characterisation, our evaluations on a relatively large dataset consisting of 139 clinically significant PCa patients and 432 controls show that the proposed deep neural network pipeline can detect csPCa patients at a level that is suitable for incidental detect...
The task of learning from point cloud data is always challenging due to the often occurrence of n... more The task of learning from point cloud data is always challenging due to the often occurrence of noise and outliers in the data. Such data inaccuracies can significantly influence the performance of state of the art deep learning networks and their ability to classify or segment objects. While there are some robust deep learning approaches, they are computationally too expensive for real-time applications. This paper proposes a deep learning solution that includes a novel robust pooling layer which greatly enhances network robustness and perform significantly faster than state-of-the-art approaches. The proposed pooling layer looks for data a mode/cluster using two methods, RANSAC, and histogram, as clusters are indicative of models. We tested the pooling layer into frameworks such as Point-based and graph-based neural networks, and the tests showed enhanced robustness as compared to robust state-of-the-art methods.
This paper approaches the problem of geometric multi-model fitting as a data segmentation problem... more This paper approaches the problem of geometric multi-model fitting as a data segmentation problem which is solved by a sequence of sampling, model selection and clustering steps. We propose a sampling method that significantly facilitates solving the segmentation problem using the Normalized cut. The sampler is a novel application of Markov-Chain-Monte-Carlo (MCMC) method to sample from a distribution in the parameter space that is obtained by modifying the Least kth Order Statistics cost function. To sample from this distribution effectively, our proposed Markov Chain includes novel long and short jumps to ensure exploration and exploitation of all structures. It also includes fast local optimization steps to target all, even fairly small, putative structures. This leads to a clustering solution through which final model parameters for each segment are obtained. The method competes favorably with the state-of-the-art both in terms of computation power and segmentation accuracy.
Classification of 3D shapes into physically meaningful categories is one of the most important ta... more Classification of 3D shapes into physically meaningful categories is one of the most important tasks in understanding the immediate environment. Methods that leverage the recent advancements in deep learning have shown to outperform the traditional approaches. However, performances of those methods have only been analyzed with relatively clean data. Three-dimensional measurement sets (point clouds) produced by 3D scanners are rarely that accurate and often contain noise, outliers or missing points. This paper presents an extensive analysis of the robustness of the state-of-the-art neural network algorithms to realistic data inaccuracies. Our experiments show that the existence of these inaccuracies can significantly affect the performance of the deep learning-based algorithms.
IEEE Transactions on Pattern Analysis and Machine Intelligence
2019 IEEE International Conference on Image Processing (ICIP)
Classification of 3D shapes into physically meaningful categories is one of the most important ta... more Classification of 3D shapes into physically meaningful categories is one of the most important tasks in understanding the immediate environment. Methods that leverage the recent advancements in deep learning have shown to outperform the traditional approaches. However, performances of those methods have only been analyzed with relatively clean data. Three-dimensional measurement sets (point clouds) produced by 3D scanners are rarely that accurate and often contain noise, outliers or missing points. This paper presents an extensive analysis of the robustness of the state-of-the-art neural network algorithms to realistic data inaccuracies. Our experiments show that the existence of these inaccuracies can significantly affect the performance of the deep learning-based algorithms.
2017 IEEE International Conference on Mechatronics (ICM)
This paper presents a track-before-detect labeled multi-Bernoulli filter tailored for industrial ... more This paper presents a track-before-detect labeled multi-Bernoulli filter tailored for industrial mobile platform safety applications. We derive two application specific separable likelihood functions that capture the geometric shape and colour information of the human targets who are wearing a high visibility vest. These likelihoods are then used in a labeled multi-Bernoulli filter with a novel two step Bayesian update. Preliminary simulation results evaluated using several video sequences show that the proposed solution can successfully track human workers wearing a luminous yellow colour vest in an industrial environment.
arXiv: Computer Vision and Pattern Recognition, 2020
Stereo vision generally involves the computation of pixel correspondences and estimation of dispa... more Stereo vision generally involves the computation of pixel correspondences and estimation of disparities between rectified image pairs. In many applications, including simultaneous localization and mapping (SLAM) and 3D object detection, the disparities are primarily needed to calculate depth values and the accuracy of depth estimation is often more compelling than disparity estimation. The accuracy of disparity estimation, however, does not directly translate to the accuracy of depth estimation, especially for faraway objects. In the context of learning-based stereo systems, this is largely due to biases imposed by the choices of the disparity-based loss function and the training data. Consequently, the learning algorithms often produce unreliable depth estimates of foreground objects, particularly at large distances~($>50$m). To resolve this issue, we first analyze the effect of those biases and then propose a pair of novel depth-based loss functions for foreground and backgroun...
2019 IEEE Intelligent Transportation Systems Conference (ITSC), 2019
With the continuous process of urbanization in many cities around the globe, construction works a... more With the continuous process of urbanization in many cities around the globe, construction works are increasingly interrupting traffic flow in urban roads. At construction sites, usually traffic control personnel are employed to instruct the traffic flow at the times obstruction. Employing manual labour flagmen for traffic control is costly and can potentially expose the workers to safety hazards. Alternative attempts such as using remotely controlled signs or timers are not ideal; the former still requires human input and the latter does not optimize traffic control based on traffic density. To tackle this problem, we propose a fully automated traffic controller solution that uses visual sensors and machine vision technology. This system can replace the human traffic controllers during lane closures on the event of a construction. The implementation of this system eliminates the requirement for conventional traffic controllers, reducing the overall health and safety risks to workers...
2018 21st International Conference on Information Fusion (FUSION), 2018
In this paper we present a novel non-Bayesian filtering method for tracking multiple objects with... more In this paper we present a novel non-Bayesian filtering method for tracking multiple objects with a particular application in time-lapse cell microscopic video sequence. In our method the heat-map of the frame sequence is extracted and represented as a pseudo-probability hypothesis density of the image. The pseudo-probability hypothesis density is used as measurements and fused with a prior Poisson random finite set density. We employed Cauchy-Schwarz divergence for information fusion. The presented algorithm was tested on a publicly available cell microscopic video sequence.
IEEE Transactions on Signal Processing, 2021
This paper presents a new solution for multi-target tracking over a network of sensors with limit... more This paper presents a new solution for multi-target tracking over a network of sensors with limited spatial coverage. The proposed solution is based on the centralized data fusion architecture. The main contribution of the paper is the introduction of a new track-to-track fusion approach in which the posterior distributions of multi-target states, reported by various sensor nodes, are fused in a way that the redundant information are combined and the rest complement each other. The proposed solution is formulated within the labeled random finite set framework in which the fused posterior incorporates all the state and label information provided by multiple sensor nodes. The performance of the proposed method is evaluated via simulation experiments that involve challenging tracking scenarios. The proposed method is implemented using sequential Monte Carlo method and the results confirm its effectiveness.
This paper proposes a novel method in order to detect the presence and obtain voxel-level segment... more This paper proposes a novel method in order to detect the presence and obtain voxel-level segmentation for three fluid lesion types (IRF/SRF/PED) in OCT images provided by the ReTOUCH challenge. The method is based on a deep neural network consisting of encoding and de-coding blocks connected with skip-connections which was trained using a combined cost function comprising of cross-entropy, dice and adversarial loss terms. The segmentation results on a held-out validation set shows that the network architecture and the loss functions used has resulted in improved retinal fluid segmentation.
IEEE Transactions on Intelligent Transportation Systems, 2021
In autonomous vehicles, depth information for the environment surrounding the vehicle is commonly... more In autonomous vehicles, depth information for the environment surrounding the vehicle is commonly extracted using time-of-flight (ToF) sensors such as LiDARs and RADARs. Those sensors have some limitations that may potentially degrade the quality and utility of the depth information to a substantial extent. An alternative solution is depth estimation from stereo pairs. However, stereo matching and depth estimation often fails at ill-posed regions including areas with repetitive patterns or textureless surfaces which are commonly found on planar surfaces. This paper focuses on designing an efficient framework for stereo depth estimation, using deep learning technique, that is robust against the mentioned ill-posed regions. With the observation that disparities of all pixels belonging to planar areas (scene plane) viewed by two rectified stereo images can be described using affine transformations, our proposed method predicts pixel-wise affine transformation parameters based on the de...
Proceedings of the 15th International Conference on Informatics in Control, Automation and Robotics
Condition monitoring of storm-water pipe systems are carried-out regularly using semi-automated p... more Condition monitoring of storm-water pipe systems are carried-out regularly using semi-automated processors. Semi-automated inspection is time consuming, expensive and produces varying and relatively unreliable results due to operators fatigue and novicity. This paper propose an innovative method to automate the stormwater pipe inspection and condition assessment process which employs a computer vision algorithm based on deep-neural network architecture to classify the defect types automatically. With the proposed method, the operator only needs to guide the robot through each pipe and no longer needs to be an expert. The results obtained on a CCTV video dataset of storm-water pipes shows that the deep neural network architectures trained with data augmentation and transfer learning is capable of achieving high accuracies in identifying the defect types.
2016 Sixth International Conference on Image Processing Theory, Tools and Applications (IPTA)
This paper approaches the problem of geometric multi-model fitting as a data segmentation problem... more This paper approaches the problem of geometric multi-model fitting as a data segmentation problem which is solved by a sequence of sampling, model selection and clustering steps. We propose a sampling method that significantly facilitates solving the segmentation problem using the Normalized cut. The sampler is a novel application of Markov-Chain-Monte-Carlo (MCMC) method to sample from a distribution in the parameter space that is obtained by modifying the Least kth Order Statistics cost function. To sample from this distribution effectively, our proposed Markov Chain includes novel long and short jumps to ensure exploration and exploitation of all structures. It also includes fast local optimization steps to target all, even fairly small, putative structures. This leads to a clustering solution through which final model parameters for each segment are obtained. The method competes favorably with the state-of-the-art both in terms of computation power and segmentation accuracy.
ArXiv, 2017
Data preprocessing is a fundamental part of any machine learning application and frequently the m... more Data preprocessing is a fundamental part of any machine learning application and frequently the most time-consuming aspect when developing a machine learning solution. Preprocessing for deep learning is characterized by pipelines that lazily load data and perform data transformation, augmentation, batching and logging. Many of these functions are common across applications but require different arrangements for training, testing or inference. Here we introduce a novel software framework named nuts-flow/ml that encapsulates common preprocessing operations as components, which can be flexibly arranged to rapidly construct efficient preprocessing pipelines for deep learning.
ArXiv, 2020
In this paper, we present a robust spherical harmonics approach for the classification of point c... more In this paper, we present a robust spherical harmonics approach for the classification of point cloud-based objects. Spherical harmonics have been used for classification over the years, with several frameworks existing in the literature. These approaches use variety of spherical harmonics based descriptors to classify objects. We first investigated these frameworks robustness against data augmentation, such as outliers and noise, as it has not been studied before. Then we propose a spherical convolution neural network framework for robust object classification. The proposed framework uses the voxel grid of concentric spheres to learn features over the unit ball. Our proposed model learn features that are less sensitive to data augmentation due to the selected sampling strategy and the designed convolution operation. We tested our proposed model against several types of data augmentation, such as noise and outliers. Our results show that the proposed model outperforms the state of a...
Predicting the presence of a disease in volumetric images is an essential task in medical imaging... more Predicting the presence of a disease in volumetric images is an essential task in medical imaging. The use of state-of-the-art techniques like deep convolutional neural networks (CNN) for such tasks is challenging due to limited supervised training data and high memory usage. This paper presents a weakly supervised solution that can be used in learning deep CNN features for volumetric image classification. In the proposed method, we use extreme value theory to model the feature distribution of the images without a pathology and use it to identify positive instances in an image that contains pathology. The experimental results show that the proposed method can learn classifiers that have similar performance to a fully supervised method and have significantly better performance in comparison with methods that use fixed number of instances from a positive image.
ArXiv, 2020
This paper outlines connections between Monotone Boolean Functions, LP-Type problems and the Maxi... more This paper outlines connections between Monotone Boolean Functions, LP-Type problems and the Maximum Consensus Problem. The latter refers to a particular type of robust fitting characterisation, popular in Computer Vision (MaxCon). Indeed, this is our main motivation but we believe the results of the study of these connections are more widely applicable to LP-type problems (at least 'thresholded versions', as we describe), and perhaps even more widely. We illustrate, with examples from Computer Vision, how the resulting perspectives suggest new algorithms. Indeed, we focus, in the experimental part, on how the Influence (a property of Boolean Functions that takes on a special form if the function is Monotone) can guide a search for the MaxCon solution.
ArXiv, 2020
In this paper we propose a real-time and robust solution to large-scale multiple rotation averagi... more In this paper we propose a real-time and robust solution to large-scale multiple rotation averaging. Until recently, Multiple rotation averaging problem had been solved using conventional iterative optimization algorithms. Such methods employed robust cost functions that were chosen based on assumptions made about the sensor noise and outlier distribution. In practice, these assumptions do not always fit real datasets very well. A recent work showed that the noise distribution could be learnt using a graph neural network. This solution required a second network for outlier detection and removal as the averaging network was sensitive to a poor initialization. In this paper we propose a single-stage graph neural network that can robustly perform rotation averaging in the presence of noise and outliers. Our method uses all observations, suppressing outliers effects through the use of weighted averaging and an attention mechanism within the network design. The result is a network that i...
Scientific Reports
Prostate cancer (PCa) is the second most frequent type of cancer found in men worldwide, with aro... more Prostate cancer (PCa) is the second most frequent type of cancer found in men worldwide, with around one in nine men being diagnosed with PCa within their lifetime. PCa often shows no symptoms in its early stages and its diagnosis techniques are either invasive, resource intensive, or has low efficacy, making widespread early detection onerous. Inspired by the recent success of deep convolutional neural networks (CNN) in computer aided detection (CADe), we propose a new CNN based framework for incidental detection of clinically significant prostate cancer (csPCa) in patients who had a CT scan of the abdomen/pelvis for other reasons. While CT is generally considered insufficient to diagnose PCa due to its inferior soft tissue characterisation, our evaluations on a relatively large dataset consisting of 139 clinically significant PCa patients and 432 controls show that the proposed deep neural network pipeline can detect csPCa patients at a level that is suitable for incidental detect...
The task of learning from point cloud data is always challenging due to the often occurrence of n... more The task of learning from point cloud data is always challenging due to the often occurrence of noise and outliers in the data. Such data inaccuracies can significantly influence the performance of state of the art deep learning networks and their ability to classify or segment objects. While there are some robust deep learning approaches, they are computationally too expensive for real-time applications. This paper proposes a deep learning solution that includes a novel robust pooling layer which greatly enhances network robustness and perform significantly faster than state-of-the-art approaches. The proposed pooling layer looks for data a mode/cluster using two methods, RANSAC, and histogram, as clusters are indicative of models. We tested the pooling layer into frameworks such as Point-based and graph-based neural networks, and the tests showed enhanced robustness as compared to robust state-of-the-art methods.
This paper approaches the problem of geometric multi-model fitting as a data segmentation problem... more This paper approaches the problem of geometric multi-model fitting as a data segmentation problem which is solved by a sequence of sampling, model selection and clustering steps. We propose a sampling method that significantly facilitates solving the segmentation problem using the Normalized cut. The sampler is a novel application of Markov-Chain-Monte-Carlo (MCMC) method to sample from a distribution in the parameter space that is obtained by modifying the Least kth Order Statistics cost function. To sample from this distribution effectively, our proposed Markov Chain includes novel long and short jumps to ensure exploration and exploitation of all structures. It also includes fast local optimization steps to target all, even fairly small, putative structures. This leads to a clustering solution through which final model parameters for each segment are obtained. The method competes favorably with the state-of-the-art both in terms of computation power and segmentation accuracy.
Classification of 3D shapes into physically meaningful categories is one of the most important ta... more Classification of 3D shapes into physically meaningful categories is one of the most important tasks in understanding the immediate environment. Methods that leverage the recent advancements in deep learning have shown to outperform the traditional approaches. However, performances of those methods have only been analyzed with relatively clean data. Three-dimensional measurement sets (point clouds) produced by 3D scanners are rarely that accurate and often contain noise, outliers or missing points. This paper presents an extensive analysis of the robustness of the state-of-the-art neural network algorithms to realistic data inaccuracies. Our experiments show that the existence of these inaccuracies can significantly affect the performance of the deep learning-based algorithms.