Open Access
23 June 2023 DDeep3M+: adaptive enhancement powered weakly supervised learning for neuron segmentation
Rong Xiao, Lei Zhu, Jiangshan Liao, Xinglong Wu, Hui Gong, Jin Huang, Ping Li, Bin Sheng, Shangbin Chen
Author Affiliations +
Abstract

Significance

Robust segmentations of neurons greatly improve neuronal population reconstruction, which could support further study of neuron morphology for brain research.

Aim

Precise segmentation of 3D neuron structures from optical microscopy (OM) images is crucial to probe neural circuits and brain functions. However, the high noise and low contrast of images make neuron segmentation challenging. Convolutional neural networks (CNNs) can provide feasible solutions for the task but they require large manual labels for training. Labor-intensive labeling is highly expensive and heavily limits the algorithm generalization.

Approach

We devise a weakly supervised learning framework Docker-based deep network plus (DDeep3M+) for neuron segmentation without any manual labeling. A Hessian analysis based adaptive enhancement filter is employed to generate pseudo-labels for segmenting neuron images. The automated segmentation labels are input for training a DDeep3M to extract neuronal features. We mine more undetected weak neurites from the probability map based on neuronal structures, thereby modifying the pseudo-labels. We iteratively refine the pseudo-labels and retrain the DDeep3M model with the pseudo-labels to obtain a final segmentation result.

Results

The proposed method achieves promising results with the F1 score of 0.973, which is close to that of the CNN model with manual labels and superior to several segmentation algorithms.

Conclusions

We propose an accurate weakly supervised neuron segmentation method. The high precision results achieved on 3D OM datasets demonstrate the superior generalization of our DDeep3M+.

1.

Introduction

Neurons are the basic units of the structure and function of the nervous system.1 And the neuronal morphology reflects the function of the brain. A complete segmentation of neuronal structure from optical microscopy (OM) images is essential to reveal the connections of brain circuits and investigate synaptic integration, neuron morphology, neuron connectivity, and brain mechanisms.2 Analyzing the morphology and connection of neurons enables us to have a deeper understanding of the operating mechanism of brains and facilitates diagnosis of brain disease.3 For example, congenital nystagmus affects starburst amacrine cells,4 and amyotrophic lateral sclerosis affects upper and lower motor neurons.5 Alzheimer’s disease (AD) is a neurodegenerative disease associated with synaptic loss and neuronal degeneration.6 A precise segmentation of 3D neuron structures is vital to probe impaired brain functions of AD animal models and to determine early treatment strategies.7 We might gain a better understanding of such disorders if we could find specific neuronal morphology in disease models.8

Through the accurate signals of neuron morphology, we can systematically classify brain cells. In recent years, labeling techniques9,10 and optical imaging methods11,12 have made a series of breakthroughs, which can generate terabyte (TB) level neuronal morphological data at whole brain scale.13,14 Micro-optical sectioning tomography (MOST)15 and its follow-up study fluorescence micro-optical sectioning tomography (fMOST)16 have realized the whole brain imaging of specific neural structures at submicron resolution, including tens of thousands of complete neuronal morphologies. Therefore, it is highly difficult to extract and analyze neuron morphology in neuron tracing caused by the uneven distribution of fluorescence,17 differences of imaging system,18 complex morphology of neurons,19 and low signal-to-noise ratio (SNR).20 Currently, many efforts have been devoted to develop automatic or semi-automatic neuron tracing algorithms. However, most of the neuron tracing methods are not applicable in challenging datasets where the 3D neurites are contaminated by strong background noises or containing weak light signals. These methods degrade in performance of neuron segmentation from more complex image blocks. Therefore, an efficient segmentation method is urgently needed to reduce the impact of noises and enhance the weak neurite signals, which should improve the result of neuron tracing.

In recent years, deep learning has achieved remarkable success in processing natural images. Convolutional neural networks (CNNs) have been employed to segment tubular structures due to its capability of feature learning and its nonlinear relationship capturing between inputs and outputs. Deep learning techniques have been adopted in neuron segmentation to improve the segmentation quality. A deep learning toolbox, DeepNeuron, was designed to trace neurons with manually reconstructed neurons as training samples.21 The general supervised deep learning algorithms require a large number of manual labels for reaching an efficient and accurate neuron segmentation, but the segmentation labeling is pretty expensive and labor-consuming. Moreover, extensive a priori knowledge of neuroscience is also required for manual labels.

To solve the above problems, we develop a weakly supervised learning framework [Docker-based deep network (DDeep3M+)] for neuron segmentation from 3D OM images without any manual labeling. It can extract accurate neuron signals robustly from noisy backgrounds. We take advantage of both conventional methods and deep learning techniques as follows. An adaptive enhancement filter is first employed to produce pseudo-labels for training a deep segmentation neural network (DDeep3M). The proposed DDeep3M+ method improves prediction of DDeep3M by iteratively optimizing the training labels and retraining model via region growing and image fusion based on neurite structure information to mine undetected weak neurites from the probability map predicted by DDeep3M. The experimental results on the fMOST16 datasets and public BigNeuron22 datasets have demonstrated the high precision and generalization achieved by the proposed neuron segmentation method; see Fig. 1.

Fig. 1

Typical segmentation result of two diverse OM image datasets, including the maximum intensity projection of fMOST16 and BigNeuron22 dataset. Red pixels indicate the ground truth. Yellow pixels represent the overlapping regions between the prediction and ground truth.

NPh_10_3_035003_f001.png

Our main contributions are summarized as follows:

  • 1. We propose a precise, automatic, and general method for neuron segmentation from noisy and low-contrast OM images. And a weakly supervised learning framework based on DDeep3M network is presented to improve 3D neuron segmentation without manual labels.

  • 2. An adaptive enhancement filter is presented in the first stage to generate pseudo-labels by enhancing weak neuronal structures based on Hessian matrix in a fibrous structure. We take advantage of adaptive enhancement filter and deep segmentation network to make them mutually complement to each other.

  • 3. Neuron segmentation results on 3D OM datasets demonstrate the high precision and generalization by the proposed method in automatic neuron segmentation. Moreover, our DDeep3M+ outperforms state-of-the-art weakly supervised neuron segmentation algorithms on fMOST and BigNeuron datasets.

2.

Related Work

Morphology of neurons is essential for the organization and function of the brain. Segmentation or tracing of 3D neuron structures from OM images are the main approaches to characterize and analyze neuronal morphology. There are 2,32,613 digitally reconstructed neurons contributed by 800 laboratories worldwide in NeuroMorpho19 dataset, which took more than 2.5 million hours of manual work in the past decades. In computational neuroscience, there are high requirements for automatic and accurate neuron segmentation methods.

With the development of optical imaging and molecular labeling technology, we can obtain large-scale sub-micron resolution neuron images through optical imaging of mammalian brain. These advances promote the generation of various optical images for different applications, and pose new challenges in neuron segmentation. The limit is the significant differences in image quality and attributes between different datasets due to a variety of factors, such as intensity range, image size, neuron structure, etc.22 Another challenge is that microscopy images usually have high background noise. It has difficulty in separating weak neuron voxels from an inhomogeneous background.23 It is tough to distinguish the neurites with low intensity and non-uniformity from the noise background, especially for the large-scale neuron images with discontinued segments of neurites and ultra-low SNR. The variability between different datasets and the low SNR of optical images increase the difficulty of generalization and improvement of neuron segmentation algorithms. Many semi-automatic or automatic methods have been proposed for neuron segmentation or neuron tracing.24 In these algorithms, various computational methods addressing global and local image features have been used to realize neuron segmentation and tracing. These include but are not limited to region growing, tubular model, full path pruning, graph theory method, and support vector machine based on self-learning.25,26 These algorithms usually show good performance on OM images with clear structure. However, the above methods are designed for specific datasets or specific problems, and their performance on different types of datasets may degrade, so complex parameter tuning is needed. In addition, most algorithms perform poorly in tracing neurites from low SNR images, and are unable to recognize weak neurite signals or prone to over-segmentation due to background noise.

Deep network27 has been widely used in many fields, including computer vision, natural language processing, speech recognition, and so on, and has achieved excellent results in these tasks.28 Since the full convolution network (FCN),29 various deep networks have been proposed to improve the accuracy of image segmentation, including U-Net,30 3D deeply supervised network (DSN),31 and VoxResNet.32 The excellent performance of deep network encourages researchers to apply it to neuron segmentation.33 In Ref. 34, a deep network is designed to improve the performance of existing tracing algorithms. The main purpose of the deep network is to denoise the image containing a single neuron or multiple sparse neurons. DeepNeuron,21 a deep learning toolbox, is designed to trace neurons with human-reconstructed neurons as training samples. But they have poor performance in the case of weak neurites. Yang et al.35 employed a two-stage 3D neuron segmentation approach to segment neuron images from the BigNeuron dataset, via learning deep features by an FCN model and enhancing weak neuronal structures based on Hessian Eigenvalues in a fibrous structure. Structure-guided segmentation network (SGSNet),36 a two-branch network for 3D neuron segmentation, contains a shared encoding path but utilizes two decoding paths to enhance weak neuronal structures and remove background noises. There are 3D CNNs specifically designed for neuron segmentation and tracing that show better performance on images with high noise.37,38 However, these algorithms require a lot of labor, time-consuming, and expensive manual annotation to segment neurons. Users need to annotate enough new samples for neuron datasets from different brain regions to achieve reliable prediction. The need for manual annotation has greatly limited the promotion of deep learning-based methods for various optical neuron images.

Here, we propose a weakly supervised learning method for deep CNN without manual labeling. Integrating the advantages of traditional image enhancement and deep learning, we can extract neuron signals accurately and robustly from noisy background, and realize neuron segmentation in OM image.

3.

Materials and Methods

To segment neuron population from low SNR optical microscopic images, our neuron segmentation framework consists of four key modules: pseudo-labels module, DDeep3M segmentation network, region growing module, and image enhancement module, which are illustrated in Fig. 2. The automatic neuron segmentation framework includes four steps: (1) initial neurite segmentation is automatically obtained by an adaptive enhancement filter and threshold segmentation algorithm from optical images as their pseudo-labels, (2) training of the DDeep3M segmentation network using input images and their pseudo-labels as training sets, and prediction of the neurite as probability map using DDeep3M, (3) refinement of pseudo-labels from probability map based on region growing, (4) combining the original image intensities with the refined probability map. Steps (1) to (4) are iterated to update the probability map. The segmentation network is optimized, and the completeness and accuracy of neuron segmentation are improved. By fusing the prediction probability map of neuron voxel with the original intensity to enhance the original image, we can preserve the global structure from probability map and local signal from original image at the same time. We gradually improve the performance of neuron segmentation by combining the traditional image enhancement method and the DDeep3M network without any manual annotation. Finally, we complete the automatic segmentation of large-scale neuronal population from noisy and low-contrast OM images through whole framework.

Fig. 2

Schematic illustration of the proposed weakly supervised learning framework (i.e., our DDeep3M+) for automatic neuron segmentation. Our method consists of four steps: (1) using adaptive enhancement filter to segment the neural signal from the OM image and generate pseudo labels; (2) training and predicting of DDeep3M network with raw images and their pseudo-labels as training sets; (3) refining pseudo labels using region growing; (4) combining the prediction probability map with the original image. Iterative steps 1 to 4 update the pseudo label of training samples to improve segmentation.

NPh_10_3_035003_f002.png

3.1.

Pseudo-Labels

There are some problems in the 3D optical images of neurons, such as weak signal, strong noise, and uneven signal distribution. All these problems seriously restrict neuron segmentation, so we enhance the original image to remove the noise in the first place. Usually, a neuron consists of a soma and a great number of fibers. The fibers are formed from an axon and dendrites. Seeing a neuron in mathematical graphical structure, we can understand that a cell body is a dot-like structure and a neurite is a fiber-like or line-like structure.39 We propose an adaptive enhancement for the 3D image of neurons based on distance transform (DT) and Hessian matrix. The purpose of enhancement filters for neuron is to get clearer and higher contrast images, and then the neuron can be segmented after being enhanced.

The Hessian matrix is a mixed second derivate matrix, which simply indicates that the gray intensity trends of continuous voxels in a 3D image. A 3D raw OM image volume was presented as I (x,y,z). Hessian matrix was defined as the second order derivatives of the image intensity

Eq. (1)

H=(GI)((GI)T)=[IxxIxyIxzIyxIyyIyzIzxIzyIzz],
where G=(2πs2)3/2exp((x2+y2+z2)/(2s2)) is a Gaussian kernel. G denotes the first order Gaussian derivatives. is convolution. Iij denotes the mixed second derivative along dimensions i and j.

First, the window size of the Hessian matrix is determined by DT: DT can measure the thickness of different neuron fibers effectively, and it can be used to calculate the Hessian matrix. Second, the method constructs a structural response function to enhance the neuron fibers: when the image signal is bright and the background is dark, the Hessian matrix eigenvalue λ1 is approximately equal to 0, λ2 and λ3 are <0 and their amplitudes are close to each other. The method takes advantage of this characteristic to enhance the neuron fibers. Third, the method enhances the soma with different strategy: the signal value of the soma is high and the radius is large. Accordingly, this method sets a higher threshold for DT to localize the soma for hole filling. Through the above three steps, the enhancement of neuron with different structures is finally achieved. Through a Hessian analysis-based neuron enhancement and neurite threshold segmentation, the initial neurite segmentation is automatically obtained, and then they are regarded as initial training labels (pseudo-labels).

We devise a way to get an adaptive window size, which can guarantee that the Hessian matrix has a proper window size in our image of uneven intensity. First, we distinguished the foreground regions and background regions with a determined value derived from a priori knowledge, the image was transformed to be binary. Then we used the Euclidean DT to describe the degree of the thickness. Here, the Euclidean DT between M and N is D(M,N)=(i=1k|mini|p)(1/p). Here M and N are k-tuples (k=3 in this study), mi and ni are the i’th coordinates of M and N, p=2 in Euclidean distances. The purpose of computing the DT is to determine the window size of the Hessian matrix. As a second-order partial derivative matrix, the Hessian matrix needs to determine how many points to calculate when calculating between voxels. The size of the Hessian matrix window is closely related to the thickness of the nerve fibers. It has a great influence on the subsequent nerve fiber enhancement. First, the center point of a cell body is set as X. After the Euclidean DT, the DT value of the cell center point is set to DT(X). Next, the window size of the Hessian matrix is calculated by the obtained DT value.

Then the window radius is determined by the following formula: R=log2DT(X). The intensity of neurites in the middle is higher than that at the edge. DT image reveals the structure information of the graph, to a certain extent. We made use of this useful information and defined an optimal window radius R for each voxel. The DT value for each location (x,y,z) is defined as D (x,y,z), and all the DT values for each voxel in the image build a DT matrix, which has the same size as the original image. Then we normalize DT(X) between 1 and 256. We defined the normalized result for each voxel as DN (x,y,z). We make a logarithmic operation for DN, the result of it comes to be the window radius. Indeed, in this way, we basically used a window with radius R between 1 and 8. The window size, that is, window diameter DI is defined as twice the radius and plus one. The above formulas are as follows: R=log2DN(x,y,z), DI=2R+1. It can provide a proper size for each voxel for the uneven gray-scale characteristics of different locations in neurites to calculate the Hessian matrix.

Since the Hessian matrix is symmetric, there exist three eigenvalues λ1, λ2, and λ3. For convenience, they are sorted based on the absolute value: |λ1||λ2||λ3|. In previous studies, such as Frangi’s40 method in 1998 when the signal is bright and the background is dark, the Eigenvalues of an ideal 3D dot-like, line-like, and plate-like area satisfy the following conditions: line-like: λ10; λ2λ3<0; dot-like: λ1λ2λ3<0; plate-like: λ1λ20; λ3<0. Initially, taking account of the neurite enhancement, Frangi’s filter proposed three derived parameters based on the above-mentioned Eigenvalues: p1=|λ2|/|λ3|, p2=|λ1||λ2λ3|, and p3=λ12+λ22+λ32. The response function was defined as follows:

Eq. (2)

(X)={0,if  λ2>0  or  λ3>0(1eap12)ebp22(1ep32c),else
where a, b, and c are thresholds that control the sensitivity of the line filter to the measures p1, p2, and p3. Through the test of multiple groups of data, we found the optimal values of parameters: a=5:55, b=2, and c=2×106. When a voxel in the image is a point in a plate-like structure, (1eap12) approaches 0. The calculation result of O(X) is 0 without response. When a voxel is a point in a dot-like structure, ebp22 approaches 0. The result of O(X) is 0 without response. When a voxel is background signal, (1ep32/c) approaches 0. The result is with no response. Therefore, nerve fibers can be specifically enhanced by the Eq. (2).

The above work is for line-like structures enhancement, the soma of a neuron will be turned into a circular ring or an irregular shape with a hole,41 as shown in Step 1 of Fig. 2. To solve the problem and make our algorithm complete and useful, we must design a method to detect the holes. In the original fluorescent neuron image, the intensity of the soma and middle part of the fiber is always higher than the other pixels. We take advantage of this feature and combine the characteristic of DT, then design the following approach to detect the holes in soma and thicker fiber. First, we choose a relatively higher determined value I2, which is bigger than I1, with a priori knowledge. Then, we recalculate DT matrix with the higher threshold I2, defined as D2, the result will show the location of the holes in soma and fiber. We can use the location to fill the hole in the enhanced image by replacing it with high intensity at the same location. If D2(X)>T, O(X)=255; otherwise, it remains O(X). Where X is the voxel, D2(X) is the local DT value for X with higher threshold, T is the threshold to determine whether a voxel is within the soma area or not. And O(X) is the enhanced result before soma detection. If the voxel is assumed as a soma voxel, we replace the intensity value. Otherwise, the previous enhanced result will be retained. The O(X) is the enhanced result of original image, and then we obtain the initial training labels (pseudo-labels) through threshold segmentation after enhancement by our Hessian analysis based adaptive enhancement filter.

3.2.

Segmentation Network—DDeep3M

Docker is an open source lightweight virtualization technology. DDeep3M42 integrates the CDeep3M43 network by Docker, and it can run on the local computer by downloading a separate Docker image file containing all necessary libraries, software and framework. DDeep3M greatly reduces the threshold for biomedical researchers to accept deep learning technology. It achieves excellent segmentation performance on multiple datasets and performs better than most related models, especially in the neuroscience image segmentation tasks.

The architecture of DDeep3M has some similarities to the U-Net architecture,30 which consists of a bottom-up path and a resolution recovery path. We aggregate multiscale context information required for classification and minimize the effect of noise via using the z-direction information. However, there are inception and residual modules in the coding path, which are some key differences from U-Net. In deep learning, deeper layers can lead to overfitting and vanishing gradient problems. The size of the receptive field is another important factor affecting network performance. The inception module allows information to pass through multiple sized cores in parallel, applying 1×1 convolution reduces the number of input feature maps before using large kernel convolution to effectively reduce the complexity of the model. Residual learning is able to alleviate the gradient vanishing problem. We apply skip connections similar to those in FCN,29 but add pyramid dilated deconvolution to further enhance the ability to represent multi-scale information in the spatial recovery path.

Because neurites only account for a small part of the data block, the categories of foreground (i.e., segmented neurites) and background are usually unbalanced, which may lead to prediction bias. The weighted cross-entropy (WCE) loss function44 and dice loss (DL) function45 reduce the impact of class level imbalance. We use a mixed loss function to combine these two loss functions to prevent class imbalance and maintain regional continuity. The WCE loss is

Eq. (3)

LWCE=iLjWkHγ·g(xijk)log(p(xijk)),γ=iLjWkHg(xijk)L×W×H.

Then, the DL is

Eq. (4)

LDL=12iLjWkHp(xijk)g(xijk)+εiLjWkHp(xijk)+iLjWkHg(xijk)+ε,
where p(xijk) is the prediction probability of pixel x; g(xijk) is the corresponding pseudo-labels of foreground or background, and its values are 1 or 0 respectively; L is the length, W is the width, and H is the height of volume data; γ is the ratio of the number of voxels of foreground pseudo labels to the number of voxels of image volume data N; ε is the smoothing parameter set to 1. Thus, the combination loss function to train the model is L=σLWCE+LDL, where σ is the weight of cross-entropy loss LWCE, which is used to balance the ratio of loss function.

3.3.

Pseudo-Label Refinement

Some weak and uneven neuron signals are not detected in the pseudo-labels obtained by an adaptive enhancement filter, which reduce the network prediction accuracy of neuron images with low SNR. Therefore, it is necessary to improve the pseudo-labels for better segmentation. Huang et al.46 proposed a weak supervised learning method for neuron reconstruction. The region growing and skeleton method is used to mine more weak neurites from the probability map predicted by CNN iteratively. Here, we use the region growing method employed by Huang et al.46 to find more weak neurites from the probability map predicted by DDeep3M iteratively. The steps of the regional growth method are as follows:

  • 1. Selection of seed region: According to the predicted probability map, the algorithm with maximum probability classification is used to generate the seed region of neurite. An adaptive threshold ρ is calculated for region growth in Step (2). ρ is the average value of voxels near the seed region (Nreg). Nreg is defined as follows:

    Eq. (5)

    Nreg={vN(v0)vOreg,v0Oreg},
    where N(v0) is the 124 voxels neighborhood of v0; Oreg is the seed region; and v, v0 are the voxels.

  • 2. Region growing: Based on neurite continuity, the seed region is expanded to include weaker neurites using region growing algorithm

    Eq. (6)

    Greg={vN(v0)s(v)>ρ,v0Oreg}
    where N(v0) is the 8 voxels neighborhood of v0, s(v) is the probability map value of v; ρ is the adaptive threshold of region growth, and Greg represents the grown region.

Compared with the seed region, the grown region contained more voxels from neurites. Some weak neuron signals in the post-processing prediction graph are marked, which can be fused with the original image to obtain new enhanced training data for the next round of network training, to improve the accuracy of network prediction.

3.4.

Image Enhancement

The output of DDeep3M network is the 3D probability map P(x), which is calculated by the SoftMax activation function after the 3D data block passes through convolution layers of the network. P(x)[0,1] denotes the probability that voxel x becomes the part of neuron. To use the prediction results, a natural method is to segment neurons directly from the prediction graph. However, because the accuracy of pseudo labels is not as good as manual annotation, especially in early iterations, some local details may be lost in the prediction graph. Therefore, we enhance the original image by fusing the prediction image and the original image to maintain the accurate neuron structure and suppress the noise signal effectively. By fusing the prediction result graph P(x) with the original image I(x) signal, the enhanced image block F(x) can be obtained, to preserve the local signal and the global structure at the same time (the image enhancement formula is shown below). When the enhanced image block F(x) is delivered to the Hessian module, a more complete neural community can be segmented and a better pseudo-label can be provided for the next iteration of network training. In the process of iterative learning, DDeep3M and Hessian modules complement and promote each other, thus gradually improving the performance of neuron segmentation.

Specifically, for an input image I(x), where x is a voxel, we identify the foreground voxels through the threshold δ (δ=2) selected by experience, to screen the probability map.34 If the foreground probability P(x) of voxel x is <δ, we set the intensity I(x) to zero, otherwise we keep the original intensity value unchanged. We use a step-edge function Θ(I(x)δ) to describe an intermediate image. if P(x)>δ, Θ(I(x)δ)=I(x); if P(x)δ, Θ(I(x)δ)=0. Then, intermediate images and probability maps are fused, and the final image F(x) adjusted by our enhancement function is defined as

Eq. (7)

F(x)=αα+β·Θ(I(x)δ)·I(x)+βα+β·(1α)·IM·P(x),
where IM is the maximum intensity of the image, and α[0,1] is the weight controlling the contribution of voxel x in the original intensity, whereas β[0,1] is the weight controlling thecontribution of voxel x in the prediction graph, which were both set between 0 and 1. Using more reliable segmentation results as pseudo-labels, the segmentation network can be further trained to learn more distinctive and representative features for generating the probability map, which is conducive to the Hessian matrix based enhancement filter to segment more complete neurons in the next iteration. To explore the influence of parameter α for image enhancement, we adopt different values for α. In this paper, we empirically select α=0.8 to increase the robustness of the whole framework.

During each iteration of the training process, we calculate the F1 value (Ft) of the test dataset. If the difference between Ft and the F1 value from the previous iteration (Ft1) is <0.005, and the difference between Ft and the F1 value from the next iteration (Ft+1) is also <0.005 (i.e., FtFt1<0.005 and Ft+1Ft<0.005), we consider the iteration to have converged to the optimal solution. Based on our experimental results (as shown in Fig. 3) and experience, we have found that the number of iterations required to converge to the best solution is usually between 4 and 6. Once convergence is achieved, the network training is considered complete.

Fig. 3

Comparisons of F1 value of the neuron segmentation at different iterations with four deep segmentation networks, respectively. After 4 iterations, DDeep3M network achieves the best segmentation result, with F1 value up to 0.973. Our DDeep3M+ effectively improves the neuron segmentation performance by combining any one of the three-neuron segmentation CNNs (U-Net,30 3D DSN,31 and VoxResNet32).

NPh_10_3_035003_f003.png

4.

Results

4.1.

Datasets and Settings

We evaluate the proposed method and state-of-the-art segmentation methods based on deep learning on various 3D optical neuron datasets, including the fMOST datasets and the BigNeuron datasets. For the fMOST dataset, we select an image stack with the voxel size of 0.5  μm×0.5  μm×1  μm from the images of mouse neurons at the single fiber level for training and testing. The fMOST dataset used in the experiment is 3D volume data obtained from transgenic mouse Thy1 sample using two-photon fMOST (2P fMOST) system, which combines slicer and two-photon microscopy to realize continuous block imaging. In this study, the prediction set included the fMOST image stack of 3687×1224×1000  voxels from the V layer of the mouse cerebral cortex. During the training phase, we input a representative fMOST image stack of size 200×200×350  voxels into the DDeep3M framework for iterative network training. The image stack was divided into three datasets for training, validation, and testing, respectively. The training dataset comprised a volume of 200×200×200  voxels, whereas the validation and testing datasets were 200×200×50  voxels and 200×200×100  voxels in size, respectively. For each iteration, we fed the training and validation datasets, along with the corresponding pseudo-labels, separately into the neural network for training. The network’s performance on the test dataset and the ground truth data were used to evaluate its prediction accuracy at each iteration, which was then used to decide whether to continue training the network or terminate the iteration.

4.2.

Training Details

The fMOST image stack is processed by a histogram equalization technology to strengthen the contrast, and then normalized by merely dividing the value of all pixels by 255. Finally, the data augmentation technique is applied to the original training set. Data augmentation can expand the effective size of training data, thereby decreasing the problem of over fitting. It is normally used to train CNN for image classification. By applying several spatial transformations to the input image, we enhance the training and test data. These transformations are a combination of horizontal and vertical mirrors, rotation +90  deg, 90  deg, and 180 deg in the XY plane, and flipping in the Z direction (horizontal mirror). A total of 5 datasets of 1000 images are employed in training. After all the converted data is provided to the network, we apply reverse conversion for each probability map. We take the predicted average of each transformation graph and the original graph as the final output of the boundary probability map.

We initialize the network by utilizing a Gaussian distribution with a standard deviation of 0.01. The mini batch size of our training model is 4, and the base learning rate is 0.01. The stochastic gradient descent optimization method with a momentum of 0.9 is adopted. Each model trains 30,000 iterations, and the mixed cross-entropy loss function is taken as our training objective. The network segmentation was trained and executed on a computer equipped with an Intel Xeon w-2123 CPU (64 GB RAM) and a Nvidia P5000 GPU (16 GB RAM). The prediction task for the fMOST image stack (3687×1224×1000  voxels) was completed by DDeep3M+ in 5  h, resulting in an average processing rate of 18.4 s per image.

4.3.

Evaluation Metrics

To conduct quantitative comparisons among different segmentation methods, we first perform a binarization operation on the segmentation results obtained by different segmentation methods with a threshold value of 200, and then utilize common metrics to quantitatively evaluate the segmentation result between the ground truth labeled by human experts. They are precision, recall, F1, and Jaccard, and their definitions are given by

Eq. (8)

Precision=TPTP+FP,

Eq. (9)

Recall=TPTP+FN,

Eq. (10)

F1=2·Precision×RecallPrecision+Recall,

Eq. (11)

Jaccard=TPFP+TP+FN,
here TP, TN, FP, and FN denote the true positive, true negative, false positive, and false positive.

To demonstrate the segmentation performance of our weakly supervised learning strategy, four widely-used deep segmentation networks, including U-Net, 3D DSN, VoxResNet, and DDeep3M, are tested to generate the neuron probability map in our framework. Eight iterations are tested on the fMOST dataset, and the F1 value improvement of segmentation results is shown in Fig. 3. It can be seen that our DDeep3M+ algorithm achieves the best segmentation effect, with the F1 value up to 0.973 after fourth iterations. After the fifth iteration, the F1 value decreases, and the network segmentation effect does not improve when the iteration number increases. It indicates that the prediction performance of the weakly supervised learning framework has reached the best after the fourth iteration. In addition, the performance of the DDeep3M network is better than the other three deep networks.

Moreover, the neuron segmentation results on a test block at different iterations are shown in Fig. 4, which depict the mining process of pseudo-label refinement. In Fig. 4, we can see that the prediction effect of DDeep3M network is more and more accurate, and many weak nerve fiber signals can be accurately segmented in the later stage, after several iterations of training.

Fig. 4

An example of pseudo-label refinement for weak neurite mining. (a1)–(a4) Probability map by DDeep3M at different iterations; (b1)–(b3) region grown to include nearby weak neurites; (c) original image. Yellow arrowheads point to the grown region. We iteratively mine more undetected weak neurites from the CNN-predicted probability map.

NPh_10_3_035003_f004.png

4.4.

Comparison with Other CNN Models

By comparing the detection performance of the proposed method against that of the common deep learning methods, we tested the effectiveness and accuracy of the proposed weakly supervised learning method for neuron detection and segmentation. The deep learning method used in this experiment is based on the supervised learning of the DDeep3M network. Figure 5 showed the segmentations of the proposed method on the fMOST datasets. Figure 6 displays the 3D rendering of the prediction results for a larger fMOST image stack (3687×1224×1000  voxels), as well as the maximum intensity projection of the original images and segmentations for 50 slices of the image stack. The proposed method detects almost all neurites from OM images. According to Table 1, the F1 score of the proposed method and the supervised deep learning method are 0.973 and 0.980, respectively; the recall and precision scores of the proposed method are 0.952 and 0.994, respectively; and the recall and precision scores of the deep learning method are 0.970 and 0.985, respectively. In Table 1, we also implement segmentation models that contain different parts of our DDeep3M+ modules in the fMOST image trainings, and the results show the performance for the different key structures. Table 1 shows that our DDeep3M+, which contains our enhancement filter to obtain pseudo-labels, has the best performance for fMOST neuron segmentation. We implement four segmentation networks with a threshold method (named Otsu) for pseudo-labels. And the results show that their performances are far worse than our DDeep3M+ method.

Fig. 5

Validation of DDeep3M with the fMOST16 dataset after four iterations. (a) Original image, (b) segmentation of DDeep3M+, (c) ground truth, (d) merge of the segmentation (green) and the ground truth (red). The segmentation of neuron signal is realized accurately by DDeep3M+.

NPh_10_3_035003_f005.png

Fig. 6

Prediction by DDeep3M+ on fMOST16 image stack. (a) 3D rendering of the prediction in a larger image stack. (b) Part of raw images in (a). (c) A magnified view of some of the nerve fibers in (b). The signal at the ends of the fibers is very weak and the connections are intermittent. (d) Segmentation of (b) with DDeep3M. (e) A magnified view of some of the nerve fibers in (d). Weak nerve fiber signals can be accurately identified and segmented, and the connections of nerve signals are continuous.

NPh_10_3_035003_f006.png

Table 1

Quantitative comparison results of neuron segmentation by different methods on fMOST16 data.

MethodsPrecisionRecallF1Jaccard
DDeep3M+Our0.9940.9520.9730.947
Otsu0.7640.7350.7500.599
U-Net30Our0.9930.9130.9510.907
Otsu0.5990.8360.5980.536
3D DSN31Our0.9940.8360.9080.831
Otsu0.3860.9260.5450.374
VoxResNet32Our0.9920.8920.9390.885
Otsu0.9480.5180.6700.504
Supervised learning420.9850.9700.9800.990

Zhao et al.47 proposed a progressive learning method to reconstruct neurons via 3D DSN network from ultra-large scale optical microscopic images, and it is currently the excellent method for weakly supervised learning in the field of neuron segmentation. Figure 7 shows that neither U-Net segmentation network nor progressive learning method has an ideal segmentation effect on MOST dataset by comparing the segmentation performance of different network. We tested the proposed method and compared the final result with other three networks with a mouse brain dataset as shown in Fig. 8, which shows more details for comparison results between our method and other three segmentation networks. The result of proposed method shows more details of the fibers. As pointed out by the yellow box images in Fig. 8, the VoxResNet network was unable to detect neurites with weak or sudden changes in intensity. Similarly, it is difficult to trace some weak and uneven neurites from high noise images by 3D DSN network. In Fig. 8(f), there are some false positives in the result of U-Net network, which failed to trace some weak neurites precisely from the noisy background. Our method, employing weakly supervised learning, can accurately segment neuron cell body and fiber, and even some weak neurites can be accurately identified, and its segmentation performance is comparable to that of traditional deep learning method with manual annotations.42

Fig. 7

Segmentation performance of different networks: (a) raw images, (b) U-Net, (c) 3D DSN, (d) supervised method, and (e) our method. Neither U-Net30 network nor other weakly supervised learning method has an ideal segmentation effect on fMOST16 dataset. Our method segments neurons more completely and accurately compared to other methods. The proposed method is comparable to that of supervised deep learning method.

NPh_10_3_035003_f007.png

Fig. 8

Enhancement comparison results in fMOST16 dataset. Images were acquired with a voxel size of 0.5  μm×0.5  μm×1  μm. (a) Original images and (b) enhanced results of our method. (c)–(h) Details of original images, ground truth images, enhanced results of our method, U-Net30 method, 3D DSN31 method, and VoxResNet32 method. Yellow box images show the situation of weak neurites.

NPh_10_3_035003_f008.png

4.5.

Comparison on Other Datasets

We also test the performance of the proposed method on other datasets, as shown in Table 2. The proposed method also performs well on the BigNeuron datasets, with the highest F1 value up to 0.880, much higher than the 0.809 of U-Net network. The F1 value of U-Net, 3D DSN, and VoxResNet were 0.809, 0.816, and 0.649, respectively, whereas the performance of DDeep3M in our experiment is 0.855 (recall), 0.907 (precision), and 0.880 (F1 value). After the comparison of various quantitative metrics, the proposed method is better than other segmentation methods in all aspects. Figure 9 shows the segmented neurons on three test images from the BigNeuron dataset.

Table 2

Quantitative comparison results of neuron segmentation by different methods on BigNeuron22 data.

MethodsPrecisionRecallF1Jaccard
DDeep3M+0.9770.8850.9290.867
U-Net300.9210.7220.8090.680
3D DSN310.7720.8810.8060.677
VoxResNet320.9140.6920.7840.649
Supervised learning420.8900.8020.9050.876

Fig. 9

Segmentation by DDeep3M+ on BigNeuron image stacks. (a)–(c) Original images; (d)–(f) segmentations of (a)–(c). The proposed method also performs well on BigNeuron dataset.

NPh_10_3_035003_f009.png

4.6.

Enhancement Parameter

To explore the influence of parameter α in enhancement function on image enhancement of our segmentation framework, we use different values of α and show the corresponding segmentation results of our method in Fig. 10. α=0 means that the original data block is directly used as the input of neuron segmentation. α=1 means that only the probability map is used as the input of neuron segmentation. This shows that the performance is improved by combining the probability map with the original image signals. The main reason is that the probability map reflects the long-range trajectory structure, whereas the original image signal carries more details of the neurites. The experimental results show that the segmentation performance can be improved by combining the probability map with the original image signal, which is mainly because the probability map reflects the long-range trajectory structures, whereas the original image signal has more details of subtle neural processes. Since we used pseudo labels to train the CNN model, its performance was limited when comparing it against the supervised learning method. To reduce the influence of false positive prediction on the probability map and combine with experimental verification, we chose α=0.8 in this paper to improve the robustness of the whole framework.

Fig. 10

Neuron segmentation performance with different α in Eq. (7) for image enhancement. We choose the best model α=0.8 to reduce the impact of false positive prediction on probability map and improve the robustness of the whole framework.

NPh_10_3_035003_f010.png

5.

Discussion

Deep learning-based algorithms provide an effective and automatic method for neurite detection in highly noisy backgrounds. However, an accurate and robust estimation relies on large numbers of voxel-wise annotations. In recent years, the method of weakly supervised learning has gradually become a new trend in biomedical image segmentation. Huang et al.,46 Zhao et al.,47 and Chen et al.48 proposed three weakly supervised methods for neuron tracing and reconstruction. They solved the problem of the scarcity of manually annotated training samples and obtained better neurite detection results. Here, we employ a well-established CNN (DDeep3M) with inception and residual modules to achieve accurate, automatic, and universal neuron segmentation of 3D optical images for different types of neuron datasets with low and non-uniform signal strength. The proposed method is superior to several novel weakly supervised segmentation methods.

First, we apply the Hessian-based adaptive enhancement filter for 3D neuron images based on DT to obtain more accurate pseud-label. Our enhancement filter can preserve the neuron signals of subtle intensity. The enhanced images have shown better visual effect, higher image quality, more abundant information. In addition, the proposed algorithm has shorter time consumption. Thus, this algorithm will benefit the morphological study of neurons. The adaptive enhancement filter based on Hessian matrix is used to extract the global feature of the neuron voxel, which can effectively provide the approximate location of the neuron. However, some weak, disconnected, and heterogeneous neurites are not detected in the pseudo-labels obtained through the Hessian matrix, which reduces the accuracy of image segmentation for low SNR neuron images. Therefore, finer neuronal signals must be mined for better segmentation. Deep CNN has achieved impressive performance in both natural and medical image segmentation, and its efficient performance in extracting image features gives us the possibility of our work. Further use of the DDeep3M network on the adaptively enhanced image can detect almost all the neurites in the neuron image, including the difficult to identify thin neurites with relatively low intensity.

The initial training labels are constructed by the traditional image enhancement algorithm based on the neurite structure features. Then, the DDeep3M network improves prediction by iteratively optimizing the training labels and retraining model, while using regional growth to mine undetected weak neurites from the probability map predicted by DDeep3M. With more iterations of network training and image enhancement, the pseudo-labels are progressively refined and the DDeep3M model can gradually learn discriminative features of the neurites and background. Therefore, DDeep3M based segmentation method and traditional image enhancement technique can complement and promote each other. Finally, completer and more accurate neuron segmentation is obtained by our DDeep3M+ method.

In the training phase, deep learning completes the model establishment and parameter optimization, and in the inference phase, it completes the specific implementation and application. After carefully adjusting the weights, the neural network is basically a bulky and huge database. To make full use of the training results and complete the segmentation task in practical application, we used data quantification, layer fusion, and model pruning to optimize the inference process of the network. After the acceleration technology in inference phase, the speed of network prediction can be increased by about 12 times. At the same time, the F1 value of the prediction results did not decrease significantly, indicating that the inference speed was improved and the accuracy was not lost.

There are still some limitations of our work. (1) Our method is currently not as good as supervised learning methods in accuracy, although we have an advantage in the cost of manual labeling. In the future work, we need to make better use of the prior knowledge of neuronal signals and improve the architecture of deep networks to improve the prediction accuracy of weakly supervised learning and achieve more accurate segmentation of neuronal protrusions. (2) The selection of the value of the parameter in the enhancement formula in the work of image fusion is empirical, and there is no theoretical support. We choose α=0.8 as the weighting value of the original image according to the experience of the experiment. The enhancement result is the best, which is only obtained from limited discrete data. (3) The selection of the number of iterations in the whole weakly supervised learning scheme is also obtained according to the experience of experiments, which also lacks the explanation of the algorithm principle. (4) We have not carried out neuron reconstruction experiments on the results of neural network segmentation. In the future work, we will use neuronal images enhanced by neural networks for neuron tracing and reconstruction, so as to analyze the advantages of our method in promoting neuron tracing, reconstruction and visualization compared with other deep learning methods.

Compared to supervised learning methods, our method adapts to different types of optical neuron datasets without manual labeling, model redesign, or parameter tuning. The high accuracy results on challenging 3D optical images from different types of datasets demonstrate the accuracy and generalization of the proposed neurite segmentation method. The proposed weakly supervised learning framework can be extended to multi-modal biomedical big data, such as soma, vessels, and tumors, which helps life science experts to efficiently use deep learning techniques and promote further brain research. In addition, the source code of our method is open on the GitHub, which helps neuroscience researchers to easily use deep learning tools for their work on neuron segmentation and reconstruction.

6.

Conclusions

We propose a weakly supervised deep learning method (i.e., DDeep3M+) for automatic neuron segmentation in 3D OM images. Our DDeep3M+ framework consists of an adaptive enhancement filter that extracts initial neuron signals as initial neurite segmentations and a deep segmentation network for precise neurite segmentations from low SNR optical images without manual labeling. The accuracy of segmentation is close to that one of the best supervised learning methods. The comparison between the proposed method and several new segmentation methods proves the superiority of the proposed method in the segmentation of neurons from low SNR images. The proposed framework is effective for both public fMOST and BigNeuron datasets, and can be used for automatic segmentation of super-large neuronal population in the entire brain region. The codes and the pretrained model weights of our method are available at GitHub, which will help neuroscience researchers to easily use deep learning tools for their work on neuron segmentation and reconstruction. We believe it will promote further brain studies, including neuron tracing, neuron reconstruction, and so on.

Disclosures

The authors declared that they have no conflicts of interest to this work.

Code, Data, and Materials Availability

The codes, datasets, and the pretrained model weights of our method are available at https://github.com/xiaoronghust/DDeep3Mplus/.

Acknowledgments

This work was supported in part by the National Natural Science Foundation of China (Grant Nos. 61890951 and 61371014), and the Director Fund of WNLO, Open Project Program of Wuhan National Laboratory for Optoelectronics (Grant No. WNLOKF027).

References

1. 

A. Peters, “Golgi, Cajal, and the fine structure of the nervous system,” Brain Res. Rev., 55 (2), 256 –263 https://doi.org/10.1016/j.brainresrev.2006.12.002 BRERD2 0165-0173 (2003). Google Scholar

2. 

S. L. Senft, “A brief history of neuronal reconstruction,” Neuroinformatics, 9 (2–3), 119 –128 https://doi.org/10.1007/s12021-011-9107-0 1539-2791 (2011). Google Scholar

3. 

H. Zeng and J. R. Sanes, “Neuronal cell-type classification: challenges, opportunities, and the path forward,” Nat. Rev. Neurosci., 18 (9), 530 –546 https://doi.org/10.1038/nrn.2017.85 NRNAAN 1471-003X (2017). Google Scholar

4. 

F. Crick and E. Jones, “Backwardness of human neuroanatomy,” Nature, 361 109 –110 https://doi.org/10.1038/361109a0 (1993). Google Scholar

5. 

J. P. Taylor, R. H. J. Brown and D. W. Cleveland, “Decoding ALS: from genes to mechanism,” Nature, 539 (7628), 197 –206 https://doi.org/10.1038/nature20413 (2016). Google Scholar

6. 

J. M. Long and D. M. Holtzman, “Alzheimer disease: an update on pathobiology and treatment strategies,” Cell, 179 (2), 312 –339 https://doi.org/10.1016/j.cell.2019.09.001 CELLB5 0092-8674 (2019). Google Scholar

7. 

J. Zhang, B. Long and A. Li, “Whole-brain three-dimensional profiling reveals brain region specific axon vulnerability in 5xFAD mouse model,” Front. Neuroanat., 14 608177 https://doi.org/10.3389/fnana.2020.608177 (2020). Google Scholar

8. 

H. Gong, D. Xu and J. Yuan, “High-throughput dual-colour precision imaging for brain-wide connectome with cytoarchitectonic landmarks at the cellular level,” Nat. Commun., 7 12142 https://doi.org/10.1038/ncomms12142 NCAOBW 2041-1723 (2016). Google Scholar

9. 

T. Ragan, L. R. Kadiri and K. U. Venkataraju, “Serial two-photon tomography for automated ex vivo mouse brain imaging,” Nat. Methods, 9 (3), 255 –258 https://doi.org/10.1038/nmeth.1854 1548-7091 (2012). Google Scholar

10. 

L. Silvestri, A. Bria and L. Sacconi, “Confocal light sheet microscopy: micron-scale neuroanatomy of the entire mouse brain,” Opt. Express, 20 (18), 20582 –20598 https://doi.org/10.1364/OE.20.020582 OPEXFF 1094-4087 (2012). Google Scholar

11. 

S. Liu, J. Nie and Y. Li, “Three-dimensional, isotropic imaging of mouse brain using multi-view deconvolution light sheet microscopy,” J. Innov. Opt. Heal. Sci., 10 (5), 1743006 https://doi.org/10.1142/S1793545817430064 (2017). Google Scholar

12. 

P. Osten and T. W. Margrie, “Mapping brain circuitry with a light microscopy,” Nat. Methods, 10 (6), 515 –523 https://doi.org/10.1038/nmeth.2477 1548-7091 (2013). Google Scholar

13. 

D. Cyranoski, “China launches brain-imaging factory,” Nat. News, 548 (7667), 268 –269 https://doi.org/10.1038/548268a (2017). Google Scholar

14. 

E. Meijering, “Neuron tracing in perspective,” Cytometry Part A, 77 (7), 693 –704 https://doi.org/10.1002/cyto.a.20895 1552-4922 (2010). Google Scholar

15. 

A. Li, H. Gong and B. Zhang, “Micro-optical sectioning tomography to obtain a high-resolution atlas of the mouse brain,” Science, 330 (6009), 1404 –1408 https://doi.org/10.1126/science.1191776 SCIEAS 0036-8075 (2010). Google Scholar

16. 

H. Gong, S. Zeng and C. Yan, “Continuously tracing brain-wide longdistance axonal projections in mice at a one-micron voxel resolution,” NeuroImage, 74 87 –98 https://doi.org/10.1016/j.neuroimage.2013.02.005 NEIMEF 1053-8119 (2013). Google Scholar

17. 

M. N. Economon, N. G. Clack and L. D. Levis, “A platform for brainwide imaging and reconstruction of individual neurons,” Elife, 5 e10566 https://doi.org/10.7554/eLife.10566 (2016). Google Scholar

18. 

M. Halavi, K. A. Hamilton and R. Parekh, “Digital reconstructions of neuronal morphology: three decades of research trends,” Front. Neurosci., 6 49 https://doi.org/10.3389/fnins.2012.00049 (2012). Google Scholar

19. 

R. Parekh and G. Ascoli, “Neuronal morphology goes digital: a research hub for cellular and system neuroscience,” Neuron, 78 (1), 206 https://doi.org/10.1016/j.neuron.2013.03.015 NERNET 0896-6273 (2013). Google Scholar

20. 

S. Li, T. Quan and H. Zhou, “Identifying weak signals in inhomogeneous neuronal images for large-scale tracing of sparsely distributed neurites,” Neuroinformatics, 17 497 –514 https://doi.org/10.1007/s12021-018-9414-9 1539-2791 (2019). Google Scholar

21. 

Z. Zhou, H. C. Kuo and H. Peng, “DeepNeuron: an open deep learning toolbox for neuron tracing,” Brain Inf., 5 (2), 3 https://doi.org/10.1186/s40708-018-0081-2 (2018). Google Scholar

22. 

H. Peng, M. Hawrylycz and J. Roskams, “BigNeuron: large-scale 3D neuron reconstruction from optical microscopy images,” Neuron, 87 252 –256 https://doi.org/10.1016/j.neuron.2015.06.036 NERNET 0896-6273 (2015). Google Scholar

23. 

K. Svoboda, “The past, present, and future of single neuron reconstruction,” Neuroinformatics, 9 (2), 97 –98 https://doi.org/10.1007/s12021-011-9097-y 1539-2791 (2011). Google Scholar

24. 

J. Lu, “Neuronal tracing for connectomic studies,” Neuroinformatics, 9 (2-3), 159 –166 https://doi.org/10.1007/s12021-011-9101-6 1539-2791 (2011). Google Scholar

25. 

L. Feng, T. Zhao and J. Kim, “NeuTube 1.0: a new design for efficient neuron reconstruction software based on the SWC format,” eNeuro, 2 (1), ENEURO.0049-14.2014 https://doi.org/10.1523/ENEURO.0049-14.2014 (2015). Google Scholar

26. 

H. Chen, H. Xiao and T. Liu, “SmartTracing: self-learning-based neuron reconstruction,” Brain Inf., 2 (3), 135 –144 https://doi.org/10.1007/s40708-015-0018-y (2015). Google Scholar

27. 

Y. LeCun, Y. Bengio and G. Hinton, “Deep learning,” Nature, 521 (7553), 436 –444 https://doi.org/10.1038/nature14539 (2011). Google Scholar

28. 

A. Krizhevsky, L. Sutskever and G. Hinton, “ImageNet classification with deep convolutional neural networks,” in Proc. 25th Int. Conf. Neural Inf. Process. Syst., 1097 –1105 (2012). Google Scholar

29. 

J. Long, E. Shelhamer and T. Darrell, “Fully convolutional networks for semantic segmentation,” IEEE Trans. Pattern Anal. Mach. Intell., 39 (4), 640 –651 (2015). Google Scholar

30. 

O. Ronneberger, P. Fischer and T. Brox, “U-Net: convolutional networks for biomedical image segmentation,” Lect. Notes Comput. Sci., 9351 234 –241 https://doi.org/10.1007/978-3-319-24574-4_28 LNCSD9 0302-9743 (2015). Google Scholar

31. 

Q. Dou et al., “3D deeply supervised network for automated segmentation of volumetric medical images,” Med. Image Anal., 41 40 –54 https://doi.org/10.1016/j.media.2017.05.001 (2017). Google Scholar

32. 

H. Chen et al., “VoxResNet: Deep voxelwise residual networks for brain segmentation from 3D MR images,” NeuroImage, 170 446 –455 https://doi.org/10.1016/j.neuroimage.2017.04.041 NEIMEF 1053-8119 (2018). Google Scholar

33. 

X. Wu, Y. Tao and G. He, “Boosting multilabel semantic segmentation for somata and vessels in mouse brain,” Front. Neurosci., 15 610122 https://doi.org/10.3389/fnins.2021.610122 1662-453X (2021). Google Scholar

34. 

R. Li, T. Zeng and S. Ji, “Deep learning segmentation of optical microscopy images improves 3D neuron reconstruction,” IEEE Trans. Med. Imaging, 36 (7), 1533 –1541 https://doi.org/10.1109/TMI.2017.2679713 ITMID4 0278-0062 (2017). Google Scholar

35. 

B. Yang, W. Chen and Y. Wang, “Neuron image segmentation via learning deep features and enhancing weak neuronal structures,” IEEE J. Biomed. Health Inf., 25 (5), 1634 –1645 https://doi.org/10.1109/JBHI.2020.3017540 (2021). Google Scholar

36. 

B. Yang, M. Liu and E. Meijering, “Structure-guided segmentation for 3D neuron reconstruction,” IEEE Trans. Med. Imaging, 41 (4), 903 –914 https://doi.org/10.1109/TMI.2021.3125777 ITMID4 0278-0062 (2022). Google Scholar

37. 

M. Dong, D. Liu and F. Wu, “3D CNN-based soma segmentation from brain images at single-neuron resolution,” in 25th IEEE Int. Conf. Image Process. (ICIP), 126 –130 (2018). https://doi.org/10.1109/ICIP.2018.8451389 Google Scholar

38. 

Q. Li and L. Shen, “3D neuron reconstruction in tangled neuronal image with deep networks,” IEEE Trans. Med. Imaging, 39 (2), 425 –435 https://doi.org/10.1109/TMI.2019.2926568 ITMID4 0278-0062 (2020). Google Scholar

39. 

T. Jerman and B. Likar, “Enhancement of vascular structures in 3D and 2D angiographic images,” IEEE Trans. Med. Imaging, 35 (9), 2107 –2118 https://doi.org/10.1109/TMI.2016.2550102 ITMID4 0278-0062 (2016). Google Scholar

40. 

R. F. Frangi, W. J. Niessen and K. L. Vincken, “Multiscale vessel enhancement filtering,” Lect. Notes Comput. Sci., 1496 130 –137 https://doi.org/10.1007/BFb0056195 LNCSD9 0302-9743 (1998). Google Scholar

41. 

Z. Zhou, S. Sorensen and H. Peng, “Adaptive image enhancement for tracing 3D morphologies of neurons and brain vasculatures,” Neuroinformatics, 13 (2), 153 –166 https://doi.org/10.1007/s12021-014-9249-y 1539-2791 (2015). Google Scholar

42. 

X. Wu, S. Chen and J. Huang, “DDeep3M: Docker-powered deep learning for biomedical image segmentation,” J. Neurosci. Methods, 342 108804 https://doi.org/10.1016/j.jneumeth.2020.108804 JNMEDT 0165-0270 (2020). Google Scholar

43. 

M. G. Haberl, C. Churas and L. Tindall, “CDeep3M–Plug-and-Play cloud-based deep learning for image segmentation,” Nat. Methods, 15 677 –680 https://doi.org/10.1038/s41592-018-0106-z 1548-7091 (2018). Google Scholar

44. 

Q. Huang, J. Sun and H. Ding, “Robust liver vessel extraction using 3D U-Net with variant dice loss function,” Comput. Biol. Med., 101 153 –162 https://doi.org/10.1016/j.compbiomed.2018.08.018 CBMDAW 0010-4825 (2018). Google Scholar

45. 

T. Falk, D. Mai and R. Bensch, “U-Net: deep learning for cell counting, detection, and morphometry,” Nat. Methods, 16 67 –70 https://doi.org/10.1038/s41592-018-0261-2 1548-7091 (2019). Google Scholar

46. 

Q. Huang, Y. Chen and S. Liu, “Weakly supervised learning of 3D deep network for neuron reconstruction,” Front. Neuroanat., 14 38 https://doi.org/10.3389/fnana.2020.00038 (2020). Google Scholar

47. 

J. Zhao, X. Chen and Z. Xiong, “Neuronal population reconstruction from ultra-scale optical microscopy images via progressive learning,” IEEE Trans. Med. Imaging, 39 (12), 4034 –4046 https://doi.org/10.1109/TMI.2020.3009148 ITMID4 0278-0062 (2020). Google Scholar

48. 

X. Chen, C. Zhang and J. Zhao, “Weakly supervised neuron reconstruction from optical microscopy images with morphological priors,” IEEE Trans. Med. Imaging, 40 (11), 3205 –3216 https://doi.org/10.1109/TMI.2021.3080695 ITMID4 0278-0062 (2020). Google Scholar

Biographies of the authors are not available.

CC BY: © The Authors. Published by SPIE under a Creative Commons Attribution 4.0 International License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.
Rong Xiao, Lei Zhu, Jiangshan Liao, Xinglong Wu, Hui Gong, Jin Huang, Ping Li, Bin Sheng, and Shangbin Chen "DDeep3M+: adaptive enhancement powered weakly supervised learning for neuron segmentation," Neurophotonics 10(3), 035003 (23 June 2023). https://doi.org/10.1117/1.NPh.10.3.035003
Received: 25 March 2023; Accepted: 6 June 2023; Published: 23 June 2023
Advertisement
Advertisement
KEYWORDS
Image segmentation

Neurons

Education and training

Machine learning

Voxels

Image enhancement

3D image processing

Back to Top