Next Article in Journal
Data Quality Analysis and Improvement: A Case Study of a Bus Transportation System
Previous Article in Journal
Cultural Influences on Saccadic Patterns in Facial Perception: A Comparative Study of American and Japanese Real and Animated Faces
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

TTH-Net: Two-Stage Transformer–CNN Hybrid Network for Leaf Vein Segmentation

Tongda College of Nanjing University of Posts & Telecommunications, Yangzhou 225127, China
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2023, 13(19), 11019; https://doi.org/10.3390/app131911019
Submission received: 4 September 2023 / Revised: 30 September 2023 / Accepted: 4 October 2023 / Published: 6 October 2023

Abstract

:

Featured Application

Leaf vein segmentation work can be applied in plant species classification, drought resistance research in plants, and climate and environmental analysis.

Abstract

Leaf vein segmentation is crucial in species classification and smart agriculture. The existing methods combine manual features and machine learning techniques to segment coarse leaf veins. However, the extraction of the intricate patterns is time consuming. To address the issues, we propose a coarse-to-fine two-stage hybrid network termed TTH-Net, which combines a transformer and CNN to accurately extract veins. Specifically, the proposed TTH-Net consists of two stages and a cross-stage semantic enhancement module. The first stage utilizes the Vision Transformer (base version) to extract globally high-level feature representations. Based on these features, the second stage identifies fine-grained vein features via CNN. To enhance the interaction between the two stages, a cross-stage semantic enhancement module is designed to integrate the strengths of the transformer and CNN, which also improves the segmentation accuracy of the decoder. Extensive experiments on the public dataset LVN are conducted, and the results prove that TTH-Net has significant advantages over other methods in leaf vein segmentation.

1. Introduction

Leaf vein segmentation plays a crucial role in the fields of botany and biology due to the important physiological functions and unique morphological characteristics of leaf veins. By segmenting leaf vein patterns, it can further enhance the accuracy of plant species classification [1], facilitate the development of smart agriculture [2], and contribute to research on climate condition analysis [3]. Similar to the vascular network in human beings, leaf vein patterns exhibit complex distributions and strong connectivity. Thicker veins are responsible for providing support to the leaf, running through the entire leaf, while thinner veins are responsible for nutrient transport, hence exhibiting strong connectivity. There are close connections and numerous loops between them, which pose significant challenges for extraction and segmentation tasks.
To segment leaf vein patterns, early works mainly rely on chemical reagents [4] and scanners [5] to extract veins. Lersten et al. [4] used a chemical method where the leaf tissue was cleaned, leaving behind the vein pattern. Larese et al. [5] utilized the scanner to obtain high-resolution leaf images, and then extracted the vein pattern via unconstrained hit-or-miss transforms and adaptive thresholds. Limited by image processing techniques, extraction steps are time consuming, and the results are difficult to reproduce. To enhance the automation level of leaf vein extraction, Price et al. [6] designed the LEAF GUI software specifically for vein extraction. This software captures leaf images and returns statistical data about the vein network and stomatal structures via a series of interactive thresholding and cleaning steps, which further deepens users’ understanding of leaf vein networks. With the development of image processing techniques, more and more machine learning-based algorithms have been employed for leaf vein segmentation. Sibi Chakkaravarthy et al. [7] used the angle between the main vein and secondary veins and the centroid vein angle as key parameters, combined with Hough lines, to extract the leaf veins. Radha et al. [8] utilized the canny edge detection algorithm to determine the nutritional status of plants by detecting the leaf edges and veins. The Scale Invariant Feature Transform (SIFT) operation in [9] was used to extract local features from leaf vein images. These features are considered important references for subsequent plant classification. Compared to the methods based on chemical reagents and scanners, machine learning-based algorithms rely on manually crafted feature extractors and have limited automation levels. With the remarkable performance of the FCN network in the field of segmentation, a lot of CNN-based deep learning algorithms [10,11,12,13] have been extended to the domain of image segmentation. Xu et al. [14] employed existing CNN models to extract leaf veins and achieved impressive performance. Convolutional Neural Networks (CNNs) utilize small convolutional kernels that slide over the image to capture local information and gradually build receptive fields. By stacking multiple convolutional layers, CNNs can aggregate and learn hierarchical representations, enabling them to capture both low-level and high-level features from the input image. However, due to the limitation of the receptive field, the continuity of the veins may be compromised.
To address the aforementioned issues, inspired by the recent success of Transformers in natural language processing (NLP), this paper leverages Transformers to enhance the accuracy of vein extraction. The core of the Transformer is the self-attention module. By extracting pixel relationships across the entire image, attention coefficients based on global receptive fields can be obtained. Multiplying this probability map with the original image via matrix multiplication significantly enhances the global dependency in the feature maps. Recently, related studies [15,16,17] have already demonstrated the advantage of Transformers over CNNs in modeling global context. Vision Transformer (ViT) [15] has demonstrated its effectiveness compared to CNN by dividing the image into patches and flattening them into vectors before inputting them into the Transformer. This approach allows ViT to capture global relationships and context, which is key to the transformer outperforming CNN. Based on ViT, Segmenter [16] designed a pure Transformer model, which is specifically designed for semantic segmentation tasks. Compared to convolution-based methods, Segmenter allows the modeling of global context from the first layer. This approach makes Segmenter highly competitive in the scene-understanding datasets. Additionally, SETR [17], a variant of the ViT model, introduces the global characteristics of ViT and conducts extensive experiments to prove the superior performance of the Transformer compared to CNN in semantic segmentation tasks. However, SETR retains the decoding approach of CNN in the decoder to fully leverage the excellent capability of CNN for understanding local details.
To achieve high-accuracy segmentation performance, this paper combines the strengths of the Transformer and CNN to design a coarse-to-fine model termed a two-stage transformer–CNN hybrid network (TTH-Net). Specifically, in the first stage, TTH-Net utilizes the backbone of ViT [15] to extract feature representations that contain rich global context. This stage is responsible for generating coarse-grained feature representations. Based on these features, TTH-Net employs CNN in the second stage for fine-grained recovery of the features generated by the ViT. Furthermore, to enhance the accuracy of the decoder, TTH-Net introduces a cross-stage semantic enhancement (CSE) module to connect the features generated by the encoders and decoders of both stages. Different from other two-stage networks, the CSE module fully utilizes the features generated in both stages and effectively integrates the advantages of Transformer and CNN, which results in the best performance of TTH-Net among all the compared methods.
In summary, our contributions are as follows:
(1)
We expand the fusion of CNN and transformer and propose a coarse-to-fine approach that combines the advantages of Transformer and CNN, which is termed a two-stage hybrid network (TTH-Net) for leaf vein segmentation.
(2)
We introduce the CSE module to guide the feature fusion between the CNN and transformer. This approach breaks away from the traditional stage-wise cascade and significantly enhances the segmentation accuracy of the decoder.
(3)
Extensive experiments conducted on the LVN dataset and our proposed TTH-Net achieve stage-of-the-art performance. Additionally, corresponding ablation experiments validate the necessity of each stage and the CSE module.
The rest of the sections in this paper are as follows. In Section 2, the typical algorithms for the leaf vein segmentation task and two-stage design are reviewed. Section 3 introduces the components of the proposed TTH-Net method. Section 4 presents the experimental results and discussions. Finally, conclusions are drawn in Section 5.

2. Related Work

In this section, we first summarize the work on leaf vein segmentation and deeply analyze the existing algorithms. Additionally, the applications of two-stage algorithms are also introduced. Finally, the motivations of our work are listed.
Leaf vein segmentation. The leaf vein segmentation task is crucial for applications in botany and biology. Many efforts [6,18,19,20,21,22,23] have been deliciated to improve the accuracy and the efficiency. Bühler et al. [18] were dedicated to automatically segmenting and analyzing leaf veins in images obtained via different imaging modalities such as microscopy and macro photography. It has designed a powerful and user-friendly image analysis tool called phenoVein, which greatly facilitates the study of plant phenotypes in ecology and genetics research. Kirchgeßner et al. [19] represented leaf veins as B-splines that contain hierarchical information about the vascular system. He accomplishes structure tracking by considering spatial information. The algorithm samples the previously estimated affine growth rates along the plant veins and perpendicular to them, representing them in physiological coordinates. To efficiently segment leaf vein patterns, Price et al. [6] designed a user-assisted software tool termed Leaf Extraction and Analysis Framework Graphical User Interface (LEAF GUI) to enhance researchers’ understanding of leaf network structures. The LEAF GUI considers the size, position, and connectivity of network veins, as well as the size, shape, and position of surrounding pores. It follows a series of interactive thresholding and cleaning steps to return statistical data and information about the leaf vein network. Blonder et al. [20] developed theories for absorption- and phase-contrast X-ray imaging. By utilizing a synchrotron accelerator light source and two commercial X-ray instruments, they successfully segment the major and minor veins. Salima et al. [21] considered leaf vein structure a crucial biological feature and combined it with the Hessian matrix to perform grouped structure segmentation of images. By analyzing the eigenvalues of pixels, the segmentation results based on the characteristics of the leaf vein structure were obtained. Katyal et al. [22] proposed a new leaf vein segmentation method that utilizes odd Gabor filters and morphological operations to achieve more precise vein segmentation. Compared to other algorithms, the Odd Gabor filter can provide more precise outputs due to its ability to effectively detect fine fibrous structures present in leaves, such as tertiary vein branches. Recently, with the development of CNN, Xu et al. [14] utilized a convolutional neural network (CNN) to end-to-end segment leaf veins, which is automatic and accurate. While CNNs have significantly improved the automation level and achieved high accuracy in leaf vein segmentation, most current methods still rely on machine learning algorithms. Therefore, there is great potential for utilizing deep learning to further boost performance.
Two-stage algorithm. The two-stage algorithms are widely used to solve image processing problems such as Image Super-Resolution [24], single-image reflection removal [25], segmentation task [26,27], intrusion detection [28], eigenvalues Detection [29], image restoration [30] and image denoising [31,32]. For example, Fan et al. [24] achieved the optimal balance between accuracy and speed by employing a deep residual design with constrained depth and combining the two-stage structure with the lightweight two-layer PConv residual block design. Li et al. [25] proposed a novel two-stage network, named the Reflection-Aware Guided Network (RAGNet), for single-image reflection removal tasks. This approach alleviates the limitations of most existing methods in utilizing the results from the previous stage to guide transmission estimation. Wang et al. [26] proposed a two-stage segmentation network to incorporate the global and local image region fitting energies by using popular active contours. The first stage of the network uses the Gaussian distribution for coarse features, while the second stage refines the local details based on the results obtained in the previous stage, which are effective for synthetical and real-world images. Ong et al. [27] designed a two-stage hierarchical artificial neural network. The first stage was responsible for capturing the main colors from the fixed feature map in an unsupervised manner, while the second stage controlled the number of color clusters for segmentation quality improvement. For single-image super-resolution, Chen et al. [30] proposed a two-step restoration approach to predict missing region features when object features are severely missing. This method significantly improves texture blurring and feature discontinuity issues. Cai et al. [31] utilized a two-phase approach to restore images corrupted by blur and impulse noise by identifying outlier candidates in the first stage and employing variational methods in the second stage. Tian et al. [32] expanded the three-stage network based on CNN by incorporating wavelet transform for image denoising. Although two-stage algorithms are widely applicable to image processing tasks, most existing methods neglect the interaction between the two stages, which results in the limited performance of the models.
In this paper, our proposed two-stage hybrid network (TTH-Net) integrates the strengths of both transformer and CNN, which also achieves state-of-the-art performance in the leaf vein segmentation task. The proposed TTH-Net utilizes deep learning methods to significantly improve the level of automation in vein segmentation while achieving high accuracy. Additionally, different from other two-stage methods, we extensively interact with the features generated from both stages and guide the representation of long-range features via auxiliary loss functions.

3. The Proposed Method

In this section, we propose a coarse-to-fine two-stage Transformer–CNN hybrid network (TTH-Net) for leaf vein segmentation. The proposed TTH-Net consists of two stages and the cross-stage semantic enhancement (CSE) module, as shown in Figure 1. The first stage utilizes ViT to extract global contexts, while the second stage employs features generated by Transformer to guide CNN for fine vein recovery. Additionally, the CSE module is introduced to facilitate the interaction between the two stages and further improve the segmentation accuracy of the decoder.

3.1. First Stage: Coarse-Grained Feature Extraction

To model the long-range dependencies of leaf veins, the Vision Transformer (ViT) [15] is used for more representative features. Different from CNN, ViT expands the receptive field based on the entire leaf image by self-attention module, which results in strong global modeling capability. Thus, in this paper, we take ViT as the encoder to capture coarse-grained vein features.
Specifically, the ViT encoder consists of self-attention modules, Multi-Layer Perceptron (MLP), and LayerNorm. The input X R H × W × C (H, W, and C are height, width, and channel dimensions, respectively) is first split into N flattened patches, where N = H × W / P 2 and P is the width/height of each patch. As a result, the original input is reshaped into X p R N × P 2 × C . By flattening these patches into one-dimensional vectors and projecting them linearly, we obtain the input E v of ViT. To compensate for the lost positional information during the patching, E p is added into E v and forms a new input z . After the LayerNorm function and linear mapping operation, we obtain the Query (Q), Key (K), and Value (V). Then, we utilize the self-attention module to model the global dependency of the vectors for accuracy improvements, which can be calculated as
A t t e n t i o n ( Q ,   K ,   V ) = S o f t m a x ( Q · K T d ) V
where d represents the dimension of feature vectors. After the residual connection, z s is generated. The LayerNorm function and MLP are used to integrate contexts and output z s , which can be formulated as
z s = M L P ( L N ( z s ) ) + z s
Finally, by repeating the above process, we obtain the coarse-grained results. To facilitate the recovery of fine-grained features in subsequent steps, we retained the results from the third, sixth, ninth, and twelfth layers and reshaped them into the 2D features. Note that these features have the same resolutions. In addition, we also introduce four additional auxiliary losses to guide ViT in extracting coarse-grained feature representations.

3.2. Second Stage: Fine-Grained Feature Recovery

To enhance the TTH-Net of recognizing fine-grained features, we expand ViT in stage one by cascading a U-shaped CNN. Although transformers excel at modeling global context, fine-grained vein recognition is equally important for leaf vein segmentation tasks. The fine-grained feature restoration network consists of the encoder and decoder. The outputs { z 3 , z 6 , z 9 , z 12 } from stage one is set as the inputs of stage two. Specifically, z 3 is first sent to stage two and convolved by 3 × 3 convolution, Batch Normalization (BN), and ReLU function. Then, z 6 2 is acquired from z 3 and further stacked together with z 6 1 , which is downsampled by z 6 . After the 3 × 3 convolution, BN, and ReLU function, the new feature map z 6 3 can be extracted more vein-related details. By repeating the above process, our proposed TTH-Net can generate more advanced representations. For the decoding part, bilinear interpolation functions are used to restore the vein semantics. To better combine the advantages of Transformer and CNN, we introduce the cross-stage semantic enhancement (CSE) module [33] to facilitate the cross-stage fusion of representations between long-distance features. For example, the output of the encoder z 12 3 and the feature z 12 are sent into the CSE module and generates z 9 4 . By concatenating z 9 4 with z 9 3 , the new feature maps z 9 5 are obtained after the 3 × 3 convolution, BN, and ReLU function. By stacking these operations, the target classes in the features are divided by adding a convolution at the end of the network. More details about the CSE module are introduced in the following subsection.

3.3. Cross-Stage Semantic Enhancement Module

The multi-stage approach produces increasingly refined estimation results by extending the depth of the network and the number of paths for information flow. However, there are semantic gaps in the cross-stage features, and simply adding them together not only fails to improve the performance but also introduces background noise to perturb the segmentation results. To bridge the semantic gap between the different stages, we designed the cross-stage semantic enhancement (CSE) module [33] to efficiently model the long-range dependencies and improve the model performance, which is shown in Figure 2.
The CSE module consists of three main branches, B 1 , B 2 , and B 3 , and each branch is implemented by the corresponding module for cross-stage semantic enhancement. Specifically, SSAM is used to guide the deeper stage to aggregate high-resolution features, and CAM is applied to reweight channels. In addition, B 3 utilizes the LRM to generate refined features, which are provided to guide SSAM and CAM to fuse the different scale features from B 1 and B 2 .
Formality, the coarse feature maps from branch B 1 are defined as G 1 , and G 2 is the low-level feature maps from branch B 2 . We first use 1 × 1 convolution f 1 ( · ) on G 1 to obtain G ^ 1 , and G ^ 2 is obtained by downsampling G 2 . By concatenating G ^ 1 with G ^ 2 , G 3 is obtained. We send G 3 into LRM and obtain the refined features G ^ 3 . Then, G ^ 1 is convolved by 3 × 3 convolution f 3 ( · ) and CAM employs global average pooling to extract global context from G ^ 3 . By multiplying their results, we obtain G 5 , which is filtered for irrelevant information. G 1 is calculated as follows:
G 1 = ξ C ξ L (   G ^ 3   )   f 1 f 3 G 1
where ξ C and ξ L represent results generated by CAM and LRM, respectively. Parallelly, we use 3 × 3 convolution on G ^ 2 , and SSAM is applied to aggregate discriminative contextual information. Based on the results, we obtain G 6 via element-wise multiplication. G 2 is given in Equation (4).
G 2 = ξ S ξ L ( G ^ 3 ) f 1 f 3 G 2
where ξ S is the feature map generated by SSAM. Finally, the output G o u t is obtained by fusing G 1 and G 2 .
G o u t = G 1 + G 2

3.4. Loss Function

The total loss of the TTH-Net includes two parts, one part being the features from multiple stage outputs in the ViT of the first stage. Via the MLP module and upsampling operations, these auxiliary losses guide the Transformer in recognizing coarse-grained features. The other part of the loss is the difference between the output of the second stage and the labels. By reweighting the two parts of the losses, TTH-Net can accurately segment leaf veins. The loss is calculated as follows:
L o s s ( x , y ) = λ · D L ( x , y ) + ( 1 λ ) · C E ( x , y )
where CE ( · ) is the cross-entropy loss function [34], D L ( · ) represents the dice loss function [35], and the hyperparameter λ controls the fusion coefficient of the loss functions.

4. Experiment

4.1. Dataset

The public Leaf Venation Network (LVN) dataset [36] is collected by images of leaf venation networks from tree species in Malaysian Borneo. It consists of 726 leaf samples from 295 species belonging to 50 families in eight forest plots of the state of Sabah. These selected plots span across a gradient of logging intensity and are mixed lowland forests of different regeneration stages. Each plot has a size of 1 hectare, and all tree stems with a diameter at breast height (DBH) ≥ 10 cm are marked. In addition, we also employ a stratified random design to sample trees to capture rare and smaller species. Within each permanent plot, three 20 × 20 m subplots are selected, and all trees within the subplots are sampled. Leaves are collected from each stem under both sunlight and shade conditions. The image resolution is approximately 50 million pixels. Limited by the GPU memory, we use the sliced images from the LVN dataset as training samples. The original image and its cropped patch are shown in Figure 3.
For the ground truth images, the LVN dataset provides labels corresponding to the cropped images. A compound microscope equipped with a 2× de-color objective and a color camera with 3840 × 2748 resolution is used to capture images of each leaf. The complete image of the sampled area is obtained by digitally stitching together 9–16 overlapping image fields per sample using a digital tablet and GIMP software (GNU). Subsequently, the images are preprocessed using MATLAB by retaining the green channel with the maximum contrast after staining and applying contrast-limited adaptive histogram equalization. All textures were manually annotated, as shown in Figure 4. Further details of the processing can be found in [36].

4.2. Evaluation Metrics

To comprehensively evaluate the experimental results, we selected four metrics: Dice coefficient ( D i c e ) [37], Sensitivity ( S e n ) [38], Specificity (Spe), and Accuracy ( A c c ) [39], which are commonly used in segmentation tasks, to test the performance of the proposed model. The indicator D i c e is a commonly used similarity metric for evaluating image segmentation results. It measures the overlap between the predicted result and the ground truth label. The S e n metric measures the model’s capability to correctly identify positive instances while the A c c estimating the proportion of samples that the model classifies correctly, which are calculated as follows:
D i c e = 2 P r e d i c i t i o n G r o u n d t r u t h P r e d i c t i o n + G r o u n d t r u t h
S p e c i f i c i t y = T N T N + F P
S e n = T P T P + F N
A c c = T P + T N T P + F N + T N + F P
where T P , T N , F P , and F N represent the number of true positives, true negatives, false positives, and false negatives, respectively.

4.3. Implementation Details

We implement the proposed method based on the Pytorch framework and the Ubuntu 18.04 system. The NVIDIA GeForce GTX 3090 graphics cards are used to accelerate the training process, which has 24 Gigabyte memory. The batch size and total iteration number are set to 8 and 120,000, respectively, for the experiments on the LVN dataset. The polynomial learning rate decay schedule and SGD optimizer are employed. The initial learning rate and the weight decay are set to 0.008 and 0.9, respectively. All the images perform data augmentation, including random angle rotation, random translation, random zooming, random shearing, and horizontal flipping with ratios between 0.5 and 2. In addition, we employ a sliding window approach [40] for both the training and testing phases. Each image was cropped into patches of size 256 × 256 with a stride of 128 × 128. The undersized regions are filled with a pixel value of 255.
For the first stage of the transformer encoder, we select the base version as our backbone. Specifically, the ViT consists of twelve layers and sets the hidden size as 768, MLP size as 3072, and head numbers as twelve. The patch size is set to 16 × 16. The outputs of the third, sixth, ninth, and final layers of ViT are saved and used for fine-grained feature recognition in the second stage. In the second stage, the downsampling operation is used to reduce the resolution by half by setting the stride of convolution as 2. In addition, the bilinear interpolation (the ratio is 2) is adopted to upsample the resolutions of the decoder. Finally, by utilizing 1 × 1 convolution, the final binary results are generated.

4.4. Comparisons with the State of the Art

In this section, we evaluate the performance of the proposed model. For comprehensive comparison, we select CNN models including FCN [10], U-Net [12] GhostNet [13], MobileNetv3 [41], and DeepLabv3+ [42] Transformer models including ViT [15], Segmenter [16], Swin Transformer [43], SETR [17], DeiT [44], as well as two-stage models, such as ALSNet [45], TSUNet [46], DDNet [47], APCNet [48], CMMANet [49], and PCDNet [50] as benchmark methods. For the GhostNet and ViT models, we preserve their encoding parts and replace the original classification head with a 1 × 1 convolution. For training and testing, all models are conducted under the same settings.
As shown in Table 1, the results of variants of CNN and transformer, as well as the outcomes of the two-stage algorithm are recorded. For leaf vein segmentation, the transformer has significant advantages over CNNs due to its global modeling capability. In addition, the two-stage algorithms based on the deep learning methods also perform well, and our proposed TTH-Net achieves the best scores in all the indicators. Specifically, compared to the DeepLabv3+ algorithm, our proposed TTH-Net surpasses it by 2.73%, 1.05%, 2.30%, and 1.89% in terms of IoU, Spe, Sen, and Acc metrics, respectively. Additionally, TTH-Net achieves an improvement of 2.1%, 1.19%, 0.61%, and 1.15% in IoU, Spe, Sen, and Acc indicators, respectively, compared to the Swin Transformer. We also record the mean (standard deviation) of the 5-fold cross-validation of each metric for the proposed method. As shown in Table 1, transformer variants such as Segmenter and DeiT exhibit stronger dependency on the data compared to CNNs. Additionally, the proposed TTH-Net, which integrates the advantages of CNN and transformer, shows a lower dependency on the data. Thus, we can conclude that the proposed TTH-Net has state-of-the-art performance on the LVN dataset for the leaf vein segmentation task and lower dependency on the data.
As shown in Figure 5, visual comparisons of the highest-scoring models for CNN, Transformer, and the two-stage algorithm are conducted, respectively. The CNN model, DeepLabv3+, exhibits relatively coarse texture detection of vein patterns, accompanied by numerous isolated pixels and some incorrectly classified curves. For the Swin Transformer, there is a more pronounced issue of misclassifying vein curve pixels as background pixels, which also results in a suboptimal performance. In contrast to our proposed TTH-Net, TTH-Net is capable of segmenting fine-grained vein textures with strong curve continuity. Additionally, the segmentation results exhibit minimal occurrence of isolated pixels. As a result, the proposed TTH-Net can segment veins accurately and achieve state-of-the-art performance for leaf vein segmentation tasks.
Finally, we record the curves of loss vs. number of iterations and accuracy and vs. number of iterations during the training and validation process, as shown in Figure 6. From Figure 6a, we can find that our proposed model shows the consistency between the loss curves during training and validation process, which indicates that TTH-Net is not overfit or underfit. In Figure 6b, as the number of iterations increases, the accuracy of the TTH-Net keeps improving, and the optimal number of iterations is 50,000.

4.5. Ablation Study

To validate the necessity of each stage and module in TTH-Net, we conduct corresponding ablation experiments in this subsection. Specifically, we add the second stage and CSE module on top of the first stage individually and record their performances in Table 2.
Firstly, by adding a segmentation head on top of the first stage, we obtain segmentation results for the first stage with an IoU score of 83.83%, Spe score of 81.57%, Sen score of 86.03%, and Acc score of 87.67%. Then, the CNN module is added in the second stage, and TTH-Net achieves an improvement of 0.77%, 0.77%, 0.94%, and 0.79% in IoU, Spe, Sen, and Acc metrics, respectively. Based on the two stages, the CSE module can bring an increase of 0.77% in IoU, 0.6% in Spe, 0.91% in Sen, and 0.78% in Acc to the TTH-Net. As a result, each stage and the CSE module is necessary for achieving high-precision leaf vein segmentation.
Additionally, we conduct corresponding ablation experiments on the loss function and the hyperparameter λ . As shown in Table 3, we select the commonly used loss functions such as focal loss [51], cross-entropy (CE) loss [34], and dice loss (DL) [35]. As shown in Table 4, among all combinations of loss functions, TTH-Net achieves the best performance when the first stage utilizes the auxiliary loss function of the Cross-Entropy (CE) loss function while the second stage adopts the DL. For the hyperparameter λ , we find that when λ = 0.4 , the TTH-Net achieves the highest scores across all metrics.

5. Conclusions

In this paper, we propose a novel coarse-to-fine algorithm termed two-stage Transformer–CNN hybrid network (TTH-Net) for leaf vein segmentation task. The proposed TTH-Net consists of two stages and a cross-stage semantic enhancement (CSE) module. The first stage utilizes a transformer to obtain the global feature representations while the second stage recognizes the fine-grained features with CNN. To achieve better performance in the decoding stage, a CSE module is added to fuse the advantages of CNN and transformer, and the integrated features are fed into the decoder. In addition, the CSE module can also bridge the semantic gaps in the cross-stage features and efficiently model the long-range dependencies for further accuracy improvements. To evaluate the proposed TTH-Net, extensive experiments are conducted, and the results demonstrate that the proposed TTH-Net has the best performance among all the competitors.

Author Contributions

Conceptualization, P.S.; methodology, Y.Y.; data curation, Y.Z.; writing—original draft preparation, Y.Z.; writing—review and editing, P.S.; visualization, Y.Z.; supervision, Y.Y.; project administration, P.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Tongda College of Nanjing University of Posts and Telecommunications grant No. XK006XZ19013.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data is unavailable due to privacy.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Larese, M.G.; Granitto, P.M. Finding local leaf vein patterns for legume characterization and classification. Mach. Vis. Appl. 2016, 27, 709–720. [Google Scholar] [CrossRef]
  2. Darwin, B.; Dharmaraj, P.; Prince, S.; Popescu, D.E.; Hemanth, D.J. Recognition of bloom/yield in crop images using deep learning models for smart agriculture: A review. Agronomy 2021, 11, 646. [Google Scholar] [CrossRef]
  3. Sack, L.; Scoffoni, C. Leaf venation: Structure, function, development, evolution, ecology and applications in the past, present and future. New Phytol. 2013, 198, 983–1000. [Google Scholar] [CrossRef] [PubMed]
  4. Lersten, N.R. Modified clearing method to show sieve tubes in minor veins of leaves. Stain Technol. 1986, 61, 231–234. [Google Scholar] [CrossRef]
  5. Larese, M.G.; Namías, R.; Craviotto, R.M.; Arango, M.R.; Gallo, C.; Granitto, P.M. Automatic classification of legumes using leaf vein image features. Pattern Recognit. 2014, 47, 158–168. [Google Scholar] [CrossRef]
  6. Price, C.A.; Symonova, O.; Mileyko, Y.; Hilley, T.; Weitz, J.S. Leaf extraction and analysis framework graphical user interface: Segmenting and analyzing the structure of leaf veins and areoles. Plant Physiol. 2011, 155, 236–245. [Google Scholar] [CrossRef]
  7. Sibi Chakkaravarthy, S.; Sajeevan, G.; Kamalanaban, E.; Varun Kumar, K.A. Automatic leaf vein feature extraction for first degree veins. In Proceedings of the Advances in Signal Processing and Intelligent Recognition Systems: Proceedings of the Second International Symposium on Signal Processing and Intelligent Recognition Systems (SIRS-2015), Trivandrum, India, 16–19 December 2015; Springer International Publishing: Berlin/Heidelberg, Germany, 2015; pp. 581–592. [Google Scholar] [CrossRef]
  8. Radha, R.; Jeyalakshmi, S. An effective algorithm for edges and veins detection in leaf images. In Proceedings of the IEEE 2014 World Congress on Computing and Communication Technologies, Trichirappalli, India, 27 February–1 March 2014; pp. 128–131. [Google Scholar] [CrossRef]
  9. Selda, J.D.S.; Ellera, R.M.R.; Cajayon, L.C.; Linsangan, N.B. Plant identification by image processing of leaf veins. In Proceedings of the International Conference on Imaging, Signal Processing and Communication, Penang, Malaysia, 26–28 July 2017; pp. 40–44. [Google Scholar] [CrossRef]
  10. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar] [CrossRef]
  11. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  12. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, 5–9 October 2015; Proceedings, Part III 18; Springer International Publishing: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar] [CrossRef]
  13. Han, K.; Wang, Y.; Tian, Q.; Guo, J.; Xu, C.; Xu, C. Automated and accurate segmentap operations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, Washington, USA, 14–19 June 2020; pp. 1580–1589. [Google Scholar] [CrossRef]
  14. Xu, H.; Blonder, B.; Jodra, M.; Malhi, Y.; Fricker, M. Automated and accurate segmentation of leaf venation networks via deep learning. New Phytol. 2021, 229, 631–648. [Google Scholar] [CrossRef]
  15. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar] [CrossRef]
  16. Strudel, R.; Garcia, R.; Laptev, I.; Schmid, C. Segmenter: Transformer for semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 7262–7272. [Google Scholar] [CrossRef]
  17. Zheng, S.; Lu, J.; Zhao, H.; Zhu, X.; Luo, Z.; Wang, Y.; Fu, Y.; Feng, J.; Xiang, T.; Torr, P.H.S.; et al. Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 6881–6890. [Google Scholar] [CrossRef]
  18. Bühler, J.; Rishmawi, L.; Pflugfelder, D.; Huber, G.; Scharr, H.; Hülskamp, M.; Koornneef, M.; Schurr, U.; Jahnke, S. phenoVein—A tool for leaf vein segmentation and analysis. Plant Physiol. 2015, 169, 2359–2370. [Google Scholar] [CrossRef] [PubMed]
  19. Kirchgeßner, N.; Scharr, H.; Schurr, U. Robust vein extraction on plant leaf images. In Proceedings of the 2nd IASTED International Conference Visualization, Imaging and Image Processing, Málaga, Spain, 9–12 September 2002. [Google Scholar]
  20. Blonder, B.; De Carlo, F.; Moore, J.; Rivers, M.; Enquist, B.J. X-ray imaging of leaf venation networks. New Phytol. 2012, 196, 1274–1282. [Google Scholar] [CrossRef]
  21. Salima, A.; Herdiyeni, Y.; Douady, S. Leaf vein segmentation of medicinal plant using hessian matrix. In Proceedings of the IEEE 2015 International Conference on Advanced Computer Science and Information Systems (ICACSIS), Depok, Indonesia, 10–11 October 2015; pp. 275–279. [Google Scholar] [CrossRef]
  22. Katyal, V. Leaf vein segmentation using Odd Gabor filters and morphological operations. arXiv 2012, arXiv:1206.5157. [Google Scholar] [CrossRef]
  23. Saleem, R.; Shah, J.H.; Sharif, M.; Yasmin, M.; Yong, H.S.; Cha, J. Mango leaf disease recognition and classification using novel segmentation and vein pattern technique. Appl. Sci. 2021, 11, 11901. [Google Scholar] [CrossRef]
  24. Fan, Y.; Shi, H.; Yu, J.; Liu, D.; Han, W.; Yu, H.; Wang, Z.; Wang, X.; Huang, T.S. Balanced two-stage residual networks for image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 161–168. [Google Scholar] [CrossRef]
  25. Li, Y.; Liu, M.; Yi, Y.; Li, Q.; Ren, D.; Zuo, W. Two-stage single image reflection removal with reflection-aware guidance. Appl. Intell. 2023, 53, 19433–19448. [Google Scholar] [CrossRef]
  26. Wang, H.; Huang, T.Z.; Xu, Z.; Wang, Y. A two-stage image segmentation via global and local region active contours. Neurocomputing 2016, 205, 130–140. [Google Scholar] [CrossRef]
  27. Ong, S.H.; Yeo, N.C.; Lee, K.H.; Venkatesh, Y.V.; Cao, D.M. Segmentation of color images using a two-stage self-organizing network. Image Vis. Comput. 2002, 20, 279–289. [Google Scholar] [CrossRef]
  28. Kaur, R.; Gupta, N. CFS-MHA: A Two-Stage Network Intrusion Detection Framework. Int. J. Inf. Secur. Priv. (IJISP) 2022, 16, 1–27. [Google Scholar] [CrossRef]
  29. Mashta, F.; Altabban, W.; Wainakh, M. Two-Stage Spectrum Sensing for Cognitive Radio Using Eigenvalues Detection. Int. J. Interdiscip. Telecommun. Netw. (IJITN) 2020, 12, 18–36. [Google Scholar] [CrossRef]
  30. Chen, Y.; Xia, R.; Zou, K.; Yang, K. FFTI: Image inpainting algorithm via features fusion and two-steps inpainting. J. Vis. Commun. Image Represent. 2023, 91, 103776. [Google Scholar] [CrossRef]
  31. Cai, J.F.; Chan, R.H.; Nikolova, M. Fast two-phase image deblurring under impulse noise. J. Math. Imaging Vis. 2010, 36, 46–53. [Google Scholar] [CrossRef]
  32. Tian, C.; Zheng, M.; Zuo, W.; Zhang, B.; Zhang, Y.; Zhang, D. Multi-stage image denoising with the wavelet transform. Pattern Recognit. 2023, 134, 109050. [Google Scholar] [CrossRef]
  33. Elhassan, M.A.; Yang, C.; Huang, C.; Legesse Munea, T.; Hong, X. S 2 -FPN: Scale-ware Strip Attention Guided Feature Pyramid Network for Real-time Semantic Segmentation. arXiv 2022. [Google Scholar] [CrossRef]
  34. Xue, Y.; Jin, G.; Shen, T.; Tan, L.; Wang, L. Template-Calibrating the Dice loss to handle neural network overconfidence for bioadaptive cross-entropy loss for UAV visual tracking. Chin. J. Aeronaut. 2023, 36, 299–312. [Google Scholar] [CrossRef]
  35. Yeung, M.; Rundo, L.; Nan, Y.; Sala, E.; Schönlieb, C.B.; Yang, G. Calibrating the Dice loss to handle neural network overconfidence for biomedical image segmentation. J. Digit. Imaging 2023, 36, 739–752. [Google Scholar] [CrossRef]
  36. Blonder, B.; Both, S.; Jodra, M.; Majalap, N.; Burslem, D.; Teh, Y.A.; Malhi, Y. Leaf venation networks of Bornean trees: Images and hand-traced segmentations. Ecology 2019, 100. [Google Scholar]
  37. Gu, Y.; Piao, Z.; Yoo, S.J. STHarDNet: Swin transformer with HarDNet for MRI segmentation. Appl. Sci. 2022, 12, 468. [Google Scholar] [CrossRef]
  38. Yang, B.; Qin, L.; Peng, H.; Guo, C.; Luo, X.; Wang, J. SDDC-Net: A U-shaped deep spiking neural P convolutional network for retinal vessel segmentation. Digit. Signal Process. 2023, 136, 104002. [Google Scholar] [CrossRef]
  39. Maqsood, A.; Farid, M.S.; Khan, M.H.; Grzegorzek, M. Deep malaria parasite detection in thin blood smear microscopic images. Appl. Sci. 2021, 11, 2284. [Google Scholar] [CrossRef]
  40. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. Unet++: Redesigning skip connections to exploit multiscale features in image segmentation. IEEE Trans. Med. Imaging 2019, 39, 1856–1867. [Google Scholar] [CrossRef]
  41. Howard, A.; Sandler, M.; Chu, G.; Chen, L.C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; et al. Searching for mobilenetv3. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1314–1324. [Google Scholar]
  42. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  43. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar] [CrossRef]
  44. Touvron, H.; Cord, M.; Douze, M.; Massa, F.; Sablayrolles, A.; Jégou, H. Training data-efficient image transformers & distillation through attention. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 10347–10357. [Google Scholar]
  45. Tang, W.; Zou, D.; Yang, S.; Shi, J.; Dan, J.; Song, G. A two-stage approach for automatic liver segmentation with Faster R-CNN and DeepLab. Neural Comput. Appl. 2020, 32, 6769–6778. [Google Scholar] [CrossRef]
  46. Jiang, Z.; Ding, C.; Liu, M.; Tao, D. Two-stage cascaded u-net: 1st place solution to brats challenge 2019 segmentation task. In Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries: 5th International Workshop, BrainLes 2019, Held in Conjunction with MICCAI 2019, Shenzhen, China, 17 October 2019; Revised Selected Papers, Part I 5; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; pp. 231–241. [Google Scholar] [CrossRef]
  47. Božič, J.; Tabernik, D.; Skočaj, D. End-to-end training of a two-stage neural network for defect detection. In Proceedings of the IEEE 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 5619–5626. [Google Scholar] [CrossRef]
  48. Liu, J.; Yang, X.; Lau, S.; Wang, X.; Luo, S.; Lee, V.C.S.; Ding, L. Automated pavement crack detection and segmentation based on two-step convolutional neural network. Comput.-Aided Civ. Infrastruct. Eng. 2020, 35, 1291–1305. [Google Scholar] [CrossRef]
  49. Park, M.H.; Cho, J.H.; Kim, Y.T. CNN Model with Multilayer ASPP and Two-Step Cross-Stage for Semantic Segmentation. Machines 2023, 11, 126. [Google Scholar] [CrossRef]
  50. Jiang, Y.; Pang, D.; Li, C.; Yu, Y.; Cao, Y. Two-step deep learning approach for pavement crack damage detection and segmentation. Int. J. Pavement Eng. 2022, 1–14. [Google Scholar] [CrossRef]
  51. Dina, A.S.; Siddique, A.B.; Manivannan, D. A deep learning approach for intrusion detection in Internet of Things using focal loss function. Internet Things 2023, 22, 100699. [Google Scholar] [CrossRef]
Figure 1. The pipeline of the proposed TTH-Net.
Figure 1. The pipeline of the proposed TTH-Net.
Applsci 13 11019 g001
Figure 2. Illustration of cross-stage semantic enhancement module.
Figure 2. Illustration of cross-stage semantic enhancement module.
Applsci 13 11019 g002
Figure 3. Leaf images in the LVN dataset.
Figure 3. Leaf images in the LVN dataset.
Applsci 13 11019 g003
Figure 4. Ground truth in the LVN dataset (Corresponding with Figure 3).
Figure 4. Ground truth in the LVN dataset (Corresponding with Figure 3).
Applsci 13 11019 g004
Figure 5. Visual comparison between TTH-Net and other competitors on the LVN dataset for lea vein segmentation. The red rectangular boxes and arrows are meant to highlight the advantages of TTH-Net.
Figure 5. Visual comparison between TTH-Net and other competitors on the LVN dataset for lea vein segmentation. The red rectangular boxes and arrows are meant to highlight the advantages of TTH-Net.
Applsci 13 11019 g005
Figure 6. Training performance plots. (a) Loss vs. iterations. (b) Accuracy vs. iterations.
Figure 6. Training performance plots. (a) Loss vs. iterations. (b) Accuracy vs. iterations.
Applsci 13 11019 g006
Table 1. The results of the proposed TTH-Net and competitors on the LVN dataset. Note that each metric is reported via the mean (standard deviation) of 5-fold cross-validation.
Table 1. The results of the proposed TTH-Net and competitors on the LVN dataset. Note that each metric is reported via the mean (standard deviation) of 5-fold cross-validation.
ModelYearDiceSpeSenAcc
CNNFCN [10]201576.52 ± 1.2977.57 ±   2 .5281.54 ± 2.6383.78 ± 0.05
U-Net [12]201581.47 ± 1.2781.38 ±   2 .4584.28 ±   2 .5787.02 ± 0.03
GhostNet [13]202081.22 ± 1.3580.87 ±   2 .6884.33 ±   2 .6686.74 ± 0.06
MobileNetv3 [41]201981.34 ± 1.3880.75 ±   2 .6084.36 ±   2 .7286.73 ± 0.07
DeepLabv3+ [42]201882.51 ± 1.2581.90 ±   2 .4185.06 ±   2 .5187.12 ± 0.03
TransformerViT [15]202082.86 ± 2.4281.41 ±   3 . 2186.30 ±   3 .8487.24 ± 0.04
Segmenter [16]202181.71 ±   2 .3780.77 ±   3 . 0284.88 ±   3 .9187.15 ± 0.05
Swin Transformer [43]202183.14 ±   2 .4481.76 ±   3 .8686.75 ±   3 .7287.86 ± 0.03
SETR [17]202182.82 ±   2 .4181.48 ±   3 .7286.69 ±   3 .8887.59 ± 0.04
DeiT [44]202182.78 ±   2 .4881.98 ±   3 .8486.67 ±   3 .9087.81 ± 0.03
Two-stageALSNet [45]202082.36 ±   1 .2381.64 ±   2 .5685.58 ±   2 .5987.44 ± 0.03
TSUNet [46]202082.93 ±   1 .2581.53 ±   2 .7285.72 ±   2 .4787.63 ± 0.02
DDNet [47]202183.89 ±   1 .3182.45 ±   2 .6486.85 ±   2 .5387.82 ± 0.01
APCNet [48]202083.40 ±   1 .4081.55 ±   2 .5186.77 ±   2 .5887.47 ± 0.03
CMMANet [49]202383.96 ±   1 .2982.03 ±   2 .4386.89 ±   2 .6187.62 ± 0.04
PCDNet [50]202283.49 ±   1 .2682.16 ±   2 .4986.38 ±   2 .5087.53 ± 0.02
TTH-Net 85.24  ±   1 .2882.95  ±   2 .4687.36  ±   2 .5589.01  ± 0.03
Table 2. Ablation studies of the proposed TTH-Net. Note that each metric is reported via the mean (standard deviation) of 5-fold cross-validation.
Table 2. Ablation studies of the proposed TTH-Net. Note that each metric is reported via the mean (standard deviation) of 5-fold cross-validation.
ModuleDiceSpeSenAcc
with 1st stage83.83 ±   1 .3581.57 ±   2 .6386.03 ±   2 .8087.67 ± 0.04
with 1st stage + 2nd stage84.47 ±   1 .2682.18 ±   2 .4786.42 ±   2 .6188.22 ± 0.03
with 1st + 2nd stage + CSE module85.24  ±   1 .2882.95  ±   2 .4687.36  ±   2 .5589.01  ± 0.03
Table 3. Ablation studies of the loss function. Note that each metric is reported via the mean (standard deviation) of 5-fold cross-validation.
Table 3. Ablation studies of the loss function. Note that each metric is reported via the mean (standard deviation) of 5-fold cross-validation.
StudyStageLoss FunctionDiceSpeSenAcc
CEDLFocal
11st85.24  ±   1 .2882.95  ±   2 .4687.36  ±   2 .5589.01  ± 0.03
2nd
21st84.56 ±   1 .3582.37 ±   2 .5486.47 ±   2 .6788.37 ± 0.04
2nd
31st83.89 ±   1 .3882.42 ±   2 .5885.68 ±   2.57 87.42 ± 0.03
2nd
41st84.78 ±   1 .2682.67 ±   2 .4186.02 ±   2 .5287.76 ± 0.04
2nd
51st84.83 ±   1 .2982.25 ±   2 .4486.14 ±   2 .4188.35 ± 0.02
2nd
Table 4. Ablation studies of the hyperparameter λ . Note that each metric is reported via the mean (standard deviation) of 5-fold cross-validation.
Table 4. Ablation studies of the hyperparameter λ . Note that each metric is reported via the mean (standard deviation) of 5-fold cross-validation.
H y p e r p a r a m e t e r λ DiceSpeSenAcc
0.384.56 ±   1 .3481.74 ±   2 .5887.03 ±   2 .6488.58 ± 0.04
0.485.24   ±   1 .2882.95   ±   2 .4687.36   ±   2 .5589.01   ± 0.03
0.585.07 ±   1 .3182.46 ±   2 .4787.10 ±   2 .5788.56 ± 0.03
0.684.57 ±   1 .3882.20 ±   2 .6287.02 ±   2 .7188.14 ± 005
0.784.42 ±   1 .3582.28 ±   2 .5387.04 ±   2 .6988.28 ± 0.04
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Song, P.; Yu, Y.; Zhang, Y. TTH-Net: Two-Stage Transformer–CNN Hybrid Network for Leaf Vein Segmentation. Appl. Sci. 2023, 13, 11019. https://doi.org/10.3390/app131911019

AMA Style

Song P, Yu Y, Zhang Y. TTH-Net: Two-Stage Transformer–CNN Hybrid Network for Leaf Vein Segmentation. Applied Sciences. 2023; 13(19):11019. https://doi.org/10.3390/app131911019

Chicago/Turabian Style

Song, Peng, Yonghong Yu, and Yang Zhang. 2023. "TTH-Net: Two-Stage Transformer–CNN Hybrid Network for Leaf Vein Segmentation" Applied Sciences 13, no. 19: 11019. https://doi.org/10.3390/app131911019

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop