Skip to main content
  • Original research
  • Open access
  • Published:

Direct reconstruction for simultaneous dual-tracer PET imaging based on multi-task learning

Abstract

Background

Simultaneous dual-tracer positron emission tomography (PET) imaging can observe two molecular targets in a single scan, which is conducive to disease diagnosis and tracking. Since the signals emitted by different tracers are the same, it is crucial to separate each single tracer from the mixed signals. The current study proposed a novel deep learning-based method to reconstruct single-tracer activity distributions from the dual-tracer sinogram.

Methods

We proposed the Multi-task CNN, a three-dimensional convolutional neural network (CNN) based on a framework of multi-task learning. One common encoder extracted features from the dual-tracer dynamic sinogram, followed by two distinct and parallel decoders which reconstructed the single-tracer dynamic images of two tracers separately. The model was evaluated by mean squared error (MSE), multiscale structural similarity (MS-SSIM) index and peak signal-to-noise ratio (PSNR) on simulated data and real animal data, and compared to the filtered back-projection method based on deep learning (FBP-CNN).

Results

In the simulation experiments, the Multi-task CNN reconstructed single-tracer images with lower MSE, higher MS-SSIM and PSNR than FBP-CNN, and was more robust to the changes in individual difference, tracer combination and scanning protocol. In the experiment of rats with an orthotopic xenograft glioma model, the Multi-task CNN reconstructions also showed higher qualities than FBP-CNN reconstructions.

Conclusions

The proposed Multi-task CNN could effectively reconstruct the dynamic activity images of two single tracers from the dual-tracer dynamic sinogram, which was potential in the direct reconstruction for real simultaneous dual-tracer PET imaging data in future.

Introduction

Positron emission tomography (PET) is a nuclear imaging technique that uses radiolabeled tracers to measure metabolic functions in vivo. It provides useful information and guidance for the detection, diagnosis, staging, treatment planning and monitoring of diseases [1, 2].

For some diseases, a PET scan with a single tracer is insufficient to reveal the complex pathological characteristics. Researchers tried out multi-tracer PET imaging for more complete information, which was usually achieved by multiple single-tracer scans. For example, Fu et al. [3] demonstrated that combining the imaging results of \(^{18}\)F-fluorodeoxyglucose (\(^{18}\)F-FDG) and \(^{18}\)F-fluorocholine (\(^{18}\)F-FCH) led to more accurate detection of low-grade glioma than using \(^{18}\)F-FDG or \(^{18}\)F-FCH only.

Instead of two separate single-tracer scans, a simultaneous dual-tracer PET scan can eliminate errors caused by image misalignment and physiological changes between scans, and also reduce the total scanning time. However, different tracers all generate the 511-keV photon pairs, which makes signal separation a difficulty in the reconstruction for simultaneous dual-tracer PET imaging. Efforts have been devoted to the development of various separation techniques, including traditional methods mainly based on estimation and more recent approaches based on deep learning or machine learning.

Traditional estimation approaches have been extensively studied [4]. The earliest simualtion study [5] and later animal study [6] achieved signal separation based on different half-lives of two tracers. Later on, the parallel compartment model [7, 8] was widely used in tracer separation, by which the kinetic parameters of two tracers were simultaneously estimated to recover the activity distributions of each tracer. Several studies have investigated the effects of tracer combination, injection interval, and scanning protocol on the results of compartment model-based signal separation [9,10,11,12]. The model-based separation method was simplified by introducing a reference region [13] or the technique of reduced parameter space [14,15,16]. A state-space representation based on the compartment model [17] was proposed for simultaneous signal separation and reconstruction. In addition, separation methods based on principal component analysis [18], generalized factor analysis [19], or basis function fitting [20] were also proposed. In particular, Andreyev and Celler [21] proposed a physical method to distinguish different tracer signals using a specially labeled tracer that can additionally emit a high-energy \(\gamma\) photon. Fukuchi et al. used special detectors to receive prompt \(\gamma\)-rays, and proposed reconstruction methods based on data subtraction for dual-tracer PET imaging when one tracer was labeled by a pure positron emitter and the other by a positron-\(\gamma\) emitter [22, 23].

Since these traditional methods explicitly use specific prior information, they also have limitations in applications. Methods based on the difference of half-lives are not applicable to tracers with same or close half-lives. Methods based on the parallel compartment model require staggered injection and arterial blood sampling. To use compartment model with reference tissue method, the firstly injected tracer must have a reference region. Besides, component analysis may have non-unique solutions, especially if the component factors of the tracers are similar.

Compared to traditional methods, deep learning or machine learning methods for dual-tracer separation or reconstruction are not constrained by prior information, thus can deal with simultaneously injected tracers or tracers with the same half-lives.

There are two kinds of deep learning approaches for dual-tracer reconstruction: the indirect ones and direct ones. Indirect methods firstly reconstruct the dual-tracer dynamic images by traditional reconstruction algorithms, then separate the dual-tracer images by deep neural networks to obtain single-tracer images [24,25,26,27,28]. These networks separate signals either from the voxel time-activity curves (TACs) [24,25,26,27] or from the entire dynamic image [28], and are easily influenced by the quality of reconstructed images. Direct methods reconstruct images of two tracers directly from the dual-tracer sinogram, which means the network learns to solve both the separation task and reconstruction task. As an example, Xu and Liu [29] imitated a filtered back-projection (FBP) procedure by network to reconstruct dual-tracer images and used a three-dimensional (3D) convolutional neural network (CNN) to provide an estimation of two tracers from the reconstructed images, referred to as FBP-CNN model. FBP-CNN behaved better than indirect reconstruction methods since it also learned spatial information from sinogram. However, the FBP part of the model included tremendous number of parameters, causing demands for both large amount of training data and memories. The development of a more trainable and memory-friendly network is necessary. Apart from these deep learning models, the recurrent extreme gradient boosting, a machine learning algorithm, was recently used to separate tissue TACs in the \(^{18}\)F-FDG/\(^{68}\)Ga-DOTATATE imaging of neuroendocrine tumors [30].

This study aims to directly reconstruct activity distributions of two tracers from the dual-tracer sinogram by a novel deep neural network named Multi-task CNN, which is characteristic of:

(a) A 3D encoder-decoder structure. The deep encoder-decoder structure is motivated by DeepPET [31] which was used for single-tracer, static PET image reconstruction. To cope with the dynamic data in dual-tracer PET imaging, we adopt 3D convolution layers to form the encoder and decoder, by which the temporal kinetic features and spatial structural features are learned simultaneously. The former plays an important role in the separation task, while the latter is useful for reconstruction.

(b) A multi-task learning framework. The multi-task learning [32] forces a model to learn multiple different but related tasks simultaneously. The learned information is shared between tasks to improve the performances of the model. It has been widely used in computer vision [33] and medical image analysis [34]. In dual-tracer PET imaging, the reconstruction of two single-tracer activity images can be regarded as two different but related tasks. In Multi-task CNN, we set one common encoder while extend the decoder part into two branches.

(c) The direct encoding of sinogram. Unlike FBP-CNN which reconstructed the dual-tracer activity images by network before encoding, the Multi-task CNN avoids the reconstruction of dual-tracer images but directly encodes the dual-tracer sinogram.

Methods

Models of dual-tracer PET imaging and reconstruction

Generally, the process of a single-tracer PET imaging is formulated as

$$\begin{aligned} y(t) = Gx(t)+n(t), \end{aligned}$$
(1)

where y(t) and x(t) are the sinogram and activity distribution at time t. The system matrix G describes the probability of each voxel in the activity distribution to be detected by each bin in the sinogram, which is related to the physical and geometrical structure of the PET scanner [35]. And n(t) is the noise mainly caused by random and scattered coincidences.

The dual-tracer PET imaging with simultaneous injection can be defined as

$$\begin{aligned} y^{I+II}(t) = G\left[ x^I(t)+x^{II}(t)\right] +n^I(t)+n^{II}(t), \end{aligned}$$
(2)

where \(y^{I+II}(t)\) is the dual-tracer sinogram, \(x^I(t)\) and \(x^{II}(t)\) are the activity distributions of Tracer I and Tracer II. The different spatiotemporal distributions of the tracers lead to different noises \(n^I(t)\) and \(n^{II}(t)\).

For a certain scanning protocol, the dynamic dual-tracer sinogram is denoted as \(Y^{I+II}=[y^{I+II}(t_1),\dots ,y^{I+II}(t_F)]\), where F is the number of frames, \(t_f\) is the mid-frame time of the \(f^{th}\) frame. The dynamic single-tracer activity images can be denoted in the same way as \(X^I\) and \(X^{II}\). The direct reconstruction for dual-tracer PET imaging can be described as

$$\begin{aligned}{}[{\hat{X}}^I,{\hat{X}}^{II}]=h_\theta (Y^{I+II}), \end{aligned}$$
(3)

where \({\hat{X}}^I\) and \({\hat{X}}^{II}\) are the estimatation of \(X^I\) and \(X^{II}\). In this study, \(\theta\) are the weights of the proposed Multi-task CNN.

Multi-task CNN

Network architecture

Fig. 1
figure 1

The architecture of Multi-task CNN. The dual-tracer dynamic sinogram is inputted to the encoder, and two decoders output the dynamic activity images of two tracers. Conv = convolution layer, BN = batch normalization layer, ReLU = rectified linear unit layer, Deconv=deconvolution layer. The kernel size and number of channels are noted in the figure

The architecture of Multi-task CNN is presented in Fig. 1. The dual-tracer dynamic sinogram of a single slice, in the shape of frames \(\times\) bins \(\times\) angles, is taken in as the network input. The spatiotemporal features are extracted from the dynamic sinogram by the encoder, which includes three downsampling blocks. In each block, a 3D convolution layer with a stride of 1 is firstly applied, with the size of each feature map maintained. Then, the second 3D convolution layer downsamples the features by a stride of 2. The number of channels in three blocks are 16, 32 and 64. At the end of the encoder, a 3D convolution layer is used to extend the features to 128 channels.

The multi-task learning scheme is reflected in using two decoders to estimate the activity images of two tracers separately. The two decoders have the same structure, both including three upsampling blocks. In each block, a 3D deconvolution layer is firstly used to upsample the features by a stride of 2. Then a 3D convolution layer with a stride of 1 is applied, with the size of features kept unchanged. Symmetry to the encoder, the number of channels are 64, 32, 16 in these blocks. Finally, a 3D convolution layer with a stride of 1 is used to reduce the number of channels to 1. All the convolution and deconvolution layers are connected by a batch normalization layer and a rectified linear unit (ReLU) layer. The size of the kernels are \(3 \times 3 \times 3\) in all convolution and deconvolution layers, except the last layer which uses 1 \(\times\) 1 \(\times\) 1 kernels.

Loss function

The loss function of a single-frame image of one tracer is a combination of the mean squared error (MSE) and the structural similarity (SSIM) [36]:

$$\begin{aligned} L({\hat{x}},x)= \alpha {\text{MSE}}({\hat{x}},x)-\beta \ln {\frac{1+{\text{SSIM}}({\hat{x}},x)}{2}}, \end{aligned}$$
(4)

where \({\hat{x}}\) is an estimated single-frame activity image, and x is the corresponding label. \(\alpha\) and \(\beta\) are the coefficients that balance MSE and SSIM.

The total loss \(L_{\text{total}}\) was summed by the loss of two tracers and averaged by frames:

$$\begin{aligned} L_{\text{total}}=\frac{1}{F}\sum _{f=1}^F L\left[ {\hat{x}}^I(t_f),x^I(t_f)\right] +L\left[ {\hat{x}}^{II}(t_f),x^{II}(t_f)\right] . \end{aligned}$$
(5)

Comparative method and evaluation metrics

The proposed Multi-task CNN was compared to FBP-CNN [29] by MSE, multiscale structural similarity (MS-SSIM) index [37] and peak signal-to-noise ratio (PSNR). Frame-wise MS-SSIM is formulated as

$$\begin{aligned} {\text{MS}-{\text{SSIM}}({\hat{x}},x)}=[l_M({\hat{x}},x)]^{\gamma _{{1},M}}\cdot \prod _{m=0}^M[c_m ({\hat{x}},x)]^{\gamma _{{2},m}}[s_m({\hat{x}},x)]^{\gamma _{{3},m}}, \end{aligned}$$
(6)

where M represents the number of dowmsampling operations by a factor of 2 performed to the estimated image \({\hat{x}}\) and label image x. The subscript m is related to the \(m^{th}\) downsampling operation. For instance, \(c_m({\hat{x}},x)\) calculates \(c(\cdot ,\cdot )\) of the m-times downsampled images, and its corresponding exponential parameter is \(\gamma _{2,m}\). \(l({\hat{x}},x)\), \(c({\hat{x}},x)\) and \(s({\hat{x}},x)\) measure the differences of luminent, contrast and structure similarity between two images. The detailed calculation of \(l(\cdot ,\cdot )\), \(c(\cdot ,\cdot )\) and \(s(\cdot ,\cdot )\), as well as the values of M and exponential parameters can refer to the previous publication [37]. And the frame-wise PSNR is formulated as:

$$\begin{aligned} {\text{PSNR}}({\hat{x}},x)=10\cdot \log _{10}\left[ \frac{x_{\max }^2}{\text{MSE}({\hat{x}},x)}\right] , \end{aligned}$$
(7)

where \(x_{\max }\) is the maximum value of x.

Experiments

In this study, we trained and tested the Multi-task CNN and FBP-CNN on four simulated datasets and one animal dataset from real PET experiments.

Simulation study

Experimental settings

Table 1 Settings of the simulation experiments

The simulation study contained four experiments: two individual difference experiments (EXP 1-1, EXP 1-2), one tracer combination experiment (EXP 2) and one scanning protocol experiment (EXP 3). The detailed settings are shown in Table 1. To simulate the physiological characteristics of different people, we applied Gaussian randomization to the tracer kinetic parameters and parameters of plasma input function in all experiments. The numbers of parameter sets (including both kinetic parameters and input function parameters) and the levels of physiological variation are listed in Table 1. In all experiments, the number of frames of the dynamic data was set to 18.

Exp 1-1 and EXP 1-2 simulated data of 30-min \(^{11}\)C-Flumazenil (\(^{11}\)C-FMZ)/\(^{11}\)C-acetate imaging. The two experiments designed the same variation range (10%) of input functions but different variation ranges (10%, 20%) of kinetic parameters. In EXP 2, we set two combinations of tracers, where \(^{11}\)C-FMZ/\(^{11}\)C-acetate had the same half-life, while \(^{18}\)F-FDG/\(^{11}\)C-FMZ had different half-lives. The scanning protocols were different accordingly. The levels of both input variation and kinetic variation of two combinations were set to 10%. In EXP 3, we designed three scanning protocols with the durations of 40 min, 50 min and 60 min for \(^{18}\)F-FDG/\(^{11}\)C-FMZ imaging. All the variation ranges were set to 10%.

Simulation process

Fig. 2
figure 2

The Zubal phantoms used in simulation experiments. a b, c, and d are representative phantoms corresponding to differenct slices of the original 3D phantom

The phantoms used for simulation were originated from the 3D Zubal brain phantom [38]. By region of interest (ROI) partition, slice selection and dowmsampling, we obtained 40 two-dimensional phantoms sized 128 \(\times\) 128. The voxel size was \(2.2 \times 2.2 \times 1.4\) mm\(^3\). The number of ROIs in each phantom varied from 2 to 5. Representative phantoms are shown in Fig. 2.

The single-tracer dynamic activity images were generated based on the two-tissue compartment model [39] which described the kinetic characteristics of the tracer:

$$\begin{aligned} \begin{aligned} \dfrac{\text{d}C_1(t)}{\text{d}t}&=K_1 C_P(t)-(k_2+k_3)C_1(t)+k_4 C_2(t) \\ \dfrac{\text{d}C_2(t)}{\text{d}t}&=k_3 C_1(t)-k_4 C_2(t) \\ C_T(t)&=C_1(t)+C_2(t) \end{aligned} \end{aligned}$$
(8)

\(C_1(t)\) and \(C_2(t)\) are the tracer concentrations in two tissue compartments. The two compartments represent different metabolic states of the tracer. \(C_P(t)\) is the tracer concentration in plasma, i.e., the plasma input function. \(K_1\), \(k_2\), \(k_3\) and \(k_4\) are the kinetic parameters reflecting the speed of directional tracer exchanges between different tissue compartments or between plasma and tissue, which are different among ROIs, individuals and tracers. The input functions of \(^{18}\)F-FDG [40], \(^{11}\)C-FMZ [41] and \(^{11}\)C-acetate [18] were generated by different models. By applying Gaussian randomization to the kinetic parameters and parameters of input function, we simulated individual differences in physiological states. The mean values of these parameters were set as previously reported [7, 15, 18, 40,41,42], and the standard deviations were set as a portion of mean values, which are shown in Table 1. After setting the aforementioned phantoms, parameters and the scanning protocols, the tissue TACs \(C_T(t)\) were obtained by numerically solving Eq. (8) by the COMKAT toolbox [43]. All voxel TACs formed the noise-free dynamic activity images. The images generated were sized 128 \(\times\) 128 \(\times\) 18 frames.

The single-tracer dynamic sinograms were generated by frame-wise projection of single-tracer activity images using Michigan Image Reconstruction Toolbox [44]. We designed the system matrix by simple strip integrals according to the geometry of the Siemens Inveon PET/CT scanner [45], without the consideration of attenuation, scatter and variations in detector efficiencies. We added 20% random noise and Poisson noise to the sinograms. The size of the dynamic sinogram was 128 bins \(\times\) 160 angles \(\times\) 18 frames.

The dual-tracer dynamic sinogram was obtained by simply adding up two single-tracer sinograms. This guaranteed the alignment of images, the match of doses and the consistency of physiological states between the dual-tracer scan and two single-tracer scans. The dual-tracer dynamic sinogram was used as the network input.

Additionally, the single-tracer dynamic activity images with noise were reconstructed frame-wisely by the ordered subsets expectation maximization (OSEM) algorithm [46] with 6 iterations and 5 subsets. Corrections and normalization were not performed. In consistence with the phantom and simulated noise-free activity images, the voxel size of the reconstructed images was \(2.2 \times 2.2 \times 1.4\) mm\(^3\). The reconstructed images of two tracers were used as labels.

Datasets and network training

Datasets used in EXP 1-1, EXP 1-2, EXP 2 and EXP 3 all included 30 parameter sets. In particular, 15 sets of parameters were simulated for each one of the two tracer combinations in EXP 2, and 10 sets of parameters were simulated for each one of the three scanning protocols in EXP 3. Therefore, the four datasets used for network training all contained 1200 groups (40 phantoms \(\times\) 30 parameter sets) of simulated PET data. Each dataset was randomly divided into training data, validation data and test data by 8:1:1. The inputs were normalized by the mean and standard deviation of the training set, and the labels were normalized by dividing 255.

We used the Adam optimizer in the training of Multi-task CNN and FBP-CNN. For Multi-task CNN, the hyperparameters were set as: learning rate = 0.0005, epochs = 100, batch size = 4, \(\alpha\) = 100 and \(\beta\) = 1. For FBP-CNN, the hyperparameters were set as: learning rate = 0.0001, epochs = 100 and batch size = 4.

Animal study

Animals and tracers

The animal experiments were approved by the Experimental Animal Ethics Committee of Southern Medical University, and were conducted in compliance with the ARRIVE guidelines, the guidelines of the US National Institutes of Health and local legal requirements. 13 SD rats (average age: 9 weeks; average weight: 300 g) with a C6 cell intracranial orthotopic xenograft glioma model were used in this experiment.

Previous PET study showed that \(^{18}\)F-FDG had a higher sensitivity but a lower specificity than O-(2-[\(^{18}\)F]fluoroethyl)-L-tyrosine (\(^{18}\)F-FET) in the detection of head and neck squamous cell carcinoma [47]. A \(^{18}\)F-FDG/\(^{18}\)F-FET dual-tracer PET scan might be benificial for better detection and location of tumors. We selected \(^{18}\)F-FDG and \(^{18}\)F-FET as tracers to evaluate the feasibility of our model on real PET data.

PET imaging

Two single-tracer PET scans were performed on each rat using a Siemens Inveon PET/CT scanner [45]. The rats underwent an 8-hour fasting before PET scans. During the scans, they were kept under anesthesia by 1% isoflurane and 1 L/min of oxygen, and also fixed by medical tapes. At the beginning of each scan, \(^{18}\)F-FDG or \(^{18}\)F-FET was injected at a dose around 37 MBq into the tail vein of the rat. The \(^{18}\)F-FDG scan and \(^{18}\)F-FET scan followed the same scanning protocol (60 s \(\times\) 10, 300 s \(\times\) 3420 s \(\times\) 5), and were carried out on two different days in random order.

Data preparation

The 3D dynamic sinogram of each single-tracer scan was obtained. Using the built-in software, the dynamic activity images were reconstructed frame-wisely by 3D-OSEM algorithm with normalization and random correction. Attenuation correction was also performed using individual computed tomography (CT) images. The reconstructed single-tracer activity images were sized 128 \(\times\) 128 \(\times\) 18 frames with a voxel size of \(0.78 \times 0.78 \times 0.80\) mm\(^3\), and were used as labels.

Since the network input should be a time sequence of a 2D sinogram, the original 3D sinogram was random corrected and re-organized by single slice rebinning to form 2D sinogram. The obtained dynamic sinograms were sized 128 bins \(\times\) 160 angles \(\times\) 18 frames. As done in previous dual-tracer PET experiments [14, 48], we added sinograms of two tracers to get the dual-tracer sinogram, which was used as the network input.

Datasets and network training

Totally, 351 brain slices from 13 rats were selected, among which 297 groups of data from 11 rats were used for training and validation, the remaining 54 groups of data from the other 2 rats were used for testing. The inputs were normalized by the mean and standard deviation of the training set, and the labels were normalized by dividing \(10^6\), which converted the unit from Bq/mL to MBq/mL.

Same as the simulation study, Adam optimizer was used in the training of Multi-task CNN and FBP-CNN. For Multi-task CNN, the hyperparameters were set as: learning rate = 0.0001, epochs = 100, batch size = 4, \(\alpha\) = 1 and \(\beta\) = 0.01. Moreover, a weight decay with a coefficient of 0.01 was added to the loss function. For FBP-CNN, the hyperparameters were set as: learning rate = 0.00002, epochs = 250, batch size = 4.

Results

Simulation study

Table 2 Quantitative results of the simulation experiments

Table 2 lists the mean values of MSE, MS-SSIM and PSNR of each test datasets. In all simulation experiments and datasets, Multi-task CNN showed lower MSE, higher MS-SSIM and PSNR than FBP-CNN. The results of each simulation experiment are as follows.

Effect of individual difference

Fig. 3
figure 3

The representative reconstructed images from the individual difference experiment. a 10% kinetic variation, b 20% kinetic variation. These images show the 10th frame of the dynamic images

As shown in Fig. 3, single-tracer activity images reconstructed by Multi-task CNN contained less noise than those reconstructed by FBP-CNN, and were even smoother than the ground truth in some regions. When the kinetic variation increased from 10% (Fig. 3a) to 20% (Fig. 3b), the reconstruction results of FBP-CNN became worse, where the image details were not well reconstructed.

As displayed in Table 2, for both tracers under two levels of individual difference, the Multi-task CNN reconstructions had significantly lower MSE, higher MS-SSIM and PSNR than FBP-CNN reconstructions. When the level of individual difference increased, the MSE of both methods increased, while Multi-task CNN increased by a smaller increment. The MS-SSIM and PSNR of FBP-CNN reconstructions of two tracers all decreased. Although the MS-SSIM and PSNR of Multi-task CNN reconstructions of Tracer II (\(^{11}\)C-acetate) were also slightly reduced, the two-tailed t tests showed that the MS-SSIM (\({T}=-2.57, {P}=0.011\)) and PSNR (\({T}=0.42, {P}=0.674\)) of reconstructed images of Tracer I (\(^{11}\)C-FMZ) were not influenced.

Effect of tracer combination

Fig. 4
figure 4

The representative reconstructed images from the tracer combination experiment. a Tracer combination I (\(^{11}\)C-FMZ/\(^{11}\)C-acetate), b Tracer combination II (\(^{18}\)F-FDG/\(^{11}\)C-FMZ). These images show the 10th frame of the dynamic images

Visually in Fig. 4, Multi-task CNN reconstructed images with better details and less noise than FBP-CNN. Table 2 quantitatively shows that the MS-SSIM and PSNR of Multi-task CNN reconstructions were higher, and the MSE was lower.

It is worth noting that \(^{11}\)C-FMZ was set as Tracer I in the first tracer combination (Fig. 4a), but as Tracer II in the second combination (Fig. 4b). In both cases, \(^{11}\)C-FMZ images reconstructed by FBP-CNN showed an “checkerboard effect”, which was more obvious in the second combination, resulting in extremely low MS-SSIM (0.4762 ± 0.0475) and PSNR (17.13 ± 0.33). Compared to FBP-CNN, metrics of Multi-task CNN reconstructions were more stable between different tracer combinations.

Effect of scanning protocol

Fig. 5
figure 5

The representative reconstructed images from Protocol III (60-min scan) of the scanning protocol experiment

Figure 5 shows the reconstructed images of \(^{18}\)F-FDG/\(^{11}\)C-FMZ under one of the scanning protocols. The 60-min scan was too long for \(^{11}\)C-FMZ imaging, causing quite different concentrations of two tracers in later frames. Even in such case, both Multi-task CNN and FBP-CNN can well reconstructed the temporal changes of tracer activities. However, the structures of \(^{18}\)F-FDG images reconstructed by FBP-CNN were inconsistent with the ground truth. Furthermore, in the 16th frame of \(^{11}\)C-FMZ, both methods overestimated the tracer activities, while the estimation of Multi-task CNN was closer to the label.

Fig. 6
figure 6

The ROI-TACs from the scanning protocol experiment

Overall, MS-SSIM and PSNR of Multi-task CNN reconstructions were higher than FBP-CNN, and the MSE was lower (Table 2). Figure 6 shows the ROI-TACs extracted from the reconstructed images. Each subplot contains estimated and label TACs of two tracers. The rows and columns of the figure represent different ROIs and scanning protocols. Obviously shown in the first row, TACs of FBP-CNN reconstructions severely deviated from the labels, which was resulted from negative values in the background. The TACs from Multi-task CNN reconstructions were closer to the label TACs in the background and ROI 1-4, while TACs from FBP-CNN reconstructions were more accurate in ROI 5. Besides, changes in scanning protocols had little influence on the performance of both methods.

Animal study

Fig. 7
figure 7

The representative reconstructed images from the animal experiment. These images show the same slice of one rat

Table 3 Quantitative results of the animal experiment

Figure 7 shows the representative images of the same brain slice of one rat. The \(^{18}\)F-FDG images reconstructed by both methods were less-accurate. As for \(^{18}\)F-FET images, the results of Multi-task CNN showed better details and contrast, and the tumor near the top could also be distinguished. Table 3 also shows that images reconstructed by Multi-task CNN had lower MSE, higher MS-SSIM and PSNR than FBP-CNN. Compared with the simulation experiments, the qualities of reconstructed images in animal experiment were significantly worse, which was mainly caused by misalignment of images from two single-tracer scans.

Discussion

Most traditional reconstruction methods for dual-tracer PET imaging could only separate two tracers when they were administrated with an interval, and were sensitive to tracer combinations and scanning protocols. These methods usually required invasive measurements of plasma input function by arterial blood sampling. Compared to the traditional methods, the proposed Multi-task CNN as well as other deep learning-based models could effectively separate signals from simultaneoulsly administrated tracers, and avoid the arterial blood sampling.

The deep learning-based indirect reconstruction models separate the dual-tracer images after reconstruction. The majority of these methods separate signals voxel-wisely, that is, by separating the TACs extracted from the reconstructed images [24,25,26,27]. Only temporal features were learned from the TACs and used for separation. There also existed a method separating from the image domain, using both spatial and temporal features [28]. However, these two types of methods were both influenced by the quality of reconstructed images. As a direct reconstruction method, the Multi-task CNN extracted and fully used the spatiotemporal information from the dual-tracer sinogram, which was not influenced by the traditional reconstruction algorithms.

In the current study, we quantitatively compared our Multi-task CNN to FBP-CNN [29], which is also a deep neural network for direct reconstruction for dual-tracer PET imaging. The robustness of Multi-task CNN and FBP-CNN to changes in tracer combiantion and scanning protocol was evaluated by simulation data. To make the simulated data more realistic, we used phantoms corresponding to different brain slices, and generated data with varied tracer kinetic parameters and input functions to imitate individual difference. Whether the models were sensitive to different levels of individual difference was also studied.

According to the results of simulation experiments, Multi-task CNN reconstructed single-tracer activity images with higher quality than FBP-CNN, and was more robust to individual difference and tracer combination than FBP-CNN. In the animal experiment, Multi-task CNN also showed its superiority over FBP-CNN.

The advantages of Multi-task CNN over FBP-CNN might be mainly due to different scales of two models. FBP-CNN explicitly consisted of two parts. Firstly, the reconstruction part unrolled the FBP algorithm by a convolution layer and a fully-connected layer to reconstruct the dual-tracer activity image. Then, the separation part estimated single-tracer images by 3D encoders and decoders. The number of parameters in the reconstruction part, especially the fully-connected layer, growed with the size of sinogram and activity image, which resulted in low generalization ability of the model. Unlike FBP-CNN, the encoder was directly operated on the sinogram in Multi-task CNN. The number of parameters in Multi-task CNN was fixed and independent of the size of sinogram or image. Nevertheless, the encoder part of the Multi-task CNN was less interpretable than the reconstruction part of the FBP-CNN.

The effectiveness of multi-task learning could be explained in two perspectives [32]. From the perspective of overfitting, it was commonly agreed that learning more than one task could prevent the model from overfitting, since it was more challenging than learning a single task. From the perspective of auxiliary learning, learning related tasks was beneficial for the model to extract appropriate and important features. In the Multi-task CNN, the reconstruction of two single-tracer activity images were regarded as two tasks. The two tasks had the same scanning protocol and system probability matrix since they were scanned simultaneously. However, the two tracers had different kinetic characteristics, input functions, as well as different levels of noise due to different activity distributions. Using Multi-task CNN could learn more substantial and important features by the encoder, and maintain the differences between two tracers by using two decoders.

However, the current study had several limitations. As shown in Fig. 6, in some ROIs, the activities reconstructed by the proposed method deviated from the ground truth. The TACs of \(^{11}\)C-FMZ tended to be inaccurate in early frames, while TACs of \(^{18}\)F-FDG were inaccurate in later frames. The introducing of attention mechanism to the model might be helpful to better extract different features from different frames. In addition, the images reconstructed by Multi-task CNN were over smoothed, which might be due to including SSIM in the loss function. A more appropriate coefficient of SSIM part in the loss function, and more suitable constraints should be further studied. Moreover, the dual-tracer sinograms in animal experiments were not obtained from dual-tracer scans, but from single-tracer scans for simplication, where the images were not aligned.

To use real experimental or clinical data to train deep learning models in future studies, standards of data acquicision and preprocessing should be set up. The prototols of simultaneous dual-tracer PET scans, and the issues concerning image alignment and dose matching need special focuses. For example, the injected doses of tracers in both single-tracer scans and dual-tracer scan are needed for scaling and dose matching. Images intra- and inter- scans should be well-aligned. The decay correction is not necessary since the concentrations of tracers in tissues are unknown. And the scanning protocol, corrections and reconstruction should be kept consistent between scans.

Conclusions

In this study, we proposed Multi-task CNN, a 3D encoder-decoder network based on multi-task learning, to directly reconstruct two single-tracer dynamic images from the dual-tracer dynamic sinogram. The proposed method was superior to the existing FBP-CNN, exhibiting its robustness to individual difference, tracer combination and scanning protocol in the simulation study. In the animal study, the feasibility of applying this method to real PET data was preliminarily verified. The proposed Multi-task CNN was thus considered a potential deep learning-based method for the direct reconstruction of simultaneous dual-tracer PET imaging.

Availability of data and materials

The datasets used or analysed during the current study are available from the corresponding author on reasonable request.

Abbreviations

PET:

Positron emission tomography

TAC:

Time-activity curve

FBP:

Filtered back-projection

3D:

Three-dimentional

CNN:

Convolutional neural network

ReLU:

Rectified linear unit

MSE:

Mean squared error

SSIM:

Structural similarity

MS-SSIM:

Multiscale structural similarity

PSNR:

Peak signal-to-noise ratio

ROI:

Region of interest

CT:

Computed tomography

References

  1. Pinker K, Riedl C, Weber WA. Evaluating tumor response with FDG PET: updates on PERCIST, comparison with EORTC criteria and clues to future developments. Eur J Nucl Med Mol Imaging. 2017;44(1):55–66. https://doi.org/10.1007/s00259-017-3687-3.

    Article  Google Scholar 

  2. Morsing A, Hildebrandt MG, Vilstrup MH, Wallenius SE, Gerke O, Petersen H, Johansen A, Andersen TL, Høilund-Carlsen PF. Hybrid PET/MRI in major cancers: a scoping review. Eur J Nucl Med Mol Imaging. 2019;46(10):2138–51. https://doi.org/10.1007/s00259-019-04402-8.

    Article  Google Scholar 

  3. Fu Y, Ong LC, Ranganath SH, Zheng L, Kee I, Zhan W, Yu S, Chow PK, Wang CH. A dual tracer 18F-FCH/18F-FDG PET imaging of an orthotopic brain tumor xenograft model. PLoS One. 2016;11(2):0148123.

    Article  Google Scholar 

  4. Kadrmas DJ, Hoffman JM. Methodology for quantitative rapid multi-tracer PET tumor characterizations. Theranostics. 2013;3(10):757–73.

    Article  Google Scholar 

  5. Huang SC, Carson RE, Hoffman EJ, Kuhl DE, Phelps ME. An investigation of a double-tracer technique for positron computerized tomography. J Nucl Med. 1982;23(9):816–22.

    CAS  Google Scholar 

  6. Figueiras FP, Jiménez X, Pareto D, Gómez V, Llop J, Herance R, Rojas S, Gispert JD. Simultaneous dual-tracer PET imaging of the rat brain and its application in the study of cerebral ischemia. Mol Imaging Biol. 2010;13(3):500–10.

    Article  Google Scholar 

  7. Koeppe RA. Compartmental analysis of [11C]flumazenil kinetics for the estimation of ligand transport rate and receptor distribution using positron emission tomography. J Cereb Blood Flow Metab. 1991;5(11):735–44.

    Article  Google Scholar 

  8. Guo J, Guo N, Lang L, Kiesewetter DO, Xie Q, Li Q, Eden HS, Niu G, Chen X. 18F-Alfatide II and 18F-FDG dual-tracer dynamic PET for parametric, early prediction of tumor response to therapy. J Nucl Med. 2014;55(1):154–60. https://doi.org/10.2967/jnumed.113.122069.

    Article  CAS  Google Scholar 

  9. Koeppe RA, Raffel DM, Snyder SE, Ficaro EP, Kilbourn MR, Kuhl DE. Dual-[11C]tracer single-acquisition positron emission tomography studies. J Cereb Blood Flow Metab. 2001;21(12):1480–92.

    Article  CAS  Google Scholar 

  10. Nishizawa S, Kuwabara H, Ueno M, Shimono T, Toyoda H, Konishi J. Double-injection FDG method to measure cerebral glucose metabolism twice in a single procedure. Ann Nucl Med. 2001;15:203–7.

    Article  CAS  Google Scholar 

  11. Rust TC, Kadrmas DJ. Rapid dual-tracer PTSM+ATSM PET imaging of tumour blood flow and hypoxia: a simulation study. Phys Med Biol. 2006;51(1):61–75.

    Article  CAS  Google Scholar 

  12. Iwanishi K, Watabe H, Hayashi T, Miyake Y, Minato K, Iida H. Influence of residual oxygen-15-labeled carbon monoxide radioactivity on cerebral blood flow and oxygen extraction fraction in a dual-tracer autoradiographic method. Ann Nucl Med. 2009;23(4):363–71.

    Article  CAS  Google Scholar 

  13. Joshi AD, Koeppe RA, Fessler JA, Kilbourn MR. Signal separation and parameter estimation in noninvasive dual-tracer PET scans using reference-region approaches. J Cereb Blood Flow Metab. 2009;29(7):1346–57.

    Article  Google Scholar 

  14. Kadrmas DJ, Rust TC, Hoffman JM. Single-scan dual-tracer FLT+FDG PET tumor characterization. Phys Med Biol. 2013;58(3):429–49. https://doi.org/10.1088/0031-9155/58/3/429.

    Article  Google Scholar 

  15. Cheng X, Li Z, Liu Z, Navab N, Huang S-C, Keller U, Ziegler SI, Shi K. Direct parametric image reconstruction in reduced parameter space for rapid multi-tracer PET imaging. IEEE Trans Med Imaging. 2015;34(7):1498–512.

    Article  Google Scholar 

  16. Zhang JL, Morey AM, Kadrmas DJ. Application of separable parameter space techniques to multi-tracer PET compartment modeling. Phys Med Biol. 2016;61(3):1238. https://doi.org/10.1088/0031-9155/61/3/1238.

    Article  CAS  Google Scholar 

  17. Gao F, Liu H, Jian Y, Shi P. Dynamic dual-tracer PET reconstruction. Inf Process Med Imaging. 2009;21:38–49.

    Google Scholar 

  18. Kadrmas DJ, Rust TC. Feasibility of rapid multitracer PET tumor imaging. IEEE Trans Nucl Sci. 2005;52(5):1341–7.

    Article  Google Scholar 

  19. El Fakhri G, Trott CM, Sitek A, Bonab A, Alpert NM. Dual-tracer PET using generalized factor analysis of dynamic sequences. Mol Imaging Biol. 2013;15(6):666–74.

    Article  Google Scholar 

  20. Verhaeghe J, Reader AJ. Simultaneous water activation and glucose metabolic rate imaging with PET. Phys Med Biol. 2013;58(3):393–411.

    Article  Google Scholar 

  21. Andreyev A, Celler A. Dual-isotope PET using positron-gamma emitters. Phys Med Biol. 2011;56(14):4539–56.

    Article  CAS  Google Scholar 

  22. Fukuchi T, Okauchi T, Shigeta M, Yamamoto S, Watanabe Y, Enomoto S. Positron emission tomography with additional γ-ray detectors for multiple-tracer imaging. Med Phys. 2017;44(6):2257–66. https://doi.org/10.1002/mp.12149.

    Article  CAS  Google Scholar 

  23. Fukuchi T, Shigeta M, Haba H, Mori D, Yokokita T, Komori Y, Yamamoto S, Watanabe Y. Image reconstruction method for dual-isotope positron emission tomography. J Instrum. 2021;16(01):01035. https://doi.org/10.1088/1748-0221/16/01/P01035.

    Article  Google Scholar 

  24. Ruan D, Liu H. Separation of a mixture of simultaneous dual-tracer PET signals: a data-driven approach. IEEE Trans Nucl Sci. 2017;64(9):2588–97.

    Article  CAS  Google Scholar 

  25. Xu J, Liu H. Deep-learning-based separation of a mixture of dual-tracer single-acquisition PET signals with equal half-lives: a simulation study. IEEE Trans Radiat Plasma Med Sci. 2019;3(6):649–59.

    Article  Google Scholar 

  26. Qing M, Wan Y, Huang W, Xu Y, Liu H. Separation of dual-tracer PET signals using a deep stacking network. Nucl Instrum Methods Phys Res A. 2021;1013: 165681. https://doi.org/10.1016/j.nima.2021.165681.

    Article  CAS  Google Scholar 

  27. Tong J, Wang C, Liu H. Temporal information-guided dynamic dual-tracer PET signal separation network. Med Phys. 2022;49(7):4585–98. https://doi.org/10.1002/mp.15566.

    Article  Google Scholar 

  28. Lian D, Li Y, Liu H. Spatiotemporal attention constrained deep learning framework for dual-tracer PET imaging. In: Yang G, Aviles-Rivero A, Roberts M, Schönlieb C-B, editors. Medical image understanding and analysis. Cham: Springer; 2022. p. 87–100.

    Chapter  Google Scholar 

  29. Xu J, Liu H. Three-dimensional convolutional neural networks for simultaneous dual-tracer PET imaging. Phys Med Biol. 2019;64(18): 185016.

    Article  CAS  Google Scholar 

  30. Ding W, Yu J, Zheng C, Fu P, Huang Q, Feng DD, Yang Z, Wahl RL, Zhou Y. Machine learning-based noninvasive quantification of single-imaging session dual-tracer 18F-FDG and 68Ga-DOTATATE dynamic PET-CT in oncology. IEEE Trans Med Imaging. 2022;41(2):347–59. https://doi.org/10.1109/TMI.2021.3112783.

    Article  Google Scholar 

  31. Haggstrom I, Schmidtlein CR, Campanella G, Fuchs TJ. DeepPET: a deep encoder-decoder network for directly solving the PET image reconstruction inverse problem. Med Image Anal. 2019;54:253–62.

    Article  Google Scholar 

  32. Ruder S. An overview of multi-task learning in deep neural networks. 2017. arxiv:arXiv:1706.05098.

  33. Ranjan R, Patel VM, Chellappa R. Hyperface: a deep multi-task learning framework for face detection, landmark localization, pose estimation, and gender recognition. IEEE Trans Pattern Anal Mach Intell. 2019;41(1):121–35. https://doi.org/10.1109/TPAMI.2017.2781233.

    Article  Google Scholar 

  34. Wachinger C, Reuter M, Klein T. DeepNAT: deep convolutional neural network for segmenting neuroanatomy. Neuroimage. 2018;170:434–45. https://doi.org/10.1016/j.neuroimage.2017.02.035.

    Article  Google Scholar 

  35. Iriarte A, Marabini R, Matej S, Sorzano COS, Lewitt RM. System models for PET statistical iterative reconstruction: a review. Comput Med Imaging Graph. 2016;48:30–48.

    Article  CAS  Google Scholar 

  36. Wang Z, Bovik AC, Sheikh HR, Simoncelli EP. Image quality assessment: from error visibility to structural similarity. IEEE Trans Image Process. 2004;13(4):600–12. https://doi.org/10.1109/TIP.2003.819861.

    Article  Google Scholar 

  37. Wang Z, Simoncelli EP, Bovik AC. Multi-scale structural similarity for image quality assessment. In: Conference record of the thirty-seventh asilomar conference on signals, systems & computers, vols 1 and 2; 2003. p. 1398–1402.

  38. Zubal IG, Harrell CR, Smith EO, Rattner Z, Gindi G, Hoffer PB. Computerized three-dimensional segmented human anatomy. Med Phys. 1994;21(2):299–302. https://doi.org/10.1118/1.597290.

    Article  CAS  Google Scholar 

  39. Gunn RN, Gunn SR, Cunningham VJ. Positron emission tomography compartmental models. J Cereb Blood Flow Metab. 2001;21(6):635–52. https://doi.org/10.1097/00004647-200106000-00002.

    Article  CAS  Google Scholar 

  40. Feng D, Huang SC, Wang X. Models for computer simulation studies of input functions for tracer kinetic modeling with positron emission tomography. Int J Biomed Comput. 1993;32(2):95–110.

    Article  CAS  Google Scholar 

  41. Wang B, Ruan D, Liu H. Noninvasive estimation of macro-parameters by deep learning. IEEE Trans Radiat Plasma Med Sci. 2020;4(6):684–95. https://doi.org/10.1109/TRPMS.2020.2979017.

    Article  Google Scholar 

  42. Chen S, Ho C, Feng D, Chi Z. Tracer kinetic modeling of 11C-acetate applied in the liver with positron emission tomography. IEEE Trans Med Imaging. 2004;23(4):426–32. https://doi.org/10.1109/TMI.2004.824229.

    Article  Google Scholar 

  43. Muzic J, Raymond F, Cornelius S. COMKAT: compartment model kinetic analysis tool. J Nucl Med. 2001;42(4):636–45.

    CAS  Google Scholar 

  44. Fessler JA. Michigan image reconstruction toolbox. https://web.eecs.umich.edu/~fessler/code/.

  45. Kemp BJ, Hruska CB, McFarland AR, Lenox MW, Lowe VJ. NEMA NU 2-2007 performance measurements of the Siemens Inveon preclinical small animal PET system. Phys Med Biol. 2009;54(8):2359–76. https://doi.org/10.1088/0031-9155/54/8/007.

    Article  Google Scholar 

  46. Hudson HM, Larkin RS. Accelerated image reconstruction using ordered subsets of projection data. IEEE Trans Med Imaging. 1994;13(4):601–9. https://doi.org/10.1109/42.363108.

    Article  CAS  Google Scholar 

  47. Haerle SK, Fischer DR, Schmid DT, Ahmad N, Huber GF, Buck A. 18F-FET PET/CT in advanced head and neck squamous cell carcinoma: an intra-individual comparison with 18F-FDG PET/CT. Mol Imaging Biol. 2011;13(5):1036–42.

    Article  Google Scholar 

  48. Black NF, McJames S, Rust TC, Kadrmas DJ. Evaluation of rapid dual-tracer (62)Cu-PTSM + (62)Cu-ATSM PET in dogs with spontaneously occurring tumors. Phys Med Biol. 2008;53(1):217–32.

    Article  Google Scholar 

Download references

Acknowledgements

We gratefully acknowledge Youqin Xu, Yang He and Junyi Tong for their technical assistance in the animal PET experiments.

Funding

This work was supported in part by the National Key Research and Development Program of China (No: 2020AAA0109502), by the National Natural Science Foundation of China (No: U1809204), by the Talent Program of Zhejiang Province (2021R51004), and by the Key Research and Development Program of Zhejiang Province (No: 2021C03029).

Author information

Authors and Affiliations

Authors

Contributions

FZ: study design, data acquisition, data analysis; JF: data acquisition, data analysis, manuscript writing; AM: data analysis; HL: study design, manuscript revision. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Huafeng Liu.

Ethics declarations

Ethics approval and consent to participate

Animal experiments were approved by the Experimental Animal Ethics Committee of Southern Medical University, and were carried out in compliance with the ARRIVE guidelines, the guidelines of the US National Institutes of Health and local legal requirements.

Consent for publication

Not applicable.

Competing of interests

The authors declare that they have no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zeng, F., Fang, J., Muhashi, A. et al. Direct reconstruction for simultaneous dual-tracer PET imaging based on multi-task learning. EJNMMI Res 13, 7 (2023). https://doi.org/10.1186/s13550-023-00955-w

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s13550-023-00955-w

Keywords