Skip to main content
  • Methodology article
  • Open access
  • Published:

Joint pre-processing framework for two-dimensional gel electrophoresis images based on nonlinear filtering, background correction and normalization techniques

Abstract

Background

Two-dimensional gel electrophoresis (2-DGE) is a commonly used tool for proteomic analysis. This gel-based technique separates proteins in a sample according to their isoelectric point and molecular weight. 2-DGE images often present anomalies due to the acquisition process, such as: diffuse and overlapping spots, and background noise. This study proposes a joint pre-processing framework that combines the capabilities of nonlinear filtering, background correction and image normalization techniques for pre-processing 2-DGE images. Among the most important, joint nonlinear diffusion filtering, adaptive piecewise histogram equalization and multilevel thresholding were evaluated using both synthetic data and real 2-DGE images.

Results

An improvement of up to 46% in spot detection efficiency was achieved for synthetic data using the proposed framework compared to implementing a single technique of either normalization, background correction or filtering. Additionally, the proposed framework increased the detection of low abundance spots by 20% for synthetic data compared to a normalization technique, and increased the background estimation by 67% compared to a background correction technique. In terms of real data, the joint pre-processing framework reduced the false positives up to 93%.

Conclusions

The proposed joint pre-processing framework outperforms results achieved with a single approach. The best structure was obtained with the ordered combination of adaptive piecewise histogram equalization for image normalization, geometric nonlinear diffusion (GNDF) for filtering, and multilevel thresholding for background correction.

Introduction

A commonly used gel-based approach for proteomic analysis is two-dimensional gel electrophoresis (2-DGE), a technique that separates proteins in a sample based on both their isoelectric point and molecular weight [1]. This technique is often used in preliminary comparative proteomic analyses, as it is capable of resolving thousands of proteins in a single run. Once the proteins in the sample have been separated, the gel is then scanned and the imaged processed using computational tools. Often these 2-DGE images exhibit anomalies due to the technique itself or to the image scan and acquisition [2]. The purpose of 2-DGE image analysis is to detect the proteins (black spots) within the gel. However, a noisy background with variable intensity, diffuse or low-intensity spots, and over-saturated spots often hinder the detection of individual proteins. Therefore, a pre-processing step that minimizes these anomalies is an open issue in the literature, as an important phase prior to analysis of these kinds of images [3].

Pre-processing techniques for 2-DGE image analysis are classified as: image normalization, background correction, and noise reduction techniques [3, 4]. Image normalization improves the detection of low abundance proteins (low-intensity spots) [5]. Satisfactory image normalization results are achieved using multiple gels, obtaining a pattern that is compared with each sample; however, aligning the multiple images is the main difficulty of this technique [6]. On the other hand, the aim of background correction is to increase contrast and decrease the effects of non-homogeneous regions, thus improving spot detection. In the literature, there are several background correction techniques reported for 2-DGE image processing, such as adjustment by either local or global minima, polynomial adjustment, and approaches based on histograms [6, 7]. Despite the advances in normalization and background correction techniques, noise reduction approaches have been the most studied for 2-DGE image pre-processing. We found several linear and nonlinear filters used for noise reduction of 2-DGE images [3, 4]. Usually, linear filters blur the spots and reduce their intensities, which is not optimal as it alters the end results [8]. Thus, it is common to use nonlinear filters, such as filters based on Wavelet [3], Contourlet [9] and total variation (TV) [10]. The most commonly used nonlinear filtering technique for 2-DGE is based on Wavelet transform, which achieves high noise reduction; however, with this technique it is difficult to preserve the spot contours [3, 4]. On the other hand, TV preserves better spot edges due to a smoothing variable operation, but is limited in terms of noise reduction [10]. Contourlet transform also performs better than Wavelet in preserving edge information [9]. Xin and Zhao [11] used a combined version of Wavelet and TV (WTTV) to reduce information loss in 2-DGE image pre-processing. In a previous work [4], we presented a comparison between Wavelet, Contourlet, TV, and WTTV filters using synthetic and real 2-DGE images, showing that with synthetic data, Wavelet and WTTV had the lowest sensitivity to noise levels, while wavelet presented the best detection rate for known proteins on real 2-DGE images. However, these results were obtained by executing each technique separately and a joint framework was not considered.

Noise reduction, image normalization and background correction techniques reduce specific anomalies in 2-DGE images. For example, noise reduction minimizes the effect of impulsive and white noise; image normalization normalizes over-saturated and low abundance spots, as well as light saturation; and background correction reduces variability, saturation and streaking. Since each approach reduces a specific anomaly in 2-DGE images, it is necessary to combine them in order to enhance the spots in the image. This paper discusses a joint framework that combines the capabilities of image normalization, background correction and nonlinear filtering. Since there are several techniques for each approach, we first present a comparative study using both synthetic and real 2-DGE images and then we evaluate the combined framework. For this comparison, we used four metrics to evaluate the performance of the techniques applied to synthetic data, and we evaluated their capabilities in reducing anomalies in real 2-DGE images.

Pre-processing framework for 2-DGE images

In the proposed framework, the first step is image normalization. This step improves the contrast of protein spots, mainly low intensity ones. As in the literature there are several normalization techniques, we compared three enhancement techniques: histogram equalization, adaptive piece-wise histogram equalization [12], and a modification of background pixel intensity [7].

As mentioned previously, image normalization improves the contrast of low intensity protein spots; however, it also increases both the intensity of isolated points and impulsive noise. Therefore, in the proposed joint pre-processing framework, noise reduction is the second step in the process. For noise reduction, nonlinear filtering techniques are recommended for low edge distortion. A comparison of the most commonly used nonlinear techniques for 2-DGE image is presented in [4]. Quantitative comparison showed that Wavelet filtering performs better than Counterlet, TV, and WTTV. However, the results in [4] showed that with Wavelet there was less noise reduction but edge information was better preserved than with other techniques. In this paper, we evaluate the use of geometric nonlinear diffusion filtering (GNDF) for the pre-processing of 2-DGE images [13].

Finally, background correction techniques achieve better results when processing images with low levels of noise, therefore it is the last step in the pre-processing framework. We compared thresholding, multilevel thresholding [7] and surface approximation [14].

Image normalization

The histogram is an estimation of the probability of occurrence of grey levels in an image. The histogram is given by [15]:

$$ p(k)=\frac{n_{k}}{n}\quad k=0,1,\ldots,L-1 $$
(1)

where n is the total number of pixels in the image, nk is the number of pixels with grey levels equal to k, L is the number of possible grey levels, and p(k) is the probability of occurrence of k. Histogram equalization is an image transformation that approaches the probability of occurrence of grey levels to a uniform probability density function. This transformation improves the use of the dynamic range for grey levels, thus improving contrast. From the histogram, the histogram equalization is obtained by computing the function Sk given by:

$$ s(k)=\sum_{j=0}^{k}\frac{n_{k}}{n}\quad k=0,1,\ldots,L-1 $$
(2)

and then mapping each pixel with level k in the equalized image with a pixel value equal to ⌊(L−1)Sk⌋.

Given that pixel intensities behave randomly due to the type of sample and the acquisition process, an adaptive piecewise histogram equalization is proposed in [12]. This technique performs multiple histogram equalizations considering the maximum and minimum intensity levels. Further details of the algorithm are in [12].

Another way to perform image normalization is to modify the background pixel intensity [7]. The background of the image is estimated using a threshold and then it is subtracted from the data.

Nonlinear filtering

GNDF [13] reduces noise while preserving edge information, so it is expected to improve spot detection in 2-DGE image analysis. GNDF solves a nonlinear differential partial equation given by:

$$ \frac{\partial I}{\partial t}=\frac{d}{dx}[C|\nabla I|*\nabla I] $$
(3)

where the initial condition I(t=0) is the 2-DGE image, ∇I is the image gradient, and C are the diffusion coefficients defined as:

$$ C(x)=\frac{1}{1+(x/k)^{2}} $$
(4)

where k is a threshold that determines the level of noise to be removed. The estimation of k is obtained from the signal to noise ratio of the image [13].

In addition to GNDF, in this study we used Wavelet Transform for noise reduction. A comparison of WT and other filtering techniques is presented in [4]. We use WT with a Daubechies family and 5 levels of decomposition [2, 4].

Background correction

We compared three background correction techniques: thresholding, multilevel thresholding [7] and surface approximation [14]. Thresholding estimates the intensities of background pixels to be subtracted from the image. Since most of the time the background of 2-DGE images is not homogeneous, techniques such as multilevel thresholding can yield better results. Multilevel thresholding divides the image into several regions, and in each region we can estimate the intensities of the background pixels. For this paper, two levels Gf1 and Gf2 are used:

$$\begin{array}{@{}rcl@{}} G_{f1} = I \in \left(I_{i}< \frac{G_{1}}{n_{1}} \right) \quad &\text{where}& \quad G_{1}= \sum P_{x}(0, \widetilde{P_{x}}) \end{array} $$
(5)
$$\begin{array}{@{}rcl@{}} G_{f2} = I \in \left(\frac{G_{1}}{n_{1}} < I_{i}< \frac{G_{2}}{n_{2}} \right) \quad &\text{where}& \quad G_{2}= \sum P_{x}(\widetilde{P_{x}}, maxP_{x}) \end{array} $$
(6)

where Gf1 is the first level, with pixels of intensities between the minimum grey level and the median of a percentile Px (\(\widetilde {P_{x}}\)), and Gf2 is the second level with pixels of intensities between \(\widetilde {P_{x}}\) and the maximum value of the percentile maxPx.

A third method used in this paper for background correction is surface approximation [7]. A B-Spline surface is used to estimate background with the iterative algorithm presented in [7].

Experiments

Databases

Database 1: synthetic dataset

Synthetic proteins were modelled as two-dimensional Gaussian distributions [16], assuming the media, μ, and standard deviation, σ, are equal for both dimensions. Size and scattering for a protein are varied through σ. Protein location within a synthetic image was randomly generated using a uniform distribution. The random distribution generated some overlapping spots. Gaussian, Rayleigh and exponential noise, given by (7), (8) and (9) respectively, were added to the synthetic images. The parameters presented in [4] were used for each noise in order to simulate images with signal-to-noise ratio -SNR between 8 and 20 db.

$$\begin{array}{@{}rcl@{}} p(z) &=& \frac{1}{ \sqrt{2 \pi \sigma} } \exp^{-(z-\mu)^{2}/ 2 \sigma^{2}} \end{array} $$
(7)
$$\begin{array}{@{}rcl@{}} p(z) &=& \frac{2}{b}(z-a)\exp^{-(z-a)^{2}/ b} \end{array} $$
(8)
$$\begin{array}{@{}rcl@{}} p(z) &=& a \exp^{-az} \end{array} $$
(9)

Database 2: ITM 2-DGE image database

This dataset was collected from previous studies carried out in the Laboratory of Molecular and Cell Biology of the Instituto Tecnologico Metropolitano ITM of Medellin (Colombia). The 2-DGE images correspond to two different sample types:

  1. a)

    Bee venom collected from africanized worker bees (samp_01–02–03 and 04).

  2. b)

    Urine samples taken from patients with prostate cancer (samp_05 and 06).

Proteins (50 μg) were loaded by passive re-hydration onto 7 cm ZOOM ® IPG (Immobilized pH gradient) strips pH 3-10 NL (ThermoFisher Scientific). Isoelectric focusing was carried out using the following voltage ramp: 200−450−600−750−950 V during 25 min, 1200−1400−1600 V during 30 min, and 2000 V during 45 min [17]. For the second dimension, the IPG strips were loaded onto SDS-PAGE NuPAGE™ 4−12% Bis-Tris Protein Gels, 1.5 mm (ThermoFisher Scientific) and run at 200 V during 40 min. After electrophoresis these were stained with SYPRO™ Ruby (Invitrogen™, ThermoFisher Scientific) and the gel images were acquired using the ChemiDoc MP System (Bio-Rad). Gel images were analyzed and compared using the PDQuest Advanced 2-D Software (Bio-Rad).

Database 3: lECB 2-D PAGE gel image database

This database consist of four 2-DGE image data sets previously analyzed with the GELLAB-II system [18]. These data sets consist of over 300 gel images (gif format) with annotations and landmark data in html, tab-delimited and xml formats. The data sets and experimental conditions are described and documented in the papers associated with each data set [19–22]. From this database, four 2-DGE images were randomly selected for this study, one from each data set:

  1. 1.

    Human leukemias/gel-HM-029 (samp_07)

  2. 2.

    HL-60 cell line/gel-HL60-HUM-MYEL-DIFF-029 (samp_08)

  3. 3.

    MOLT-4 cell line/gel-MOLT-4-004 (samp_09)

  4. 4.

    Fetal Alcohol Syndrome (FAS) - serum/gel-FAS-NA-NA-001 (samp_10)

This database is available for public use and can be downloaded from

http://www.bioinformatics.org/lecb2dgeldb/.

Validation measures

In this study four indicators were used for evaluating the performance of pre-processing techniques. For evaluating normalization, we used the percentage of low-abundance proteins detected (LPD) defined as the ratio between the number of low-abundance spots detected (LAS det) and the total number of low-abundance spots (LAS tot) in the image:

$$ \text{LPD} = \frac{\text{LAS}_{det}}{\text{LAS}_{tot}} $$
(10)

In the case of noise reduction techniques, the signal to noise ratio (SNR), based on the normalized mean square error (MSE n), was used and can be given by:

$$\begin{array}{@{}rcl@{}} \text{MSE}_{n} &=& \frac{\sum_{i=1}^{n}(x_{i}- \widehat{x}_{i})^{2}}{\sum_{i=1}^{n}(x_{i})^{2}} \end{array} $$
(11)
$$\begin{array}{@{}rcl@{}} \text{SNR} &=& 10*\log_{10}\frac{1}{\text{MSE}_{n}} \end{array} $$
(12)

where xi is a pixel in the original image and \(\widehat {x}_{i}\) is the same pixel in the filtered image. Additionally, spot efficiency (Ξ) was used to evaluate the performance of noise reduction techniques, in terms of the number of true detected spots (ςt), false detected spots (ςf) and lost spots (ςl) [3, 4]:

$$ \Xi = \frac{\varsigma_{t} - \varsigma_{f}}{\varsigma_{t} + \varsigma_{l}} $$
(13)

Finally, the background correction methods were evaluated using the background subtraction index (BSI), which was calculated in terms of the number of detected pixels that belong to the background (ϱdet) and the total number of pixels that belong to the background (ϱtot). Thus, BSI means the percentage of pixels identified as background:

$$ \text{BSI} = \frac{\varrho_{det}}{\varrho_{tot}} $$
(14)

Proposed approach

According to the measures expressed by (10), (12), (13) and (14), several configurations of stages for normalization, noise reduction and background correction were tested in a sequential structure made up by three stages, named in this work as the joint pre-processing framework. In this sense, the order of the stages was an important aspect to evaluate and the performance of several techniques in each stage was registered, in order to find the most effective structure configuration, which was validated by experts. It is important to note that the training was executed using synthetic images, but the validation was performed using real 2-DGE images, where the algorithm results were compared with the expert’s opinions.

Results and discussion

Comparison of normalization techniques

As image normalization seeks to enhance low-abundance proteins, we used a synthetic image with these kinds of spots (see Fig. 1a). The synthetic image had 1024 x 1024 pixels, with an opaque background and 150 spots, which were generated by a Gaussian distribution with standard deviation between 0.3 and 0.8. The spot intensity was controlled to simulate low-abundance proteins with a grey level between 0.1 and 0.8. We compared histogram equalization, adaptive piecewise histogram equalization [12], and a modification of background pixel intensity [7] for image normalization, and used the percentage of low-abundance proteins detected (LPD) to evaluate the performance of each technique.

Fig. 1
figure 1

Synthetic protein spots modelled as a 2-D Gaussian distribution. a Example of a synthetic image. b Synthetic image normalized using histogram equalization. c Synthetic image normalized using adaptive piecewise histogram equalization. d Synthetic image normalized using modification of background pixel intensity

The LPD results are presented in Table 1. The technique based on background pixel intensity detected only 48.7% of low-abundance spots. On the other hand, the histogram and adaptive piecewise histogram equalizations detected 82.1% and 88.9% of low abundance spots, respectively. As can be seen in Fig. 1b and c, the techniques based on equalization enhanced the contrast of the low-abundance spots.

Table 1 Performance of image normalization techniques for a synthetic image with low-abundance spots evaluated using LPD

Figure 2 presents the normalization results for a real 2-DGE image (samp_05). The equalization-based approach improves contrast by increasing the grey level intensity of the protein spots and decreasing the intensity of the background pixels (see Fig. 2b and c). However, normalization also increases the background noise, so it was necessary to combine image normalization with a noise reduction technique.

Fig. 2
figure 2

Two-dimensional gel electrophoresis – 2-DGE – image from a human urine sample (samp_05). a Original image. b 2-DGE image normalized using histogram equalization. c 2-DGE image normalized using adaptive piecewise histogram equalization. d 2-DGE image normalized using modification of background pixel intensity

Comparison of noise reduction techniques

Wavelet transform (WT) is one of the nonlinear filters that presents the best performance for noise reduction in 2-DGE images [4]. However, there are other nonlinear methods that allow noise reduction without smoothing spot edges. We compared WT with geometric nonlinear diffusion filtering - GNDF. GNDF has been shown to perform well with several types of medical images but has not been used with 2-DGE images. For WT filter, a Daubechies wavelet family was used with five decomposition levels [4]. For GNDF, we used 35 smoothing iterations with a diffusion coefficient equal to 0.2 and windows of 5x5 pixels. The performance was evaluated using the signal-to-noise ratio (SNR) and spot efficiency [4]. WT and GNDF were tested with synthetic images with Gaussian, Rayleigh and exponential noise with SNR from 20 to 8 dB. Each synthetic image has 512x512 pixels with 250 spots.

Table 2 presents the spot efficiency comparison using WT and GNDF filters for the synthetic images with noise. In terms of spot efficiency, WT and GNDF yielded very similar results for most noise levels, with differences close to 2%. However, for the synthetic image with Gaussian noise of 8 dB (i.e. the higher noise level), GNDF presented a spot efficiency of 77.86%, while WT obtained 67.5%. On the other hand, better results were obtained by GNDF in terms of SNR. Table 3 shows the SNR comparison for WT and GNDF filters. In the case of the image with SNR of 8dB, WT obtained images with 19.31 dB, 9.78 dB and 12.71 dB for the Gaussian, Rayleigh and exponential noise respectively; while GNDF obtained images with 20.11 dB, 10.5 dB and 15.61 dB for Gaussian, Rayleigh and exponential noise respectively.

Table 2 Performance of noise reduction techniques evaluated using spot efficiency (%)
Table 3 Performance of noise reduction techniques evaluated using SNR (dB)

Both nonlinear filtering techniques, WT and GNDF, were applied to real 2-DGE images (samp_05). As can be seen in the results in Fig. 3, the effect of filtering can be noted in the background, as GNDF reduces the background noise while preserving the spot contours.

Fig. 3
figure 3

Two-dimensional gel electrophoresis – 2-DGE – image from a human urine sample (samp_05). a Original image. b 2-DGE image filtered using Wavelet Transform -WT. c 2-DGE image filtered using geometric nonlinear diffusion filtering - GNDF

Comparison of background correction

We compared three background correction techniques: thresholding, multilevel thresholding [7] and surface approximation [14]. First, we generated a synthetic image with changes in background intensity (see Fig. 4a). The background variation was obtained by increasing the initial intensity up to 155%. A percentile of 60% was used for both thresholding techniques. A B-Spline equation [14] was used for the surface approximation techniques optimizing the parameters with 150 iterations. The performance was evaluated by the Subtraction Index (SI) that compares the number of background pixels with the estimated.

Fig. 4
figure 4

Synthetic protein spots modelled as a 2-D Gaussian distribution with background. a Example of a synthetic image. b Synthetic image with background correction using thresholding. c Synthetic image with background correction using multilevel thresholding. d Synthetic image with background correction using surface approximation

Figure 4 presents the background correction results in the synthetic image. Using thresholding, the background was partially removed, but as can be seen in part B of the figure, the background is divided in two regions. Conversely, a uniform background was obtained with multilevel thresholding. The surface approximation removed most of the background, but this technique did not work for pixels close to the spots. The SI results are presented in Table 4. Thresholding detected 71.8% of background pixels, while surface approximation and multi-level thresholding detected 97.9% and 98.5% of background pixels for the synthetic images respectively.

Table 4 Performance of background correction techniques for a synthetic image with variable background using BSI

Figure 5 presents the background correction for a real 2-DGE image (samp_05). Thresholding preserved background intensities around spots, but the background obtained from multi-level thresholding and surface approximation approaches was uniform and increased spot contrast. However, background noise was also preserved; hence, it is necessary to combine background correction with noise reduction techniques for pre-processing of 2-DGE images.

Fig. 5
figure 5

Two-dimensional gel electrophoresis – 2-DGE – image from a human urine sample (samp_05). a Original image. b 2-DGE image with background correction using thresholding. c 2-DGE image with background correction using multilevel thresholding. d 2-DGE image with background correction using surface approximation

Proposal novelties I: joint pre-processing framework

Based on the comparison of image normalization, noise reduction and background correction techniques, we show that a joint pre-processing framework is needed. The proposed framework takes advantage of the capabilities of image normalization to increase the contrast of low-abundance proteins, of nonlinear filtering to reduce noise while preserving edge information, and of background correction to homogenize background pixels. According to previous results, we used piecewise histogram equalization for image normalization, GNDF for filtering and multi-level thresholding for background correction. The joint pre-processing framework was evaluated using both synthetic and real 2-DGE images.

The joint pre-processing framework was evaluated using a synthetic image generated by a 2-D Gaussian distribution, where the 150 spots have a standard deviation between 0.1 and 0.8. The image includes an intensity variation in the background along the horizontal axis. Additionally, the image has Gaussian noise with a median of zero, standard deviation equal to 1.535 and Rayleigh noise with a = 0 and b = 0.0539. Table 5 presents the performance results using LPD, spot efficiency and SI. The SI metric was only computed for the images obtained from the background correction and joint pre-processing techniques, as it measures the background subtracted from the image.

Table 5 Performance of the joint pre-processing framework for a synthetic image with variable background and noise using LPD, spot efficiency (Ξ), and BSI

The best LPD was obtained using the joint pre-processing framework with 60% of low-abundance spots detected in the image. By comparison, this percentage was 40% when only the normalization technique was implemented. In terms of spot efficiency, the proposed framework detected 63.84% of spots, while lower percentages were obtained when using a single technique: 3.57% for normalization, 17.69% for the filtered image, and 6.69% using background correction. Furthermore, the best subtraction index was also obtained by the proposed framework, with a 78.62% in comparison with 11.37% using only the modified histogram-based technique for background correction.

Figure 6 presents the effects of the joint pre-processing framework in three of the real 2-DGE images (samp_05–09–10). In the three processed images (Fig. 6b, d, and f), we can see the effect of noise reduction and background homogenization. Additionally, the enhancement of low abundance spots is noticeable.

Fig. 6
figure 6

Results of the joint pre-processing framework in real two-dimensional gel electrophoresis – 2-DGE – images. a Original 2-DGE image from a human urine sample (samp_05). b 2-DGE image of samp_05 with the joint pre-processing framework. c Original 2-DGE image from Molt-4 cell line (samp_09). d 2-DGE image of samp_09 with joint pre-processing framework. e Original 2-DGE image from Fetal Alchohol Syndrome serum (samp_10). f 2-DGE image of samp_10 with joint pre-processing framework

Proposal novelties II: validation with real 2-DGE images

The joint pre-processing framework was validated using real 2-DGE images captured from four apitoxin (honey bee venom) samples, two urine samples from patients with prostate cancer, and four 2D images from the LECB 2-D PAGE Gel Image Database. Table 6 presents the number of detected spots from the original and pre-processed samples, as well as the true positives and false positives. We obtained the false positive reduction percentages comparing the original and pre-processed images. For the 2-DGE images of apitoxin (samp_01–02–03–04), the joint pre-processing framework reduced the false positives between 43% and 72%. For the urine samples (samp_05–06), the false positives from the pre-processed images decreased by 91% and 85% respectively. And for the four images from the LECB 2-D PAGE Gel Image Database (samp_07–08–9–10), the false positives were reduced between 71% and 93%. From these results, we can see that the joint pre-processing framework improves protein detection by reducing the false positives caused by noise and non-homogeneous background.

Table 6 True positive and false positive spots detected from original and processed real 2-DGE images

Conclusions

2-DGE images commonly present several anomalies that hinder spot detection and analysis. In this paper, the use of several digital image processing techniques were tested and validated in three stages, i.e., normalization, noise reduction and background correction, achieving an enhancement of the image for posterior analysis. Each approach helps improve specific anomalies, and here we introduce a new joint pre-processing framework that combines the capabilities of the selected techniques for each of the three stages.

The techniques used in each of the stages of image pre-processing were compared on synthetic images, using four validation measures, i.e., LPD, SNR, spot efficiency (Ξ) and BSI, which offered representative and consistent values associated with pre-processing performance, so these quantitative indicators proved to be a very useful measure for 2-DGE image applications.

Experimental results from synthetic images demonstrated that the order of the stages impacts the final results. E.g., if the noise reduction stage is executed before normalization, the faint spots, that have important information for the interpretation of the image, are often removed. Consequently, the order with the best performance was the following: 1) normalization, 2) noise reduction and 3) background correction. In particular, the best normalization technique was adaptive piecewise histogram equalization, according to the LPD validation measure. Equalization techniques enhance the contrast of low-abundance spots, but also increase the background noise. In noise reduction tests, the nonlinear technique GNDF was implemented, which is a new technique for these kinds of images and reduces noise while preserving edges. GNDF showed a similar spot efficiency (Ξ) to WT, but a better SNR using synthetic data with different types of noise. Finally, three techniques were compared for background correction using the Background Subtraction Index (BSI). The best results of BSI were obtained using multi-level thresholding.

Results with real 2-DGE images showed that the joint framework outperforms results from a single approach. According to these results, the use of adaptive piecewise histogram equalization, GNDF and multi-level thresholding, is recommended for these kinds of images. However, as future work, the joint pre-processing framework could implement other kinds of techniques for each step, that were not considered in this study.

Availability of data and materials

The datasets used and/or analyzed during the current study are available from the corresponding author on reasonable request.

Abbreviations

2-DGE:

two-dimensional gel electrophoresis

TV:

total variation

WT:

wavelet transform

WTTV:

Wavelet and total variation

GNDF:

geometric nonlinear diffusion filtering

LPD:

low-abundance protein detected

SI:

subtraction index

SNR:

signal-to-noise ratio

References

  1. Færgestad EM, Rye M, Walczak B, Gidskehaug L, Wold JP, Grove H, Jia X, Hollung K, Indahl UG, Westad F, Van Den Berg F, Martens H. Pixel-based analysis of multiple images for the identification of changes: A novel approach applied to unravel proteome patters of 2-D electrophoresis gel images. Proteomics. 2007; 7(19):3450–61. https://0-doi-org.brum.beds.ac.uk/10.1002/pmic.200601026.

    Article  Google Scholar 

  2. Sengar RS, Upadhyay AK, Singh M, Gadre VM. Segmentation of two dimensional electrophoresis gel image using the wavelet transform and the watershed transform. In: 2012 National Conference on Communications, NCC 2012: 2012. https://0-doi-org.brum.beds.ac.uk/10.1109/NCC.2012.6176861.

  3. Kaczmarek K, Walczak B, De Jong S, Vandeginste BGM. Preprocessing of two-dimensional gel electrophoresis images. In: Proteomics: 2004. p. 2377–89. https://0-doi-org.brum.beds.ac.uk/10.1002/pmic.200300758.

  4. Goez MM, Torres-Madroñero MC, Röthlisberger S, Delgado-Trejos E. Preprocessing of 2-Dimensional Gel Electrophoresis Images Applied to Proteomic Analysis: A Review. Beijing Genomics Inst. 2018. https://0-doi-org.brum.beds.ac.uk/10.1016/j.gpb.2017.10.001.

  5. Keeping AJ, Collins RA. Data Variance and statistical significance in 2D-gel electrophoresis and DIGE experiments: Comparison of the effects of normalization methods. J Proteome Res. 2011; 10(3):1353–60. https://0-doi-org.brum.beds.ac.uk/10.1021/pr101080e.

    Article  CAS  Google Scholar 

  6. Rye M, Fargestad EM. Preprocessing of electrophoretic images in 2-DE analysis. Chemometr Intell Lab Syst. 2012; 117:70–79. https://0-doi-org.brum.beds.ac.uk/10.1016/j.chemolab.2011.09.012.

    Article  CAS  Google Scholar 

  7. Sarkar S, Das S. Multilevel image thresholding based on 2D histogram and maximum tsallis entropy - A differential evolution approach. IEEE Trans Image Process. 2013; 22(12):4788–97. https://0-doi-org.brum.beds.ac.uk/10.1109/TIP.2013.2277832.

    Article  Google Scholar 

  8. Dowsey AW, Dunn MJ, Yang GZ. The role of bioinformatics in two-dimensional gel electrophoresis. In: Proteomics: 2003. p. 1567–96. https://0-doi-org.brum.beds.ac.uk/10.1002/pmic.200300459.

  9. Do MN, Vetterli M. The contourlet transform: An efficient directional multiresolution image representation. IEEE Trans Image Process. 2005; 14(12):2091–106. https://0-doi-org.brum.beds.ac.uk/10.1109/TIP.2005.859376.

    Article  Google Scholar 

  10. Chan TF, Osher S, Shen J. The digital TV filter and nonlinear denoising. IEEE Trans Image Process. 2001; 10(2):231–41. https://0-doi-org.brum.beds.ac.uk/10.1109/83.902288.

    Article  CAS  Google Scholar 

  11. Xin H, Zhao F. Effective denoising methods for two-dimensional gel electrophoresis images. In: Proceedings - 2011 4th International Conference on Biomedical Engineering and Informatics, BMEI 2011, vol 3: 2011. p. 1571–1574. https://0-doi-org.brum.beds.ac.uk/10.1109/BMEI.2011.6098614.

  12. Ling Z, Liang Y, Wang Y, Shen H, Lu X. Adaptive extended piecewise histogram equalisation for dark image enhancement. IET Image Process. 2015; 9(11):1012–9. https://0-doi-org.brum.beds.ac.uk/10.1049/iet-ipr.2014.0580.

    Article  Google Scholar 

  13. Gerig G, Kbler O, Kikinis R, Jolesz FA. Nonlinear Anisotropic Filtering of MRI Data. IEEE Trans Med Imaging. 1992; 11(2):221–32. https://0-doi-org.brum.beds.ac.uk/10.1109/42.141646.

    Article  CAS  Google Scholar 

  14. Huang J, Wang B, Wang W, Sen P. A Surface Approximation Method for Image and Video Correspondences. IEEE Trans Image Process. 2015; 24(12):5100–13. https://0-doi-org.brum.beds.ac.uk/10.1109/TIP.2015.2462029.

    Article  Google Scholar 

  15. Gonzalez RC, Woods RE. Digital Image Processing. 4th ed. New York: Pearson; 2018.

    Google Scholar 

  16. Natale M, Caiazzo A, Bucci EM, Ficarra E. A Novel Gaussian Extrapolation Approach for 2D Gel Electrophoresis Saturated Protein Spots. Genomics Proteomics Bioinforma. 2012; 10(6):336–44. https://0-doi-org.brum.beds.ac.uk/10.1016/j.gpb.2012.06.005.

    Article  CAS  Google Scholar 

  17. Pineda-Guerra Y, Betancur-Echeverri J, Pedroza-Diaz J, Delgado-Trejos E, Rothlisberger S. Proteomic analysis of africanized bee venom: a comparison of protein extraction methods. Acta Biol Colomb. 2016; 21(3):619–26.

    Article  Google Scholar 

  18. Lemkin PF. The GELLAB-II 2D Gel Exploratory Analysis System. Washington, DC: National Cancer Institute; 1993. Reference manual, pp 677.

    Google Scholar 

  19. Lester EP, Lemkin PF, Lipkin LE. A two-dimensional gel analysis of autologous T and B lymphoblastoid cell lines. Clin Chem. 1982; 28(4 Pt 2):828–39.

    Article  CAS  Google Scholar 

  20. Lester EP, Lemkin PF, Lipkin LE. Protein indexing in leukemias and lymphomas. Ann N Y Acad Sci. 1984; 428:158–72.

    Article  CAS  Google Scholar 

  21. Lester EP, Lemkin PF, Lipkin LE, Cooper HL. Computer-assisted analysis of two-dimensional electrophoreses of human lymphoid cells. Clin Chem. 1980; 26(10):1392–402.

    Article  CAS  Google Scholar 

  22. Robinson MK, Myrick JE, Henderson LO, Coles CD, Powell MK, Orr GA, Lemkin PF. Two-dimensional protein electrophoresis and multiple hypothesis testing to detect potential serum protein biomarkers in children with fetal alcohol syndrome. Electrophoresis. 1995; 16(7):1176–83.

    Article  CAS  Google Scholar 

Download references

Acknowledgements

The authors would like to thank the Smart Machine and Pattern Recognition Laboratory (MIRP Lab) and the Measurement Analysis and Decision Support Laboratory (AMYSOD Lab) of Parque i, Medellin, Colombia. We also acknowledge Johanna Pedroza-Díaz for contributing the 2-DGE images from urine samples.

Funding

This work was supported by the Ministry of Science, Technology and Innovation (MinCiencias) of the Republic of Colombia (Grant No. 115077758163, Contract RC693-2018) awarded to SR and the Instituto Tecnologico Metropolitano ITM of Medellin, Colombia (Grant No. P14227) awarded to SR. The funding bodies had no role in the design of the study and collection, analysis, and interpretation of data, or in writing the manuscript.

Author information

Authors and Affiliations

Authors

Contributions

MG carried out the algorithm implementations, performed the experiments and helped draft the manuscript. MT participated in the design of the study and drafted the manuscript. SR and ED formulated the study and participated in its design. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Maria C. Torres-Madronero.

Ethics declarations

Ethics approval and consent to participate

The six 2-DGE images used from the ITM 2-DGE Image Database were obtained from two previous studies approved by the Ethics Committee for Scientific Research of the Instituto Tecnologico Metropolitano ITM of Medellin, Colombia, in sessions held on October 19, 2010 and August 14, 2014. The images in this database have no personal or identifying information attached.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Goez, M.M., Torres-Madronero, M.C., Rothlisberger, S. et al. Joint pre-processing framework for two-dimensional gel electrophoresis images based on nonlinear filtering, background correction and normalization techniques. BMC Bioinformatics 21, 376 (2020). https://0-doi-org.brum.beds.ac.uk/10.1186/s12859-020-03713-0

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://0-doi-org.brum.beds.ac.uk/10.1186/s12859-020-03713-0

Keywords