Skip to main content
  • Research article
  • Open access
  • Published:

Multitask learning for biomedical named entity recognition with cross-sharing structure

Abstract

Background

Biomedical named entity recognition (BioNER) is a fundamental and essential task for biomedical literature mining, which affects the performance of downstream tasks. Most BioNER models rely on domain-specific features or hand-crafted rules, but extracting features from massive data requires much time and human efforts. To solve this, neural network models are used to automatically learn features. Recently, multi-task learning has been applied successfully to neural network models of biomedical literature mining. For BioNER models, using multi-task learning makes use of features from multiple datasets and improves the performance of models.

Results

In experiments, we compared our proposed model with other multi-task models and found our model outperformed the others on datasets of gene, protein, disease categories. We also tested the performance of different dataset pairs to find out the best partners of datasets. Besides, we explored and analyzed the influence of different entity types by using sub-datasets. When dataset size was reduced, our model still produced positive results.

Conclusion

We propose a novel multi-task model for BioNER with the cross-sharing structure to improve the performance of multi-task models. The cross-sharing structure in our model makes use of features from both datasets in the training procedure. Detailed analysis about best partners of datasets and influence between entity categories can provide guidance of choosing proper dataset pairs for multi-task training. Our implementation is available at https://github.com/JogleLew/bioner-cross-sharing.

Background

Biomedical named entity recognition (BioNER) aims at annotating named entity mentions with their entity types (e.g., genes, proteins [1], and diseases [2]) in the input biomedical text. The outputs of model indicate not only the locations of entity mentions but also their types. BioNER models provide useful information for downstream tasks of biomedical literature mining, such as entity relation extraction [35], and biomedical network construction [68].

BioNER task requires to detect boundaries of biomedical entities and predict their entity types. Most previous systems treat the task as a sequence labeling problem. Traditional neural network models for BioNER rely on features designed for each task. These BioNER models use hand-crafted rules [9] and domain-specific features [10], such as orthographic features, morphological features [1114]. The drawback of these neural network models is that features are specially designed for each dataset or each entity type in order to achieve good performance; thus, features used in one BioNER model may not work well in another. Recent studies showed that the neural network model is capable of feature generation work without manual choosing. Some of these models use bi-directional Long Short-Term Memory with Conditional Random Field (BiLSTM-CRF) [15], and other models have extra character-level CNN [16, 17] or character-level LSTM [18, 19] to capture character features of entities.

Recently, multi-task learning (MTL) [20] has been adopted successfully to applications of biomedical literature mining, such as drug discovery [21], entity linking [22]. The multi-task model trains several datasets at the same time, and transfers domain information between datasets. By sharing representations between the main task and auxiliary task, the multi-task model improves the performance on the main task. For MTL BioNER models, the number of successful examples is growing. Crichton et al. [23] uses convolution layer as the shared part and fully connected layer as task-specific part. Wang et al. [19] experiments shared character Bi-LSTM, shared word Bi-LSTM, and shared both. Although the multi-task model can optimize the performance of the main dataset, using different combinations of training datasets may have discrepancy performances. Some other models use special methods to improve performance, such as adversarial loss [24], label-aware MMD [25], Learn What to Share Structure [26].

In this paper, we compare some different multi-task models and propose our new model with the cross-sharing structure for BioNER. No hand-crafted feature is required in our model. The proposed model is based on the BiLSTM-CNN-CRF model [16] which is a single-task neural network model. In our model, shared Bi-LSTM unit is used to learn the shared features, and private Bi-LSTM units are for the task-specific features. Besides, a cross-sharing structure helps to share information between private units. We compare the proposed model with other multi-task models [19, 24] on four main datasets of different domains. We also discover the influence of dataset pairs and dataset size to the performance of our proposed model. Results demonstrate that the proposed model achieves good results. Our method provides a novel structure of multi-task sharing in BioNER task and improves the overall performance on BioNER datasets.

Preliminaries

In this section, some basic concepts related to our multi-task neural network are introduced.

Bi-directional long short-Term memory (Bi-LSTM)

Long Short-Term Memory (LSTM) [27] is a special edition of Recurrent neural network (RNN), and LSTM avoids the gradient vanishing or exploding problems appearing in RNN. A normal LSTM cell contains a input gate, a output gate and a forget gate, and there are connections between these gates. We denote X={x1,x2,...,xT} as the series input of LSTM, where T is the sequence length of input vector. The output of LSTM is a sequence of vector H={h1,h2,...,hT}. The LSTM cell calculates ht via the following calculation:

$$\begin{array}{*{20}l} \boldsymbol{f}_{t} &= \sigma (\boldsymbol{W}_{f} [\boldsymbol{h}_{t-1}, \boldsymbol{x}_{t}] + \boldsymbol{b}_{f}) \end{array} $$
(1)
$$\begin{array}{*{20}l} \boldsymbol{i}_{t} &= \sigma (\boldsymbol{W}_{i} [\boldsymbol{h}_{t-1}, \boldsymbol{x}_{t}] + \boldsymbol{b}_{i}) \end{array} $$
(2)
$$\begin{array}{*{20}l} \tilde{\boldsymbol{C}_{t}} &= tanh (\boldsymbol{W}_{C} [\boldsymbol{h}_{t-1}, \boldsymbol{x}_{t}] + \boldsymbol{b}_{C}) \end{array} $$
(3)
$$\begin{array}{*{20}l} \boldsymbol{C}_{t} &= \boldsymbol{f}_{t} \odot \boldsymbol{C}_{t-1} + \boldsymbol{i}_{t} \odot \tilde{\boldsymbol{C}_{t}} \end{array} $$
(4)
$$\begin{array}{*{20}l} \boldsymbol{o}_{t} &= \sigma (\boldsymbol{W}_{o} [\boldsymbol{h}_{t-1}, \boldsymbol{x}_{t}] + \boldsymbol{b}_{o}) \end{array} $$
(5)
$$\begin{array}{*{20}l} \boldsymbol{h}_{t} &= \boldsymbol{o}_{t} \odot tanh (\boldsymbol{C}_{t}) \end{array} $$
(6)

In these equations, denotes element-wise multiplication. σ and tanh are element-wise sigmoid function and tanh function, respectively. ft,it,ot are the forget gate, the input gate, and the output gate, respectively. \( \tilde {\boldsymbol {C}_{t}} \) indicates some information from current input applied to cell state. ht calculates the cell output by the input and current cell state. Wj,bj(j=f,i,C,o) are the trainable parameters. The LSTM cell is designed to avoid the long-term dependency problem, and it is capable of capturing information for long periods.

Bi-LSTM is the two-direction version of LSTM. For original LSTM, the cells take input in one direction, so ht will capture some information only from previous LSTM cells. In order to capture the information from the following cells, another set of LSTM cells is used in Bi-LSTM. As shown in Figure 1, the bi-directional long short-term memory (Bi-LSTM) model contains two directions of LSTM network, original direction and reversed direction.

Fig. 1
figure 1

Bi-LSTM Structure. The figure displays a part of Bi-LSTM network. Input vectors are fed to two directions of LSTM, and the output of two directions of LSTM is concatenated as the whole output

$$ \overrightarrow{\textbf{h}}_{t} = LSTM\left(\overrightarrow{\textbf{h}}_{t - 1}, {\textbf{x}}_{t}\right) $$
(7)
$$ \overleftarrow{\textbf{h}}_{t} = LSTM\left(\overleftarrow{\textbf{h}}_{t + 1}, {\textbf{x}}_{t}\right) $$
(8)
$$ \textbf{o}_{t} = \overrightarrow{\textbf{h}}_{t} \oplus \overleftarrow{\textbf{h}}_{t} $$
(9)

In these equations, \( \overrightarrow {\boldsymbol {h}}_{t} \) and \( \overleftarrow {\boldsymbol {h}}_{t} \) are the cells output of two directions. denotes vector concatenation. The vectors, \( \overrightarrow {\boldsymbol {h}}_{t} \) and \( \overleftarrow {\boldsymbol {h}}_{t} \), are concatenated as the final output. In this way, ot keeps the information from previous and following LSTM cells.

Conditional random field (CRF)

Conditional Random Field (CRF) [28] is a conditional probability distribution model and widely used in sequence labeling tasks to generate new tag based on recent tags. When a set of random variables are given as input, CRF outputs another set of random variables according to some rules. For example, in biomedical NER task with IOB annotation, the tag after B-Gene can be I-Gene rather than I-Disease. If the previous tag is B-Gene, CRF would output I-Disease in a low probability to avoid the error of mixing different types of tags. CRF has been adopted in many state-of-art models to help to generate meaningful and legal annotations.

Let the input of CRF is vector Z=(z1,z2,...,zn), and the generated output sequence is \( \boldsymbol {\hat {Y}} = (\hat {y}_{1}, \hat {y}_{2},..., \hat {y}_{n}) \). For BioNER task, the input zi can be a feature vector representing the ith word. CRF model describes the probability of generating the whole label sequence based on Z, shown as below:

$$ p(\boldsymbol{\hat{Y}}|\boldsymbol{Z}; \boldsymbol{W}, \boldsymbol{b})=\frac{\prod_{i=1}^{n} f_{i}(\hat{y}_{i-1}, \hat{y}_{i}, \boldsymbol{Z})}{\sum_{y' \in\phi(\boldsymbol{Z})} \prod_{i=1}^{n} f_{i}(y'_{i-1}, y'_{i}, \boldsymbol{Z})} $$
(10)

In this equation, ϕ(Z) represents all of the possible label sequences for Z. The function \( f_{i}(y_{j}, y_{k}, \boldsymbol {Z}) = exp(\boldsymbol {W}_{y_{j},y_{k}}\boldsymbol {z_{i}} + \boldsymbol {b}_{y_{j},y_{k}}) \phantom {\dot {i}\!}\), where the weight \(\phantom {\dot {i}\!} \boldsymbol {W}_{y_{j},y_{k}} \) and the bias \( \boldsymbol {b}_{y_{j},y_{k}} \phantom {\dot {i}\!}\) are the trainable parameters corresponding to the pair of labels (yj,yk).

In the training procedure, we use the negative log-likelihood function to calculate the loss function J and find the optimal sequence y by minimum the loss function. The Viterbi algorithm is used to calculate the loss and the optimal sequence.

$$\begin{array}{*{20}l} J(\boldsymbol{W}, \boldsymbol{b}) &= -\sum_{i} \log(p(\boldsymbol{\hat{Y}}|\boldsymbol{Z}; \boldsymbol{W}, \boldsymbol{b})) \end{array} $$
(11)
$$\begin{array}{*{20}l} y^{*} &= \underset{y \in \phi(\boldsymbol{Z})}{\arg\min} \ \ J(\boldsymbol{W}, \boldsymbol{b}) \end{array} $$
(12)

Methods

In this section, we introduce our baseline single-task model and some multi-task models for BioNER tasks.

Baseline single-task model (STM)

We choose the model from Ma and Hovy [16] as our baseline single-task model. Unlike the vanilla BiLSTM-CRF model, this model uses an extra CNN layer to capture character-level features. All the multi-task models in the paper are implemented based on this single-task model; thus, we choose it as our baseline model. The model structure is shown in Fig. 2.

Fig. 2
figure 2

Single-task Model (STM). The input is a sentence from the BioNER dataset. The dotted rectangles represent words in a sentence, and the solid rectangles represent Bi-LSTM cells. The circles represent CNN units, and the double circles represent CRF units. The tags in the double circles, e.g., “O”, “B-GENE”, are the output of the CRF layer

For simplicity, wt denotes word embedding of word t and the ct denotes character embeddings of word t. The shape of ct is dc by lc, where dc is the dimension of character embedding and lc is the count of characters in the word.

In the embedding layer, the character representation rt is calculated based on character embedding ct by CNN to extract morphological information. The CNN scheme we use is the same as Ma and Hovy [16]. The convolution has the filter size of dc by lf and padding length of lf−1, where lf is a hyperparameter. After the convolution calculation, the output is a new vector of shape dc by (lc+lf−1). Then max pooling is used to produce a vector of size dc as the final char representation rt. A dropout layer is adopted at the input of CNN. Finally, word embedding wt and character representation rt are concatenated as xt.

After the embedding layer, resulting sequence of embeddings X={x1,x2,...,xn} are fed into Bi-LSTM layer to get O={o1,o2,...,on}. Two dropout layers are applied at the input and output of the Bi-LSTM layer. The Bi-LSTM layer is used to extract information from the word representation xt.

The top layer of the model is the CRF layer. This layer takes output vectors O to predict label sequences. As shown in Fig. 2, the word “28S” and the word “rRNA” are predicted as B-Gene and I-Gene, respectively, which suggests that the model recognizes the entity “28S rRNA”.

Fully-shared multi-task model (FS-MTM)

Our fully-shared multi-task model is based on MTM-CW from Crichton et al. [23]. All the multi-task models in this paper are designed for two datasets. If modifications applied, these models are suitable for three or more datasets. The embedding layer, Bi-LSTM layer and CRF layer in the multi-task models are the same as those in the baseline single-task model.

In the fully-shared multi-task model, we use an embedding layer and a Bi-LSTM layer as shared parts, and two CRF layers for two datasets, as shown in Fig. 3. When training and testing, word embeddings and character embeddings are first fed to the embedding layer, and then the Bi-LSTM layer takes the output of embedding layer. In the end, the output of Bi-LSTM is fed to one of the CRF layers. If source data is from dataset 1, CRF layer for dataset 1 is activated with another CRF layer ignored, and vice versa. In this model, Bi-LSTM captures all the features of dataset 1 and 2, and CRF layer produces different tags according to the input dataset.

Fig. 3
figure 3

Fully-shared Multi-task Model (FS-MTM). The embedding layer and the Bi-LSTM layer are shared by two datasets, and two CRF layer are used for two datasets

Shared-private multi-task model (SP-MTM)

Our shared-private multi-task model is based on SP-MTL from Liu et al. [24]. As shown in Fig. 4, there are two private Bi-LSTMs for two tasks and one shared Bi-LSTM. Word embeddings and character embeddings are first fed to the embedding layer. Then the output of the embedding layer is replicated and fed into shared Bi-LSTM and corresponding private Bi-LSTM, according to the source dataset. Finally, the output of shared and private Bi-LSTMs are concatenated and fed into corresponding CRF layer. In this model, shared Bi-LSTM and private Bi-LSTM captures shared and task-independent features, respectively. CRF layer produces different tags based on task-related feature representations.

Fig. 4
figure 4

Shared-private Multi-task Model (SP-MTM). The embedding layer and shared Bi-LSTM are shared by two datasets. Two CRF layer and two private Bi-LSTMs are used for two datasets

Adversarial multi-task model (ADV-MTM)

As shown in Fig. 5, our adversarial multi-task model is based on the adversarial shared-private model from Liu et al. [24]. The basic network structure of the adversarial multi-task model is the same as the shared-private multi-task model, but the calculation of loss is different.

Fig. 5
figure 5

Adversarial Multi-task Model (ADV-MTM). The embedding layer and shared Bi-LSTM are shared by two datasets. Two CRF layer and two private Bi-LSTMs are used for two datasets. Three kinds of losses are marked on the figure

We deem the current data as d, and source datasets are \( \mathcal {D}_{1}, \mathcal {D}_{2} \). Ltask is the task loss calculated by CRF layer. shown as Eq. 13.

$$ L_{task} = \left\{ \begin{array}{lr} L_{task1}, & d \in \mathcal{D}_{1}; \\ L_{task2}, & d \in \mathcal{D}_{2}. \end{array} \right. $$
(13)

Ldiff is calculated by the output of shared Bi-LSTM and private Bi-LSTM. Ldiff describes the similarity of these two output vectors. Minimizing Ldiff encourages shared and private Bi-LSTM to extract different features of input. Ldiff is calculated as Eq. 14:

$$ L_{diff} = \sum_{k=1, 2}||{\boldsymbol{S}}^{\top}\boldsymbol{P}^{k}||_{F}^{2} $$
(14)

where S is the output of shared Bi-LSTM and Pk is the output of private Bi-LSTM of dataset k. \( ||\cdot ||_{F}^{2} \) is the squared Frobenius norm.

Ladv is task adversarial loss. The shared Bi-LSTM can be regarded as generative model G which produce vector to hide the information of source dataset, and we use a discriminative model D to identify the source dataset against generative model G. Discriminative model D is shown as Eq. 15:

$$ D(\boldsymbol{s}_{T}^{k}, \theta_{D}) = softmax\left(\boldsymbol{W}\boldsymbol{s}_{T}^{k} + \boldsymbol{b}\right) $$
(15)

where \( \boldsymbol {s}_{T}^{k} \) is the output of shared Bi-LSTM of dataset k at time T. W and b are trainable parameters. And the adversarial loss function is:

$$ L_{adv} = -\max_{\theta_{G}}\left(\min_{\theta_{D}}\left(\sum_{k=1}^{K} \boldsymbol{d}_{i}^{k} \log\left[D\left(E\left(\boldsymbol{x}^{k}\right)\right)\right]\right)\right) $$
(16)

Discriminative model D is able to recognize source dataset by task-dependent features, and generative model G tends to keep common features to confuse discriminative model D; therefore, minimizing Ladv encourages shared Bi-LSTM to keep more shared features of two datasets.

The final loss is the weighted sum of these three kinds of losses.

$$ L = L_{task} + \alpha L_{adv} + \beta L_{diff} $$
(17)

where α and β are hyperparameters.

Grid search can be used to find the optimized hyperparameters α and β. By using the gradient reversal layer [29] before the discriminative model, the whole network can be trained with backpropagation.

Multi-task model with cross-sharing structure (CS-MTM)

In this section, we introduce our multi-task model with cross-sharing structure. This model captures features from both datasets and takes advantage of all the feature representations.

As shown in Fig. 6, the word embeddings and character embeddings of the input sentence are first fed to the embedding layer. The structure of the embedding layer is the same as that in the baseline single-task model. The embedding layer captures the information in word embeddings and character embeddings. The output of the embedding layer is the word representations, which can be used in the Bi-LSTM layers.

Fig. 6
figure 6

Cross-sharing Multi-task Model (CS-MTM). The embedding layer and shared Bi-LSTM are shared by two datasets. Gated interaction unit is used to adjust the output of private Bi-LSTMs. P1,P2: Output of private Bi-LSTMs. S: Output of the shared Bi-LSTM. G1,G2: Output of the gated interaction unit

After the embedding layer, the word representations are replicated as the input of shared Bi-LSTM and both private Bi-LSTMs. P1,P2 denote the output of two private Bi-LSTMs. S denotes the output of shared Bi-LSTM. Intuitively, the private Bi-LSTMs are used to capture task-independent features; thus, P1,P2 are the feature representations of dataset 1 and 2. The shared Bi-LSTM captures the common features from both datasets and S is the representation of common features.

In our previous SP-MTM and ADV-MTM, either P1 or P2 is calculated depending on source dataset. In this way, only feature representation of source dataset is calculated, but the other feature representation which may still be useful is not calculated. In multi-task learning, using information from other datasets to improve the performance of origin dataset is the main idea, so both P1 and P2 are used in this model.

The gated interaction unit then takes P1,P2 as input and produces a mixed feature representation. G1,G2 denote the output of gated interaction unit for two datasets. Eq. 18 and (19) show how gated interaction unit works.

$$\begin{array}{*{20}l} \boldsymbol{G}_{1} &= \boldsymbol{P}_{1} \odot \sigma(\boldsymbol{W}_{2 \rightarrow 1}\boldsymbol{P}_{2} + \boldsymbol{b}_{2 \rightarrow 1}) \end{array} $$
(18)
$$\begin{array}{*{20}l} \boldsymbol{G}_{2} &= \boldsymbol{P}_{2} \odot \sigma(\boldsymbol{W}_{1 \rightarrow 2}\boldsymbol{P}_{1} + \boldsymbol{b}_{1 \rightarrow 2}) \end{array} $$
(19)

where is element-wise multiplication, σ is a sigmoidal function, and W1→2,W2→1,b1→2,b2→1 are trainable parameters.

We deem the current data as d, and source datasets are \( \mathcal {D}_{1}, \mathcal {D}_{2} \). The final output of gated interaction unit G is determined by the source dataset, shown as Eq. 20.

$$ \boldsymbol{G}=\left\{ \begin{array}{lr} \boldsymbol{G}_{1}, & d \in \mathcal{D}_{1}; \\ \boldsymbol{G}_{2}, & d \in \mathcal{D}_{2}. \end{array} \right. $$
(20)

In the gated interaction unit, two private feature representations P1,P2 share feature information with each other. When training, four trainable parameters are adjusting to learning what to share between two representations. For dataset 1, P2 contains the information of features from dataset 2, and these features are task-independent and cannot be used directly to improve the performance of dataset 1; otherwise, these features should be captured by shared Bi-LSTM. The operation in gated interaction unit provides an indirect way to make use of the information in P2. In this way, both feature representations P1,P2 are used to produce a new mixed feature representation.

Before the CRF layer, output vectors of gated interaction unit and shared Bi-LSTM are concatenated, shown as Eq. 21.

$$ \boldsymbol{V} = \boldsymbol{G} \oplus \boldsymbol{S} $$
(21)

In this way, information of shared feature representation and private feature representation is combined and fed to the CRF layer. CRF layer produces predicted tags based on V.

Experiment settings

In this section, we introduce our datasets, evaluation metrics, and training details.

Datasets

We conduct experiments on several BioNER datasets from Crichton et al. [23]. The detailed information about the datasets used in our experiments is listed in Table 1. We use datasets with IOB format. These datasets are available to the open, and you can access https://github.com/cambridgeltl/MTL-Bioinformatics-2016 to get these datasets.

Table 1 Biomedical NER datasets used in the experiments

As these datasets use various BioNER tags to mark out entities, we divide them into six categories: Cell, Chemical, Disease, Gene, Protein and Species. For the entity types column in Table 1, BioNER tags are counted according to these six categories. In different datasets, BioNER tags belonging to the same category may vary. For example, in Gene categories, B-GENE/I-GENE tags are used in BC2GM dataset, while B-DNA/I-DNA are in JNLPBA dataset. In our experiments, tags are kept as they are rather than changed to be the same.

In our multi-task models, two datasets are used in the training procedure. We focus on one dataset and try to optimize the performance on it. This dataset is called the main dataset, and the other is called the auxiliary dataset. By observation, we find that some datasets contain entities from just one category, while some others from multiple categories. In order to diminish the influence between different entity categories, we prefer datasets which contain entities from one category to be main datasets. In our experiments, BC2GM, Ex-PTM, NCBI-disease, and Linnaeus are chosen as main datasets, and auxiliary datasets are picked from JNLPBA, BioNLP09, BioNLP11ID, BioNLP13PC, and BC5CDR. The performance of the main datasets is recorded in experimental results.

Evaluation metrics

We use the training set and the development set to train the model, and report the performance on the test set. We deem each predicted tag is correct only if it is the same as the ground-truth tag. We calculate macro-averaged precision, recall, F1 scores of main dataset, and these scores are recorded as final dataset scores.

Training details

Word embeddings We use pre-trained word vectors of GloVe model, and the pre-trained corpus is Wikipedia 2014 + Gigaword 5 (6B tokens, 400K vocab). The dimension of word vectors is 100.

Character embeddings The dimension of character embeddings dc is 30. Number of filters in CNN is 30. lf in the CNN is set to 3.

Bi-LSTM layers Bi-LSTM in our model uses the same hyperparameters, including Bi-LSTM in baseline single-task model, and shared/private Bi-LSTM in multi-task models. We set the dimension of hidden vectors to 256. For Bi-LSTM layers in all of our models, we use a linear unit to reshape hidden vectors to 128-dimensional vector as output. The dropout rate of all the dropout layers is 0.5.

CRF layers We use Linear-chain CRF to train and test. The Viterbi algorithm is used in the training procedure.

Training settings Our training procedure contains 80 epochs. Parameter optimization is performed with RMSprop. The decay rate of RMSProp is set to 0.95, and momentum is set to 0. Batch size is set to 16. Learning rate is 0.001 at initial, and decay at the end of every epoch at the rate of 3%. Besides, We use gradient clipping to limit max L2 norm of the gradients to 5.0 in order to avoid gradient exploding.

MTM training When performing multi-task training, batches of data from 2 datasets train in turns. To be specific, a batch of data from dataset 1 is used to train, then a batch of data from dataset 2 is used to train, this procedure is regarded as a turn. In one turn, two batches of data are randomly picked from their source datasets. In an epoch, the number of turns is set to the number of batches in the main dataset. In this case, we ensure the main dataset to be fully trained.

Grid search For the adversarial multi-task model, grid search is used to find the hyperparameters α and β. We try out α from {0, 0.1, 0.01}, and β from {0, 0.01, 0.001}. An extra gradient reverse layer is applied before the task discriminator unit in order to train the model with gradient descent.

Results

In this section, we display and analyze the experiment results, and compare our proposed model with related ones.

Performance comparison

We compare the baseline single-task model (STM) and other multi-task models (MTM). The results are shown in Table 2. It shows the performance (precision, recall, F1) of different models on four target datasets. The bold number in one row indicates the best F1 score for the dataset.

Table 2 Model Performance Comparison

FS-MTM achieves better performance than STM on BC2GM and Ex-PTM datasets but degrades on other two datasets. FS-MTM uses the most basic multi-task structure, and the only shared Bi-LSTM may not separate task-specific features for each task.

SP-MTM improves the performance comparing to FS-MTM and STM, also achieves higher F1 score than baseline STM on all of four main datasets. Intuitively, the private Bi-LSTMs are added and capable of capturing task-specific features.

We observe that both the ADV-MTM and CS-MTM improve the performance of STM, and especially CS-MTM achieves higher F1 score than baseline STM on all of four datasets. On BC2GM dataset, improvements of ADV-MTM are marginal compared with STM. Besides, CS-MTM outperforms ADV-MTM in F1 score on BC2GM, Ex-PTM, and NCBI-disease datasets. Comparing the structure of ADV-MTM and CS-MTM to SP-MTM, it indicates that the adversarial loss calculation and cross-sharing structure could help to improve the performance.

According to the precision and recall score of datasets, CS-MTM tends to produce a higher recall score, and ADV-MTM tends to improve the precision score. Intuitively, minimizing the adversarial loss in ADV-MTM helps to separate shared features and task-specific features and reduce the number of false positives. Unlike ADV-MTM, gated interaction unit in CS-MTM makes use of both feature representations, resulting in less number of false negatives.

When training, we find that the performance of ADV-MTM is not very stable, and the adversarial model uses more epochs to converge. This model has limited performance improvement comparing to SP-MTM and exposes the weakness of GAN.

We list the trainable parameter number of each model in Table 3. In the table, the parameter numbers of STM and FS-MTM are close, and SP-MTM, ADV-MTM, CS-MTM have more parameters. We can conclude that the gated interaction unit in CS-MTM has only a few parameters but improves the overall performance. It suggests that our performance improvement is not just based on the increase in the huge amount of parameters.

Table 3 Parameter numbers of all models

Performance with different auxiliary datasets

Different dataset pairs could produce different results in multi-task learning. We try out all the combinations of one main dataset and one auxiliary dataset. The results are shown in Table 4. The numbers in the table are the F1 scores of dataset pairs. BC2GM, Ex-PTM, NCBI-disease, and Linnaeus are the main dataset that we focus on. The bold number in one row indicates the best F1 score for the dataset. The / indicates the positive/negative improvement comparing to STM.

Table 4 Performance with different auxiliary datasets

From experiment results, JNLPBA is the best partner for BC2GM, and BC5CDR, BioNLP09 are helpful to BC2GM. All these five auxiliary datasets are helpful to Ex-PTM, but the best partner of Ex-PTM is BioNLP09. As for NCBI-disease, BioNLP09 is the best partner, and JNLPBA is also helpful. Auxiliary datasets except JNLPBA and BioNLP13PC are helpful to Linnaeus, and BC5CDR improves its performance significantly.

In auxiliary datasets, JNLPBA is of the biggest size, and BioNLP13PC is the smallest. Using JNLPBA as the auxiliary dataset still degrades on Linnaeus dataset, while using BioNLP13PC as the auxiliary dataset in this experiment improves the performance on Ex-PTM. For these five auxiliary datasets, we cannot observe a tendency of performance increasing or decreasing with the size of dataset changing. This phenomenon indicates that the size of the dataset is not the major factor of performance. If auxiliary dataset lacks beneficial information for the main dataset, the performance of multi-task model would be unfavorable.

BC2GM contains gene tags, and its best partner JNLPBA also contains gene tags. The situation is similar for Ex-PTM and BioNLP09. It could indicate that the dataset pair could work if the auxiliary dataset contains the categories of tags that main dataset also has. But for Linnaeus and its best partner BC5CDR, although they share no same categories of tags, BC5CDR can still provide biomedical information of other categories which is helpful to Linnaeus.

In conclusion, there is no simple rule to find the best partner, the most accurate way is to try out all the combinations.

Performance with different entity types in auxiliary datasets

In our five auxiliary datasets, some of them contain multiple categories of tags. In order to discover which category of tags is the major factor of performance, we use sub-datasets to perform the experiments. The BC5CDR, BioNLP11ID, BioNLP13PC datasets provide sub-datasets that contain the single category of tags. In this experiments, We choose our four main datasets and BioNLP11ID-chem (Chemical), BioNLP11ID-ggp (Protein), BioNLP11ID-species (Species) as auxiliary datasets. This experiment aims to check which category of tags is the most important for main datasets in CS-MTM. The results are shown in Table 5. The / indicates the positive/negative improvement comparing to STM.

Table 5 Performance with different entity types in BioNLP11ID

Ex-PTM dataset contains tags of protein category, and its best partner BioNLP11ID-ggp also contains that category of tags. Besides, as for Linnaeus and BioNLP11ID-species, these two datasets are the best pair and both contain tags of species category. It indicates that protein tags and species tags are the major factors for Ex-PTM and Linnaeus datasets, respectively, when BioNLP11ID as the auxiliary dataset. As for other tags, chemical and species tags in the BioNLP11ID dataset are hardly helpful to Ex-PTM dataset, while chemical and protein tags would make the performance of Linnaeus ever worse.

BC2GM and NCBI-disease datasets contain no tags of chemical, protein and species categories. In experiment results, we could observe that chemical and protein tags in BioNLP11ID dataset are helpful to BC2GM while species tags are harmful. For NCBI-disease dataset, all categories of tags make performance worse.

When a dataset contains multiple categories of tags, mutual influences may exist between them. For BC2GM datasets, chemical tags improve performance and species tags reduce performance, but the result of all tags is still positive. It indicates that categories of tags with the opposite effect would neutralize each other. For Ex-PTM dataset, all the categories of tags improve performance, and the performance of all tags is better than a single category of tags. Similarly, for NCBI-disease dataset, the merged result is worse. It demonstrates that the categories of tags with the same effect could cooperate and accumulate their effects. Exceptionally, for Linnaeus dataset, categories of tags are all negative, but the result of all tags is positive. We don’t have an analysis to explain this phenomenon if just base on the current experiment result, but it suggests that the influence of different categories of tags is not a simple linear calculation.

Impact of dataset size

In this part, we discover the performance of CS-MTM on smaller datasets. Using reduced-size main datasets, we record the performance (precision, recall, F1) of different situations. The results of CS-MTM in this experiment are produced using the best pairs in “Performance with different auxiliary datasets” section. The reduced-size datasets are produced by removing sentences in training sets randomly, and the development sets and test sets are not modified. To compare with, we also use the reduced-size dataset on baseline single-task model. The results are shown in Table 6. The better F1 scores for each training set size are bold.

Table 6 Impact of dataset size

For STM and CS-MTM, the F1 score decreases when the size of training data is limited. When the training set is reduced and the test set is kept, the missing of information in removed sentences makes the model produce worse results. In CS-MTM, the missing information could be found in auxiliary datasets, so CS-MTM could improve the performance back if a suitable auxiliary dataset is chosen.

For 50%-size and 25%-size datasets, CS-MTM outperforms STM on F1 score by providing a higher recall score. But for 10%-size datasets, CS-MTM outperforms STM on BC2GM and NCBI-disease datasets and degrades on Ex-PTM and Linnaeus datasets. In this case, our CS-MTM may not learn missing information from auxiliary dataset well.

Performance with different word embeddings

In this part, we discover the performance of STM and CS-MTM by using different pre-trained word embeddings. In our previous experiments, we just use the pre-trained GloVe to produce our word embeddings. Our CS-MTM model may have better performance when using other word embeddings. In this experiment, we obtain the performance with several different pre-trained Word2Vec and compare them with the performance with the original pre-trained GloVe. The results are shown in Table 7. The best F1 scores for the model on each dataset are bold.

Table 7 Performance with different word embeddings

Four pre-trained Word2Vec word embeddings are used in this experiment. One trains with PMC corpus, one trains with PubMed corpus, one trains with PMC + PubMed corpora, one trains with PMC + PubMed + Wikipedia corpora. These pre-trained Word2Vec word embeddings are available at http://bio.nlplab.org/. They report that Word2Vec was run using the skip-gram model with a window size of 5, hierarchical softmax training, and a frequent word subsampling threshold of 0.001 to create 200-dimensional vectors.

For STM, we have the best performance on BC2GM dataset when choosing PMC + PubMed + Wikipedia word embedding, and the best performance on the other three datasets is achieved by our original GloVe word embedding. For CS-MTM, PMC + PubMed on BC2GM and other three datasets on GloVe word embedding can produce the best performance. This phenomenon shows that different word embeddings can produce discrepant performance.

Our GloVe word embedding achieves good performance on three datasets, but the coverage of Glove might be relatively small because it is not trained with the biomedical corpus. An important reason is that CNN in the embedding layer builds character embeddings to compensate for the missing of words. Besides, according to the overall performance, GloVe embeddings work better with our models than Word2Vec embeddings. But on certain datasets, such as BC2GM, character embeddings may not work well, and using word embedding which trains with specialized corpus can improve the performance.

Case study

In this part, we use some examples from datasets to illustrate the effect of the multi-task model. The examples are shown in Table 8.

Table 8 Case Study: Bold text: ground-truth entity; Underlined text: model prediction

Case 1 and 2 are picked from the test set of Ex-PTM. The main dataset, Ex-PTM, and the auxiliary dataset, BioNLP09, only have entity tags of protein category. In case 1, STM cannot recognize the entity Nef but CS-MTM can find it out, because the training data of auxiliary dataset directly provides entity information about Nef protein. In case 2, STM recognizes Vitamin K as a protein entity, which is incorrect. For the CS-MTM, in the training data of auxiliary dataset, there is no information about Vitamin K, but other Vitamins, such as Vitamin C and Vitamin E, appear in the dataset. The character embedding in the model can capture the morphological information; therefore, the multi-task model can recognize these Vitamins as non-protein entities.

Case 3 is picked from the test set of Linnaeus. Linnaeus contains entity tags of species category, but the auxiliary dataset, BC5CDR, have no species entity tags. In case 3, STM recognizes slept as a species entity. Because our model use no pre-defined feature, such as Part-of-Speech feature, STM may not learn that slept is not an entity if there are few appearances of this word. For the CS-MTM, it can learn from auxiliary training data which exists the information of sleep; therefore, CS-MTM can recognize it as a non-species entity.

Discussion

In this part, we compare our models with other BioNER models as well as the state-of-the-art models.

For the multi-task model from Crichton et al. [23], they experiment with many BioNER datasets. They report their best model achieves the F1 of 73.17% on BC2GM, 74.90% on Ex-PTM, 80.37% on NCBI-disease, and 84.04% on Linnaeus. Our model has better performance on BC2GM and NCBI-disease datasets, because both word embedding and character embedding are used as input in our model, while only word embedding is used in their model. In Crichton’s work, many more combinations of datasets are tried in the experiment, so this could be the reason why they have better performance on Ex-PTM and Linnaeus.

For the multi-task model from Wang et al. [19], they achieve the F1 of 83.14% on BC2GM and 86.37% on NCBI-disease. Our model outperforms their model on these two datasets, because we use shared and private Bi-LSTMs to capture different features, as well as the gated interaction unit to make use of features from the auxiliary dataset.

For the BioBERT model from Lee et al. [30], they report their best model achieves the F1 of 84.40% on BC2GM, 89.36% on NCBI-disease, and 89.81% on Linnaeus. Their model outperforms ours because BioBERT has much more trainable parameters than ours. In BioBERT’s paper, the authors don’t report the number of parameters, but BioBERT should be similar to the original BERT which has more than 100M parameters to train.

For the CollaboNet model from Yoon et al. [31], they achieve the F1 of 78.56% on BC2GM and 86.36% on NCBI-disease. This model uses a special structure to achieve good performance, but our model uses multi-task learning to achieve better performance on BC2GM dataset.

As for state-of-the-art models, BioCreative II Gene Mention Tagging System [10] achieves the F1 of 87.21% on BC2GM dataset, MO-MTM from Crichton et al. [23] achieves the F1 of 74.90% on Ex-PTM dataset, BioBERT [30] achieves the F1 of 89.36% on NCBI-disease dataset, and the original LINNAEUS system [32] achieves the F1 of 95.68% on Linnaeus dataset. Although BioCreative II and LINNAEUS system have the best performance on certain datasets, they rely heavily on hand-craft features which are not used in our model. Besides, these systems can pre-process the input data or have some special process using field knowledge, which benefits the performance.

Conclusion

In this paper, we propose a new multi-task learning framework for BioNER. We also implement some other multi-task models and compare our new model with them. Our proposed model achieves better performance, even if the size of the training data is smaller. Detailed analysis about best partners of datasets and influence between entity categories can provide guidance of choosing proper dataset pairs for multi-task training. Furthermore, our analysis suggests that the cross-sharing structure in our model is a key point to improve performance in the way of cross-dataset feature sharing.

Limitations to the work include that it is difficult to predict whether one dataset can help another before running the model. Another limitation is that the current implementation of the model may not produce promising results for all datasets, in our experiment we find the performance of the proposed model on Linnaeus dataset worse than the ADV-MTM.

There are several further directions with our cross-sharing multi-task model. First, training more datasets at the same time could provide more cross-dataset information and obtain better performance. Besides, we can adjust our cross-sharing structure to improve the performance on certain datasets or combine the current multi-task model with the newly proposed structure, such as BioBERT. Finally, our work may have entity type conflict problem, we could use an entity type unifier to recognize by source datasets in order to get the performance improvement.

Availability of data and materials

BioNER datasets are available at https://github.com/cambridgeltl/MTL-Bioinformatics-2016. Our implement of cross-sharing multi-task model is available at https://github.com/JogleLew/bioner-cross-sharing.

Abbreviations

ADV-MTM:

Adversarial multi-task model

Bi-LSTM:

Bi-directional long short-term memory

BiLSTM-CRF:

Bi-directional long short-term memory with conditional random field

BioNER:

Biomedical named entity recognition

CRF:

Conditional random field

CS-MTM:

Multi-task model with cross-sharing structure

FS-MTM:

Fully-shared multi-task model

LSTM:

Long short-term memory

MTL:

Multi-task learning

RNN:

Recurrent neural network

SP-MTM:

Shared-private multi-task model

STM:

Single-task model

References

  1. Pan X, Fan Y-X, Jia J, Shen H-B. Identifying rna-binding proteins using multi-label deep learning. Sci China Inf Sci. 2019; 62(1):19103.

    Article  Google Scholar 

  2. HE Y, LIU S, QIAN L, ZHOU G. Disease name recognition based on syntactic and semantic features. Sci Sin Inf. 2018; 48(11):1546–57.

    Article  Google Scholar 

  3. Wei C-H, Peng Y, Leaman R, Davis AP, Mattingly CJ, Li J, Wiegers TC, Lu Z. Assessing the state of the art in biomedical relation extraction: overview of the biocreative v chemical-disease relation (cdr) task. Database. 2016; 2016:8.

    Article  Google Scholar 

  4. Donaldson I, Martin J, De Bruijn B, Wolting C, Lay V, Tuekam B, Zhang S, Baskin B, Bader GD, Michalickova K, et al. Prebind and textomy–mining the biomedical literature for protein-protein interactions using a support vector machine. BMC Bioinformatics. 2003; 4(1):11.

    Article  Google Scholar 

  5. Rajpal D, Qu X, Freudenberg J, Kumar V. Mining emerging biomedical literature for understanding disease associations in drug discovery. Methods Mol Biol (Clifton, NJ). 2014; 1159:171–206.

    Article  CAS  Google Scholar 

  6. Montojo J, Zuberi K, Rodriguez H, Bader GD, Morris Q. Genemania: Fast gene network construction and function prediction for cytoscape. F1000Research. 2014; 3:153.

    Article  Google Scholar 

  7. Wang J, Peng X, Peng W, Wu F-X. Dynamic protein interaction network construction and applications. Proteomics. 2014; 14(4-5):338–52.

    Article  CAS  Google Scholar 

  8. Ji M, He Q, Han J, Spangler S. Mining strong relevance between heterogeneous entities from unstructured biomedical data. Data Min Knowl Disc. 2015; 29(4):976–98.

    Article  Google Scholar 

  9. Fukuda K-i, Tsunoda T, Tamura A, Takagi T, et al. Toward information extraction: identifying protein names from biological papers. In: Proceedings of the 3rd Pacific Symposium on Biocomputing. Hawaii: PSB: 1998. p. 707–718.

    Google Scholar 

  10. Ando RK. Biocreative ii gene mention tagging system at ibm watson. In: Proceedings of the Second BioCreative Challenge Evaluation Workshop. Madrid: Centro Nacional de Investigaciones Oncologicas (CNIO). 2007;23:101–3.

  11. Campos D, Matos S, Oliveira JL. Biomedical named entity recognition: a survey of machine-learning tools. In: Theory and Applications for Advanced Text Mining. London: IntechOpen: 2012. p. 175–95.

    Google Scholar 

  12. Leaman R, Wei C-H, Lu Z. tmchem: a high performance approach for chemical named entity recognition and normalization. J Cheminformatics. 2015; 7(1):3.

    Article  Google Scholar 

  13. Li L, Jin L, Jiang Z, Song D, Huang D. Biomedical named entity recognition based on extended recurrent neural networks. In: Bioinformatics and Biomedicine (BIBM), 2015 IEEE International Conference On. Washington: IEEE: 2015. p. 649–52.

    Google Scholar 

  14. Lu Y, Ji D, Yao X, Wei X, Liang X. Chemdner system with mixed conditional random fields and multi-scale word clustering. J Cheminformatics. 2015; 7(S1):4.

    Article  Google Scholar 

  15. Lample G, Ballesteros M, Subramanian S, Kawakami K, Dyer C. Neural architectures for named entity recognition. In: Proceedings of NAACL-HLT. US: Association for Computational Linguistics: 2016. p. 260–70.

    Google Scholar 

  16. Ma X, Hovy E. End-to-end sequence labeling via bi-directional lstm-cnns-crf. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics. US: Association for Computational Linguistics. 2016;1:1064–74.

  17. Chiu J, Nichols E. Named entity recognition with bidirectional lstm-cnns. Trans Assoc Comput Linguist. 2016; 4(1):357–70.

    Article  Google Scholar 

  18. Habibi M, Weber L, Neves M, Wiegandt DL, Leser U. Deep learning with word embeddings improves biomedical named entity recognition. Bioinformatics. 2017; 33(14):37–48.

    Article  Google Scholar 

  19. Wang X, Zhang Y, Ren X, Zhang Y, Zitnik M, Shang J, Langlotz C, Han J. Cross-type biomedical named entity recognition with deep multi-task learning. Bioinformatics. 2018; 35(10):1745–52.

    Article  Google Scholar 

  20. Collobert R, Weston J. A unified architecture for natural language processing: Deep neural networks with multitask learning. In: Proceedings of the 25th International Conference on Machine Learning. New York City: ACM: 2008. p. 160–7.

    Google Scholar 

  21. Speck-Planche A, Cordeiro MNDS. Multitasking models for quantitative structure–biological effect relationships: current status and future perspectives to speed up drug discovery. Expert Opin Drug Discov. 2015; 10(3):245–56.

    Article  CAS  Google Scholar 

  22. Yang M, Simm J, Lam CC, Zakeri P, van Westen GJ, Moreau Y, Saez-Rodriguez J. Linking drug target and pathway activation for effective therapy using multi-task learning. Sci Rep. 2018; 8:8322.

    Article  Google Scholar 

  23. Crichton G, Pyysalo S, Chiu B, Korhonen A. A neural network multi-task learning approach to biomedical named entity recognition. BMC Bioinformatics. 2017; 18(1):368.

    Article  Google Scholar 

  24. Liu P, Qiu X, Huang X. Adversarial multi-task learning for text classification. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), vol. 1. US: Association for Computational Linguistics. 2017;1:1–10.

  25. Wang Z, Qu Y, Chen L, Shen J, Zhang W, Zhang S, Gao Y, Gu G, Chen K, Yu Y. Label-aware double transfer learning for cross-specialty medical named entity recognition. In: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. US: Association for Computational Linguistics. 2018;1:1–15.

  26. Abdou M, Kulmizev A, Ravishankar V, Abzianidze L, Bos J. What can we learn from semantic tagging? In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. US: SIGDAT: 2018. p. 4881–9.

    Google Scholar 

  27. Hochreiter S, Schmidhuber J. Long short-term memory. Neural Comput. 1997; 9(8):1735–80.

    Article  CAS  Google Scholar 

  28. Lafferty J, McCallum A, Pereira FC. Conditional random fields: Probabilistic models for segmenting and labeling sequence data. In: Proceedings of the 17th International Conference on Machine Learning (ICML). New York City: ACM: 2001. p. 282–9.

    Google Scholar 

  29. Ganin Y, Lempitsky V. Unsupervised domain adaptation by backpropagation. In: Proceedings of the 32nd International Conference on International Conference on Machine Learning. New York City: ACM. 2015;37:1180–9.

  30. Lee J, Yoon W, Kim S, Kim D, Kim S, So CH, Kang J. Biobert: a pre-trained biomedical language representation model for biomedical text mining. 2019. arXiv preprint arXiv:1901.08746.

  31. Yoon W, So CH, Lee J, Kang J. Collabonet: collaboration of deep neural networks for biomedical named entity recognition. BMC Bioinformatics. 2019; 20(10):249.

    Article  Google Scholar 

  32. Gerner M, Nenadic G, Bergman CM. Linnaeus: a species name identification system for biomedical literature. BMC Bioinformatics. 2010; 11(1):85.

    Article  Google Scholar 

Download references

Acknowledgements

We thank the members of the lab for thoughtful discussions.

Funding

This work was supported in part by the National Natural Science Foundation of China (NSFC) [Grant No. 61421003] and the fund of the State Key Lab of Software Development Environment [Grant No. SKLSDE-2017ZX-05]. The funding bodies did not play any role in the design of the study, data collection and analysis, or writing the manuscript.

Author information

Authors and Affiliations

Authors

Contributions

JL contributed to the implementation of the methods, to performing the experiments and to writing the manuscript. XW contributed to the design of the methods and supervised the experiments. LD and KX supervised the whole work. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Ke Xu.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver(http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Wang, X., Lyu, J., Dong, L. et al. Multitask learning for biomedical named entity recognition with cross-sharing structure. BMC Bioinformatics 20, 427 (2019). https://0-doi-org.brum.beds.ac.uk/10.1186/s12859-019-3000-5

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://0-doi-org.brum.beds.ac.uk/10.1186/s12859-019-3000-5

Keywords