Learning A Unified Named Entity Tagger From Multiple Partially Annotated Corpora For Efficient Adaptation
Named entity recognition (NER) identifies typed entity mentions in raw text. While the task is well-established, there is no universally used tagset: often, datasets are annotated for use in downstream applications and accordingly only cover a small set of entity types relevant to a particular task. For instance, in the biomedical domain, one corpus might annotate genes, another chemicals, and another diseases—despite the texts in each corpus containing references to all three types of entities. In this paper, we propose a deep structured model to integrate these “partially annotated” datasets to jointly identify all entity types appearing in the training corpora. By leveraging multiple datasets, the model can learn robust input representations; by building a joint structured model, it avoids potential conflicts caused by combining several models’ predictions at test time. Experiments show that the proposed model significantly outperforms strong multi-task learning baselines when training on multiple, partially annotated datasets and testing on datasets that contain tags from more than one of the training corpora.111The code and the datasets will be made available at https://github.com/xhuang28/NewBioNer
Named Entity Recognition (NER), which identifies the boundaries and types of entity mentions from raw text, is a fundamental problem in natural language processing (NLP). It is a basic component for many downstream tasks, such as relation extraction hasegawa2004discovering; mooney2005subsequence, coreference resolution soon2001machine, and knowledge base construction craven1998learning; craven1999constructing.
One problem in NER is the diversity of entity types, which vary in scope for different domains and downstream tasks. Traditional NER for the news domain focuses on three coarse-grained entity types: person, location, and organization tjong2003introduction. However, as NLP technologies have been applied to a broader set of domains, many other entity types have been targeted. For instance, \newciteritter2011named add seven new entity types (e.g., product, tv-show) on top of the previous three when annotating tweets. Other efforts also define different but partially overlapping sets of entity types walker2006ace; ji2010overview; ere13; aguilar2014comparison. These non-unified annotation schemas result in partially annotated datasets: each dataset is only annotated with a subset of possible entity types.
One approach to this problem is to train individual NE taggers for each partially annotated dataset and combine their results using some heuristics. Figure 1 shows an example that demonstrates the possible shortcomings of this approach, using the biomedical domain as a case study.222https://corposaurus.github.io/corpora/ summarizes dozens of partially annotated biomedical datasets. Here, we train four separate models on four partially annotated datasets: AnatEM pyysalo2013anatomical annotated for the anatomy type, BC2GM smith2008overview for the gene type, JNLPBA kim2004introduction for cell types, and Linnaeus gerner2010linnaeus for the species type. We can see that the models’ predictions contradict each other when applied to the same test sentence—making it a challenge to accurately combine them.
In this paper, we propose a deep structured model to leverage multiple partially annotated datasets, allowing us to jointly identify the union of all entity types presented in the training data. The model leverages supervision signals across diverse datasets to learn robust input representations, thus improving the performance for each entity type. Moreover, it makes joint predictions to avoid potential conflicts among models built on different entity types, allowing further improvement for cross-type NER.
Experiments on both real-world and synthetic datasets show that our model can efficiently adapt to new corpora that have more types than any individual dataset used for training and that it achieves significantly better results compared to strong multi-task learning baselines.
2 Problem Statement
We formally define the problem by first defining our terminology.
Global Tag Space. Let denote a corpus, and denote the set of entity types that are tagged in corpus . When there are a set of corpora , each has its own tag space concerning different entity types, the global tag space is defined as the union of the local tag space. Formally, .
Partially Annotated Corpus. If , then is a partially annotated corpus.
Global Evaluation. When a model is trained on a set of partially annotated corpora and predicts tags for the whole global tag space , we say it is making global predictions. Accordingly, the evaluation of the models’ performance on is called global evaluation.
Our goal is to train a single unified NE tagger from several partially annotated corpora for efficient adaptation to new corpora that have more types than any individual dataset used during training. Formally, we have a set of corpora , and we propose to train a joint model on such that it makes predictions for the global tag space . One benefit of this joint model is that it can be easily adapted to a new tag space where , and .
3 Background and Related Work
In this section, we first introduce neural architectures for NER which our work builds upon and then summarize previous work on imperfect annotation problems.
3.1 Neural Architectures for NER
With recent advances using deep neural networks, bi-directional long short-term memory networks with conditional random fields (BiLSTM-CRF) have become standard for NER lample2016neural. A typical architecture consists of a BiLSTM layer to learn feature representations from the input and a CRF layer to model the inter-dependencies between adjacent labels and perform joint inference. \newcitema2016end introduce additional character-level convolutional neural networks (CNNs) to capture subword unit information. In this paper, we use a BiLSTM-CRF with character-level modeling as our base model. We now briefly review the BiLSTM-CRF model.
Long Short Term Memory networks (LSTMs) hochreiter1997long are a variation of RNNs that are designed to avoid the vanishing/exploding gradient problem bengio1994learning. Specifically, BiLSTMs take as input a sequence of words and output a sequence of hidden vectors: BiLSTMs combine a left-to-right (forward) and a right-to-left (backward) LSTM to capture both left and right context. Formally, they produce a hidden vector for each input, where and are produced by the forward and the backward LSTMs respectively; denotes vector concatenation.
Following \newcitewang2018cross, we use a BiLSTM for character-level modeling. We concatenate the hidden vector of the space after a word from the forward LSTM and the hidden vector of the space before a word from the backward LSTM to form a character-level representation of the word: . The word-level BiLSTM then takes the concatenation of and the word embedding as input to learn contextualized representations.
Conditional Random Fields (CRFs) lafferty2001conditional are sequence tagging models that capture the inter-dependencies between the output tags; they have been widely used for NER mccallum2003early; lu2015chemdner; pengnamed; pengimproving; peng2016multi. Given a set of training data , a CRF minimizes negative log-likelihood:
where is any possible tag sequence with the same length as , is the potential of the tag sequence , and is the potential of the gold tag sequence. The numerator is called the gold energy function, and the denominator is the partition function. The likelihood function using globally annotated data is illustrated in Figure 1(a). The potential of a tag sequence can be computed as:
where is the th element in ( is the start of the sequence), and
where is the transition score from to , and is the emission score of computed based on the output of the BiLSTM.
3.2 Learning from Imperfect Annotations
Learning from multiple partially annotated datasets could be more generally thought of as learning from imperfect annotations. In that broad sense, there are several notable areas of prior work. One of the most prominent concerns learning from incomplete annotations (noisy labels), where some occurrences of entities are neglected in the annotation process and falsely labeled as non-entities (negative). A related problem is learning from unlabeled data with distant supervision.
A major challenge of all these settings, including ours, is that a positive instance might be labeled as negative. A well-explored solution to this problem is proposed by \newciteC08-1113, which instead of maximizing the likelihood of the gold tag sequence, we maximize the total likelihood for all possible tag sequences consistent with the gold labels. \newciteC08-1113,Yang2014SemiSupervisedCW applied this idea to the incomplete annotation setting; \newciteshang2018learning,D14-1093 applied it to the unlabeled data with distant supervision setting; and \newcitegreenberg2018marginal applied it to the partial annotation setting. While this is a general solution, its primary drawback is that it assumes a uniform prior on all labels consistent with the gold labels. This may have the result of overly encouraging the prediction of entities, resulting in low precision.
To tackle the problem of incomplete annotations, \newciteCarlson2009LearningAN,yang2018distantly explored bootstrap-based semi-supervised learning on unlabeled data, iteratively identifying new entities with the taggers and then re-training the taggers. \newcitebellare2007learning,li2005learning,fernandes2011learning explored an EM algorithm with semi-supervision.
For the partial annotation problem, most previous work has focused on building individual taggers for each dataset and using single-task learning liu2017empower or multi-task learning crichton2017neural; wang2018cross. In single-task learning, each model is trained separately on each dataset , and makes local predictions on . Based on the neural-CRF architecture, multi-task learning uses a different CRF layer for each dataset (each task) to make local predictions on , and shares the lower-level representation learning component across all tasks. Both single-task learning and multi-task learning make local predictions and have to apply heuristics to combine the model predictions, resulting in the collision problem demonstrated in Figure 1.
To the best of our knowledge, \newcitegreenberg2018marginal is the only prior work trying to build a unified model from multiple partially annotated corpora. We will show that their model, which is reminiscent of \newciteC08-1113, is a special case of ours and that our other variations achieve better performance. In addition, they only evaluated the model on the training corpora while we conduct evaluations to test the model’s ability to adapt to new corpora with different tag spaces.
As mentioned above, we use a BiLSTM-CRF with character-level modeling as our base model. Our goal is to build a unified model to make global predictions. That is, our model will be jointly trained on multiple partially annotated datasets and make predictions on the global tag space . Such a unified model will enjoy the benefit of learning robust representations from multiple datasets just like multi-task learning while maintaining a joint probability distribution of the global tag space to avoid possible conflicts from individual models.
4.1 Naive Approach
A simple solution to the problem is to merge all the datasets into one giant corpus. A single model can then be trained on this corpus to make global predictions. However, such a corpus will be missing many correct annotations, since each portion will be annotated with only a subset of the target entity types. Figure 1(b) shows an example: here, a location (Texas) exists but is labeled as a non-entity, because the original dataset from which this sentence is drawn does not annotate locations at all. As a result, this approach suffers from false penalties when applying the original likelihood function (Eq. 2-4) to train the model, meaning that it penalizes predictions that correctly identify entities with types that are not annotated for a particular sentence.
4.2 Improving the Gold Energy Function
One way to improve performance is to explicitly acknowledge the incompleteness of the existing “gold” annotations and to give the model credit for predicting any tag sequence that is consistent with the partial annotations. This can be done by modifying the CRF’s gold energy function, illustrated in the upper part of Figure 1(c). Specifically, in this example, John is labeled as PER, so PER is the only possible correct tag at that position. However, lives, in, and Texas are labeled as O (non-entity), which here means only that they may not be PER—but any of them could be LOC, since locations are not annotated for this sentence. Therefore, any sequence that assigns either O or LOC for any of these three positions is consistent with the gold labels. To account for this, we modify the gold energy function to credit all tag sequences that are consistent with the gold annotations, encouraging the model to predict other consistent labels when the gold label is O. \newciteC08-1113 propose a specific solution that applies this idea on incomplete annotations: instead of maximizing the likelihood of the gold tag sequence when optimizing the CRF model, they maximize the total likelihood of all possible tag sequences consistent with the gold labels. This approach is later used by \newcitegreenberg2018marginal to handle the problem of partial annotation. We will address a potential problem with their method and propose a generalized version in Section 4.4.
4.3 Improving the Partition Function
Modifying the gold energy function will give credit to a system for producing alternative entity labels for words tagged as O in the partially annotated training. A different solution is to simply not penalize predictions of such alternative labels. This can be done by modifying the partition function and keeping the gold energy function unchanged. The lower part of Figure 1(c) gives an illustration. As stated above, LOC is a consistent alternative entity label for lives, in, and Texas. We therefore exclude from our calculations any paths that include LOC at any of those positions. More generally, we exclude all such consistent but alternative tag sequences from the computation of the CRF’s partition function. Section 4.4 gives formal definitions with equations. The improved partition function sets the model free to predict alternative labels without penalty (as long as they are consistent with the known gold annotations), but it does not give them any positive credit for doing so (as in the previous approach). We hypothesize that the improved partition function would work better than the improved gold energy function in our setting because it addresses the false penalties problem more precisely. We will verify this hypothesis in our experiments.
4.4 Discounting Alternative Sequences
There is a potential problem with naively applying the improved gold energy function: when the gold label is O, the model is encouraged to predict other consistent labels as strongly as it is encouraged to predict O. However, many O labels are confident annotations of O. As a result, naively training with the improved gold energy function tends to over-predict entities and not predict Os. To mitigate this problem, we discount the energy of tag sequences that go through alternative labels. This can be achieved by introducing a hyper-parameter as a discounting factor for the gold energy function. Formally, we modify Eq 3 to:
where is the set of alternative labels. We thus have the improved gold energy function:
where is the set of all tag sequences that are consistent with the gold sequence, including the gold sequence itself.
Similarly, for the improved partition function, we can use the same strategy to discount the energy of alternative sequences rather than completely removing them. We thus introduce another and the improved partition function becomes:
4.5 Combining Improved Functions
For generality, we combine the improved gold energy and the improved partition function to make a new likelihood function as our final model:
To ensure Equation 7 is a valid likelihood function (the probabilities of all sequences sum to 1), we need a constraint that . Note that Equation 7 subsumes all models discussed in this section. Specifically, when , the model is the Naive Model discussed in Section 4.1; when , the model is the same as \newcitegreenberg2018marginal discussed in Section 4.2; when , the model is the same as proposed in Section 4.3. We have a general perspective of all the models by simply treating and as hyper-parameters.
Note that for the Naive Model, since , the Equation 7 is not always a valid likelihood function333This may be confusing because when it looks exactly the same as the original CRF likelihood function. But in the partial annotation setting, this means that the scores of alternative sequences will be zero in the numerator but non-zero in the denominator, which makes the total likelihood less than 1. It suggests that the original CRF likelihood function is not suitable for the partial annotation setting.. This may partially explain why the Naive Model performs so poorly under this setting. We posit that the model will work the best when .
5 Experimental Setup
Our goal is to train a unified NER model on multiple partially annotated datasets. This model will make global predictions and can efficiently adapt to new corpora that contain tags from more than one training corpus. To fully test this capability, we would need a single test set annotated with all types of interest. However, the motivation behind this effort is that such a dataset typically does not exist. We therefore take two approaches to approximate such an evaluation.
In the first evaluation setting, we take advantage of the fact that although there may not be a single dataset annotated with all types of named entities of interest, there exist several datasets that cover types from more than one of the training corpora. Specifically, we are able to select test corpora that each cover types of interest from multiple training corpora. Table 1 shows the biomedical corpora we use and their entity types. For example, we use BC5CDR for global evaluation, because its entity types (Chemical and Disease) cover multiple training corpora (BC4CHEM for Chemical and NCBI for Disease).
|For Global Evaluations|
In the second evaluation setting, we create synthetic datasets from the CoNLL 2003 NER dataset to simulate training and global evaluations. Specifically, the CoNLL 2003 dataset is annotated with four entity types: location, person, organization, and miscellaneous entities. We randomly split the training set into four portions, each containing only one entity type (all other types are removed). In this setting, the four portions of the training set are used for training and the original dataset with all entities annotated is used as a global corpus.
More details about all the datasets can be found in Appendix A.1.
5.1.1 Biomedical Dataset Analysis
The motivation for this work rests on the assumption that even when a dataset is annotated for a certain set of entity types, it likely contains other types of entities that are unlabeled. To verify this assumption, we expand the annotations of each dataset using heuristics and compute the pairwise mention-level overlap between the datasets. Specifically, suppose we are comparing two datasets, A and B. We first construct A’ and B’, where A’ contains all mentions in A but is augmented with new mentions found by taking all strings annotated in B and marking them as named entities in A (regardless of context; there may obviously be some errors). We do the same (in the opposite direction) to construct B’. We then compute the pairwise overlap coefficient between A’ and B’ according to the following criterion:
Figure 3 shows the heat maps. For the training group, BC2GM, BC4CHEMD, and Linnaeus are considerably overlapped, although they are annotated with different entity types (GP, Chemical, and Species). This confirms our assumption that although the datasets are annotated for a subset of entity types, they contain other types that are unlabeled.444We further verified this conclusion by computing the heat maps on the original datasets. The overlaps between BC2GM and BC4CHEMD, and BC2GM and Linnaeus are nearly .
|Corpus||Trained on Other Biomedical Datasets||Traind on CoNLL|
We borrow most of the best hyper-parameters reported by \newcitewang2018cross. The hidden sizes of the BiLSTMs are tuned, and the best value we found is 100 for the character-level BiLSTM, and 300 for the word-level BiLSTM. We also tuned both discounting factors and in the range of [0,0.2,0.4,0.6,0.8,1.0]. It turns out that (using improved partition function) and (using improved gold energy function) make two local optimums. Therefore we report the performance of three special cases of our proposed framework, with , , and (the naive model), respectively.
5.3 Compared Models.
We compare different variations of our unified model and other models in different settings. We first train models on all training corpora, and then perform evaluations under two scenarios: (1) no-supervision: directly evaluating the trained models on each global corpora; (2) limited-supervision: fine-tuning the models on a small subset of the training portion of each global corpus before the evaluations.
Under both scenarios, we report performance of four different models:
MTM/MTM-vote: Train a multi-task model (MTM) on training corpora, using a separate CRF for each corpus. (This is the current state-of-the-art structure wang2018cross when evaluated on the training corpora.)
Under the no-supervision setting, we heuristically combine all existing CRF’s predictions to make global predictions. Specifically, we apply two heuristics to resolve conflicts while preserving entity chunk-level consistency. First, where predictions from more than one model overlap, we expand each prediction’s boundary to the outermost position. Second, we always favor the predictions of named entities over the predictions of non-entity.555A lower recall and f1 score was observed in the initial experiment without this heuristic.
Under the limited-supervision setting, for each global corpus, we add a new CRF and train it along with the LSTMs.
Unified-01: Use the naive training approach described in 4.1; this corresponds to our unified model with settings .
Unified-11: Use the improved gold energy function described in 4.2; this corresponds to our unified model with settings and is equivalent to the model proposed by \newcitegreenberg2018marginal.
Unified-00: Use the improved partition function proposed in 4.3; this corresponds to our unified model with settings .
Among the compared models, Unified-01 (the naive model) and MTM/MTM-Vote are either simple or commonly used methods and thus are treated as baselines. Unified-00 is a novel approach. Although \newcitegreenberg2018marginal used the approach of Unified-11, they only evaluated the model on training corpora/tasks while we apply it for task adaptation. Moreover, it is a special case of our proposed framework, thus we argue that people can simply tune and to get good performance for adaptations to new tasks.
As mentioned above, we compare the results of four different approaches in no-supervision and limited-supervision settings, both with real-world biomedical data and synthetic news data.
As a sanity check, we also evaluate the models on the test sets of the training corpora. The results can be found in Appendix A.2. It is shown that our MTM performs comparably with state-of-the-art systems evaluated on the training corpora, and thus is a strong baseline.
6.1 No-Supervision Setting
Table 2 demonstrates the results for task adaptation in the no-supervision setting. We report precision and recall in addition to f1 scores to better show the differences between the models.
Comparing on f1 scores, Unified-00 (our new model) significantly outperforms all other models on three out of four datasets, demonstrating its effectiveness. Unified-11 also achieves good results, with higher recall but lower precision than Unified-00. This aligns well with our hypothesis that it encourages predictions of entities. Conversely, Unified-01 (the naive approach) achieves the highest precision but lowest recall, which is reasonable considering the problem of false penalties that discourages the model from predicting entities. We also found that the model achieves better performance when , which supports our hypothesis in 4.5 that the model works better with a valid likelihood function.
6.2 Limited-Supervision Setting
To further demonstrate the models’ ability to adapt to new datasets with a small amount of supervision, we sample a small subset of the training portion of each global evaluation corpus to fine-tune the trained models. We show the performance of the models fine-tuned with different amounts of sampled data. For each global corpus, we show a single-task model (STM) trained on it with a reasonable amount of data (two thousand sentences for the biomedical corpora). In the CoNLL 2003 setting, we train the STM on the entire training data for a fair comparison, because all other models are first trained on the four training portions, which essentially look through the entire training set (just partially annotated). The results of the STMs are used as benchmarks. Experimental results are presented in Figure 4.
Firstly, with much less training data, all the models achieve comparable or noticeably better performance than the STMs trained from scratch, demonstrating that training on the partially annotated corpora does help to boost performance on global evaluation corpora. Additionally, MTMs are worse than all the unified models, because they only share the LSTM layers, but lose all the knowledge in the CRFs when adapted to new corpora. The unified models have the advantage that they can reuse the robust CRFs learned from a large amount of data. This is more obvious in the CoNLL 2003 evaluation setting, where the unified models that reuse the pre-trained CRFs achieve good performance trained with only 50 sentences, but the MTM, which does not reuse the CRFs, needs a larger amount of training data to catch up.
In general, Unified-00, our novel approach proposed here, still performs the best on every dataset. We note that although Unified-01 has an extremely low recall on the CoNLL 2003 dataset in the no-supervision setting, it works surprisingly well in the limited-supervision setting. On the other hand, Unified-00 and Unified-11 generally perform better than Unified-01 on real-world biomedical datasets, especially when fine-tuned on less data. Again, since all the unified models are special cases of our proposed framework, we argue that, for adapting to new datasets, people can simply tune the discounting factors and to get good results.
7 Conclusion and Future Work
In this paper, we propose a unified model that learns from multiple partially annotated datasets to make joint predictions on the union of entity types appearing in any training dataset. The model integrates learning signals from different datasets and avoids potential conflicts that would result from combining independent predictions from multiple models. Experiments show that the proposed unified model can efficiently adapt to new corpora that have more entity types than any of the training corpora, and performs better than the baseline approaches.
In future work, we plan to explore other algorithms (e.g. imitation learning) that allow the model the explore the unknown space during training, using delayed rewards to decide whether the model should trust its exploration. Analysis of the global evaluation results suggests that the unified model is under-predicting, meaning there is still room for improvement specifically on recall. We plan to explore further changes to the current objectives to encourage more entity predictions.
Finally, the approach proposed in this paper also does not handle entity types of varying granularities or tagsets with mismatched guidelines (e.g. one dataset annotates only for-profit companies as ORG and one annotates all formalized groups). Effectively modeling these complications is an interesting area for future work.
We thank the anonymous reviewers for their constructive comments, as well as the members of the USC PLUS lab for their early feedbacks. We thank Tianyu Meng and Yuxin Zhou for their help with initial data processing and experimental setup. This work is supported in part by DARPA (HR0011-15-C-0115) and an NIH R01 (LM012592). Approved for Public Release, Distribution Unlimited. The views expressed are those of the authors and do not reflect the official policy or position of the sponsors.
Appendix A Appendix
|JNLPBA||Gene/Protein, DNA, Cell-type, Cell-line, RNA||24,806||595,994||59,965|
|BioNLP13CG||Gene/Protein, Disease, Chemical, Others||1,906||52,771||6881|
|BioNLP11ID||Gene/Protein, Chemical, Others||5178||166416||11084|
Below we introduce the datasets in the biomedicine domain and the news domain.
a.1.1 Biomedicine domain: Local training group
The training group consists of five datasets: BC2GM, BC4CHEM, NCBI-disease, JNLPBA, and Linnaeus. The first two datasets are from different BioCreative shared tasks smith2008overview; krallinger2015chemdner; wei2015overview. NCBI-disease is created by \newcitedougan2014ncbi for disease name recognition and normalization. JNLPBA comes from the 2004 shared task from joint workshop on natural language processing in biomedicine and its applications kim2004introduction, and Linnaeus is a species corpus composed by \newcitegerner2010linnaeus. More information about the datasets can be found in Table 3.
Below are detailed descriptions of the datasets:
BC2GM is a gene/protein corpus. The annotation is Gene. It’s provided by the BioCreative II Shared Task for gene mention recognition.
BC4CHEM is a chemical corpus. The annotation is Chemical. It’s provided by the BioCreative IV Shared Task for chemical mention recognition.
NCBI-disease is a disease corpus. The annotation is Disease. It was introduced for disease name recognition and normalization.
JNLPBA consists of DNA, RNA, Gene/Protein, Cell line, Cell Type. The annotation is same as the NE names, except the Gene/Protein is annotated with Protein. It was provided by 2004 JNLPBA Shared Task for biomedical entity recognition.
Linnaeus is a species corpus. The annotation is Species. The original project was created for entity mention recognition.
a.1.2 Biomedicine domain: Global evaluation group
We reemphasize here that the purpose of the global evaluation is to test the model’s ability to making global predictions and efficiently adapt to global corpora. While no corpus is globally annotated, we identify several existing corpora to approximate the global evaluation. Each test corpus is annotated with a superset of several training corpora to test the model’s generalizability outside of the local tag spaces.
The global evaluation group contains three datasets: BC5CDR, BioNLP13CG, and BioNLP11ID. Each is annotated with multiple entity types. BC5CDR comes from the BioCreative shared tasks smith2008overview; krallinger2015chemdner; wei2015overview. BioNLP13CG and BioNLP11ID come from the BioNLP shared task kim2013genia. More information about the global evaluation datasets can be found in Table 4.
Below are detailed descriptions of the datasets:
BC5CDR is a chemical and disease corpus. The annotation is Chemical and Disease. It’s provided by BioCreative V Shared Task for chemical and disease mention recognition.
BioNLP13CG consists of Gene/Protein and Related Product, Cancel, Chemical, Anatomy and Organism and others. BioNLP11ID consists of Gene/Protein, Chemical, and Organism. The annotation is same as the NE types but has a finer ontology scope.
There are inconsistencies between the entity type names in different datasets, mainly due to different granularities. To remove this unnecessary noise, we manually merged some entity types. For example, we unify Gene and Protein into Gene/Protein as they are commonly used interchangeably; we merge “Simple Chemical” to “Chemical” and leave the problem of entity type granularity for future work. The information in Table 3 and 4 reflects the merged types.
a.1.3 News domain: CoNLL 2003 NER dataset
We use the CoNLL 2003 NER dataset (sang2003introduction) to evaluate the models in news domain. More information about the dataset can be found in Table 5. We use synthetic data from the dataset to simulate local training and global evaluation. Specifically, the CoNLL 2003 NER dataset is annotated with four entity types: location, person, organization, and miscellaneous entities. We randomly split the training set into four portions, each contains only one entity type respectively, with other types changed to ”O”. The models are trained on the four training portions and we test on the original test set with all entity types annotated.
a.1.4 Data split
For the news domain, we use the default train, dev, test portion of the CoNLL 2003 NER dataset. For the biomedicine domain, we follow the data split in \newcitecrichton2017neural for both the training and the evaluation groups. All datasets are divided into three portions: train, dev, and test. We train the model on the training set of the training group and tune the hyper-parameters on the corresponding development set. Global evaluations are performed on the test set of the evaluation group.
a.2 Local Evaluation
For a sanity check, we evaluate the models on the training corpora and compare the results with state-of-the-art systems. In this setting, all the models are trained on the training set of the training corpora (without fine-tuning on global evaluation corpora) and evaluated on their test set. The results are shown in Table 6. STM is the single-task models we implemented, following the settings in \newcitewang2018cross. The SOTA is achieved by \newcitewang2018cross with multi-task model, which is shown in the table as MTM \newcitewang2018cross. They trained their model on BC2GC, BC4CHM, NCBI, JNLPBA, and BC5CDR. MTM (ours) is the multi-task model we trained on our five training corpora and used as a baseline in the global evaluations. It has the same architecture as \newcitewang2018cross.
As we can see, MTM \newcitewang2018cross achieves the best results on 3 out of 4 datasets. And our MTM achieves very similar results, showing it is a strong model on training corpora. Our proposed models do not perform very well when evaluated on the training corpora. But in the global evaluation setting, they perform much better compared to our strong MTM. This demonstrates the superiority of our proposed models on task adaptation.