This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Deep Span Representations for Named Entity Recognition

Enwei Zhu    Yiyang Liu    Jinpeng Li111Corresponding author.
Ningbo No.2 Hospital
Ningbo Institute of Life and Health Industry, University of Chinese Academy of Sciences
{zhuenwei,liuyiyang,lijinpeng}@ucas.ac.cn
Abstract

Span-based models are one of the most straightforward methods for named entity recognition (NER). Existing span-based NER systems shallowly aggregate the token representations to span representations. However, this typically results in significant ineffectiveness for long entities, a coupling between the representations of overlapping spans, and ultimately a performance degradation. In this study, we propose DSpERT (Deep Span Encoder Representations from Transformers), which comprises a standard Transformer and a span Transformer. The latter uses low-layered span representations as queries, and aggregates the token representations as keys and values, layer by layer from bottom to top. Thus, DSpERT produces span representations of deep semantics.

With weight initialization from pretrained language models, DSpERT achieves performance higher than or competitive with recent state-of-the-art systems on six NER benchmarks.111Our code is available at https://github.com/syuoni/eznlp. Experimental results verify the importance of the depth for span representations, and show that DSpERT performs particularly well on long-span entities and nested structures. Further, the deep span representations are well structured and easily separable in the feature space.

1 Introduction

As a fundamental information extraction task, named entity recognition (NER) requires predicting a set of entities from a piece of text. Thus, the model has to distinguish the entity spans (i.e., positive examples) from the non-entity spans (i.e., negative examples). In this view, it is natural to enumerate all possible spans and classify them into the entity categories (including an extra non-entity category). This is exactly the core idea of span-based approaches (Sohrab and Miwa, 2018; Eberts and Ulges, 2020; Yu et al., 2020).

Analogously to how representation learning matters to image classification (Katiyar and Cardie, 2018; Bengio et al., 2013; Chen et al., 2020), it should be crucial to construct good span representations for span-based NER. However, existing models typically build span representations by shallowly aggregating the top/last token representations, e.g., pooling over the sequence dimension (Sohrab and Miwa, 2018; Eberts and Ulges, 2020; Shen et al., 2021), or integrating the starting and ending tokens (Yu et al., 2020; Li et al., 2020d). In that case, the token representations have not been fully interacted before they are fed into the classifier, which impairs the capability of capturing the information of long spans. If the spans overlap, the resulting span representations are technically coupled because of the shared tokens. This causes the representations less distinguishable from the ones of overlapping spans in nested structures.

Inspired by (probably) the most sophisticated implementation of attention mechanism — Transformer and BERT (Vaswani et al., 2017; Devlin et al., 2019), we propose DSpERT, which stands for Deep Span Encoder Representations from Transformers. It consists of a standard Transformer and a span Transformer; the latter uses low-layered span representations as queries, and token representations within the corresponding span as keys and values, and thus aggregates token representations layer by layer from bottom to top. Such multi-layered Transformer-style aggregation promisingly produces deep span representations of rich semantics, analogously to how BERT yields highly contextualized token representations.

With weight initialization from pretrained language models (PLMs), DSpERT performs comparably to recent state-of-the-art (SOTA) NER systems on six well-known benchmarks. Experimental results clearly verify the importance of the depth for the span representations. In addition, DSpERT achieves particularly amplified performance improvements against its shallow counterparts222In this paper, unless otherwise specified, we use “shallow” to refer to models that construct span representations by shallowly aggregating (typically top) token representations, although the token representations could be “deep”. on long-span entities and nested structures.

Different from most related work which focuses on the decoder designs (Yu et al., 2020; Li et al., 2020b; Shen et al., 2021; Li et al., 2022), we make an effort to optimize the span representations, but employ a simple and standard neural classifier for decoding. This exposes the pre-logit representations that directly determine the entity prediction results, and thus allows further representation analysis widely employed in a broader machine learning community (Van der Maaten and Hinton, 2008; Krizhevsky et al., 2012). This sheds light on neural NER systems towards higher robustness and interpretability (Ouchi et al., 2020).

2 Related Work

The NER research had been long-term focused on recognizing flat entities. After the introduction of linear-chain conditional random field (Collobert et al., 2011), neural sequence tagging models became the de facto standard solution for flat NER tasks (Huang et al., 2015; Lample et al., 2016; Ma and Hovy, 2016; Chiu and Nichols, 2016; Zhang and Yang, 2018).

Recent studies pay much more attention to nested NER, which a plain sequence tagging model struggles with (Ju et al., 2018). This stimulates a number of novel NER system designs beyond the sequence tagging framework. Hypergraph-based methods extend sequence tagging by allowing multiple tags for each token and multiple tag transitions between adjacent tokens, which is compatible with nested structures (Lu and Roth, 2015; Katiyar and Cardie, 2018). Span-based models enumerate candidate spans and classify them into entity categories (Sohrab and Miwa, 2018; Eberts and Ulges, 2020; Yu et al., 2020). Li et al. (2020b) reformulates nested NER as a reading comprehension task. Shen et al. (2021, 2022) borrow the methods from image object detection to solve nested NER. Yan et al. (2021) propose a generative approach, which encodes the ground-truth entity set as a sequence, and thus reformulates NER as a sequence-to-sequence task. Li et al. (2022) describe the entity set by word-word relation, and solve nested NER by word-word relation classification.

The span-based models are probably the most straightforward among these approaches. However, existing span-based models typically build span representations by shallowly aggregating the top token representations from a standard text encoder. Here, the shallow aggregation could be pooling over the sequence dimension (Eberts and Ulges, 2020; Shen et al., 2021), integrating the starting and ending token representations (Yu et al., 2020; Li et al., 2020d), or a concatenation of these results (Sohrab and Miwa, 2018). Apparently, shallow aggregation may be too simple to capture the information embedded in long spans; and if the spans overlap, the resulting span representations are technically coupled because of the shared tokens. These ultimately lead to a performance degradation.

Our DSpERT addresses this issue by multi-layered and bottom-to-top construction of span representations. Empirical results show that such deep span representations outperform the shallow counterpart qualitatively and quantitatively.

3 Methods

Deep Token Representations.

Given a TT-length sequence passed into an LL-layered dd-dimensional Transformer encoder (Vaswani et al., 2017), the initial token embeddings, together with the potential positional and segmentation embeddings (e.g., BERT; Devlin et al., 2019), are denoted as 𝗛0T×d\boldsymbol{\mathsf{H}}^{0}\in\mathbb{R}^{T\times d}. Thus, the ll-th (l=1,2,,Ll=1,2,\dots,L) token representations are:

𝗛l=TrBlock(𝗛l1,𝗛l1,𝗛l1),\boldsymbol{\mathsf{H}}^{l}=\operatorname{TrBlock}(\boldsymbol{\mathsf{H}}^{l-1},\boldsymbol{\mathsf{H}}^{l-1},\boldsymbol{\mathsf{H}}^{l-1}), (1)

where TrBlock(𝗤,𝗞,𝗩)\operatorname{TrBlock}(\boldsymbol{\mathsf{Q}},\boldsymbol{\mathsf{K}},\boldsymbol{\mathsf{V}}) is a Transformer encoder block that takes 𝗤T×d\boldsymbol{\mathsf{Q}}\in\mathbb{R}^{T\times d}, 𝗞T×d\boldsymbol{\mathsf{K}}\in\mathbb{R}^{T\times d}, 𝗩T×d\boldsymbol{\mathsf{V}}\in\mathbb{R}^{T\times d} as the query, key, value inputs, respectively. It consists of a multi-head attention module and a position-wise feed-forward network (FFN), both followed by a residual connection and a layer normalization. Passing the same matrix, i.e., 𝗛l1\boldsymbol{\mathsf{H}}^{l-1}, for queries, keys and values exactly results in self-attention (Vaswani et al., 2017).

The resulting top representations 𝗛L\boldsymbol{\mathsf{H}}^{L}, computed through LL Transformer blocks, are believed to embrace deep, rich and contextualized semantics that are useful for a wide range of tasks. Hence, in a typical neural NLP modeling paradigm, only the top representations 𝗛L\boldsymbol{\mathsf{H}}^{L} are used for loss calculation and decoding (Devlin et al., 2019; Eberts and Ulges, 2020; Yu et al., 2020).

\includestandalone

[width=]tikz/deep-span

Figure 1: Architecture of DSpERT. It comprises: (Left) a standard LL-layer Transformer encoder (e.g., BERT); and (Right) a span Transformer encoder, where the span representations are the query inputs, and token representations (from the Transformer encoder) are the key/value inputs. There are totally K1K-1 span Transformer encoders, where KK is the maximum span size; and each has LL layers. The figure specifically displays the case of span size 3; the span of positions 1–3 is highlighted, whereas the others are in dotted lines.

Deep Span Representations.

Figure 1 presents the architecture of DSpERT, which consists of a standard Transformer encoder and a span Transformer encoder. In a span Transformer of size kk (k=2,3,,Kk=2,3,\dots,K), the initial span representations 𝗦0,k(T+k1)×d\boldsymbol{\mathsf{S}}^{0,k}\in\mathbb{R}^{(T+k-1)\times d} are directly aggregated from the corresponding token embeddings:

𝒔i0,k=Aggregating(𝗛[i:i+k]0),\boldsymbol{\mathit{s}}^{0,k}_{i}=\operatorname{Aggregating}(\boldsymbol{\mathsf{H}}^{0}_{[i:i+k]}), (2)

where 𝒔i0,kd\boldsymbol{\mathit{s}}^{0,k}_{i}\in\mathbb{R}^{d} is the ii-th vector of 𝗦0,k\boldsymbol{\mathsf{S}}^{0,k}, and 𝗛[i:i+k]0=[𝒉i0;;𝒉i+k10]k×d\boldsymbol{\mathsf{H}}^{0}_{[i:i+k]}=[\boldsymbol{\mathit{h}}^{0}_{i};\dots;\boldsymbol{\mathit{h}}^{0}_{i+k-1}]\in\mathbb{R}^{k\times d} is a slice of 𝗛0\boldsymbol{\mathsf{H}}^{0} from position ii to position i+k1i+k-1; Aggregating()\operatorname{Aggregating}(\cdot) is a shallowly aggregating function, such as max-pooling. Check Appendix A for more details on alternative aggregating functions used in this study. Technically, 𝒔i0,k\boldsymbol{\mathit{s}}^{0,k}_{i} covers the token embeddings in the span (i,i+k)(i,i+k).

The computation of high-layered span representations imitates that of the standard Transformer. For each span Transformer block, the query is a low-layered span representation vector, and the keys and values are the aforementioned token representation vectors in the positions of that very span. Formally, the ll-th layer span representations are:

𝒔il,k=SpanTrBlock(𝒔il1,k,𝗛[i:i+k]l1,𝗛[i:i+k]l1),\boldsymbol{\mathit{s}}^{l,k}_{i}=\operatorname{SpanTrBlock}(\boldsymbol{\mathit{s}}^{l-1,k}_{i},\boldsymbol{\mathsf{H}}^{l-1}_{[i:i+k]},\boldsymbol{\mathsf{H}}^{l-1}_{[i:i+k]}), (3)

where SpanTrBlock(𝗤,𝗞,𝗩)\operatorname{SpanTrBlock}(\boldsymbol{\mathsf{Q}},\boldsymbol{\mathsf{K}},\boldsymbol{\mathsf{V}}) shares the exactly same structure with the corresponding Transformer block, but receives different inputs. More specifically, for span (i,i+k)(i,i+k), the query is the span representation 𝒔il1,k\boldsymbol{\mathit{s}}^{l-1,k}_{i}, and the keys and values are the token representations 𝗛[i:i+k]l1\boldsymbol{\mathsf{H}}^{l-1}_{[i:i+k]}. Again, the resulting 𝒔il,k\boldsymbol{\mathit{s}}^{l,k}_{i} technically covers the token representations in the span (i,i+k)(i,i+k) on layer l1l-1.

In our default configuration, the weights of the standard and span Transformers are independent, but initialized from a same PLM. Given the exactly same structure, the weights can be optionally shared between the two modules. This reduces the model parameters, but empirically results in slightly lower performance (See Appendix F).

The top span representations 𝗦L,k\boldsymbol{\mathsf{S}}^{L,k} are built through LL Transformer blocks, which are capable of enriching the representations towards deep semantics. Thus, the representations of overlapping spans are decoupled, and promisingly distinguishable from each other, although they are originally built from 𝗦0,k\boldsymbol{\mathsf{S}}^{0,k} — those shallowly aggregated from token embeddings. This is conceptually analogous to how the BERT uses 12 or more Transformer blocks to produce highly contextualized representations from the original static token embeddings.

The top span representations are then passed to an entity classifier. Note that we do not construct a unigram span Transformer, but directly borrow the token representations as the span representations of size 1. In other words,

𝗦L,1𝗛L.\boldsymbol{\mathsf{S}}^{L,1}\equiv\boldsymbol{\mathsf{H}}^{L}. (4)

Entity Classifier.

Following Dozat and Manning (2017) and Yu et al. (2020), we introduce a dimension-reducing FFN before feeding the span representations into the decoder. According to the preceding notations, the representation of span (i,j)(i,j) is 𝒔iL,ji\boldsymbol{\mathit{s}}^{L,j-i}_{i}, thus,

𝒛ij=FFN(𝒔iL,ji𝒘ji),\boldsymbol{\mathit{z}}_{ij}=\operatorname{FFN}(\boldsymbol{\mathit{s}}^{L,j-i}_{i}\oplus\boldsymbol{\mathit{w}}_{j-i}), (5)

where 𝒘jidw\boldsymbol{\mathit{w}}_{j-i}\in\mathbb{R}^{d_{w}} is the (ji)(j-i)-th width embedding from a dedicated learnable matrix; \oplus means the concatenation operation. 𝒛ijdz\boldsymbol{\mathit{z}}_{ij}\in\mathbb{R}^{d_{z}} is the dimension-reduced span representation, which is then fed into a softmax layer:

𝒚^ij=softmax(𝗪𝒛ij+𝒃),\hat{\boldsymbol{\mathit{y}}}_{ij}=\operatorname{softmax}(\boldsymbol{\mathsf{W}}\boldsymbol{\mathit{z}}_{ij}+\boldsymbol{\mathit{b}}), (6)

where 𝗪c×dz\boldsymbol{\mathsf{W}}\in\mathbb{R}^{c\times d_{z}} and 𝒃c\boldsymbol{\mathit{b}}\in\mathbb{R}^{c} are learnable parameters, and 𝒚^ijc\hat{\boldsymbol{\mathit{y}}}_{ij}\in\mathbb{R}^{c} is the vector of predicted probabilities over entity types. Note that Eq. (6) follows the form of a typical neural classification head, which receives a single vector 𝒛ij\boldsymbol{\mathit{z}}_{ij}, and yields the predicted probabilities 𝒚^ij\hat{\boldsymbol{\mathit{y}}}_{ij}. Here, the pre-softmax vector 𝗪𝒛ij\boldsymbol{\mathsf{W}}\boldsymbol{\mathit{z}}_{ij} is called logits, and 𝒛ij\boldsymbol{\mathit{z}}_{ij} is called pre-logit representation (Müller et al., 2019).

Given the one-hot encoded ground truth 𝒚ijc\boldsymbol{\mathit{y}}_{ij}\in\mathbb{R}^{c}, the model could be trained by optimizing the cross entropy loss for all spans:

=0i<jT𝒚ij𝖳log(𝒚^ij).\mathcal{L}=-\sum_{0\leq i<j\leq T}\boldsymbol{\mathit{y}}_{ij}^{\mathsf{T}}\log(\hat{\boldsymbol{\mathit{y}}}_{ij}). (7)

We additionally apply the boundary smoothing technique (Zhu and Li, 2022), which is a variant of label smoothing (Szegedy et al., 2016) for span-based NER and brings performance improvements.

4 Experiments

4.1 Experimental Settings

Datasets.

We perform experiments on four English nested NER datasets: ACE 2004333https://catalog.ldc.upenn.edu/LDC2005T09., ACE 2005444https://catalog.ldc.upenn.edu/LDC2006T06., GENIA (Kim et al., 2003) and KBP 2017 (Ji et al., 2017); and two English flat NER datasets: CoNLL 2003 (Tjong Kim Sang and De Meulder, 2003) and OntoNotes 5555https://catalog.ldc.upenn.edu/LDC2013T19.. More details on data processing and descriptive statistics are reported in Appendix B.

Implementation Details.

To save space, our implementation details are all placed in Appendix C.

4.2 Main Results

Table 1 shows the evaluation results on English nested NER benchmarks. For a fair and reliable comparison to previous SOTA NER systems,666We exclude previous systems relying on extra training data (e.g., Li et al., 2020c), external resources (e.g., Yamada et al., 2020), extremely large PLMs (e.g., Yuan et al., 2022), or neural architecture search (e.g., Wang et al., 2021). we run DSpERT for five times on each dataset, and report both the best score and the average score with corresponding standard deviation.

With a base-sized PLM, DSpERT achieves on-par or better results compared with previous SOTA systems. More specifically, the best F1F_{1} scores are 88.31%, 87.42%, 81.90% and 87.65% on ACE 2004, ACE 2005, GENIA and KBP 2017, respectively. Except for ACE 2005, these scores correspond to 0.17%, 0.13% and 3.15% absolute improvements.

ACE 2004
Model Prec. Rec. F1
Li et al. (2020b) 85.05 86.32 85.98
Yu et al. (2020) 87.3 86.0 86.7
Yan et al. (2021) 87.27 86.41 86.84
Shen et al. (2021) 87.44 87.38 87.41
Li et al. (2022) 87.33 87.71 87.52
Zhu and Li (2022) 88.43 87.53 87.98
Shen et al. (2022) 88.48 87.81 88.14
DSpERT† 88.29 88.32 88.31
DSpERT‡ 87.90 88.21 88.05±0.18
ACE 2005
Model Prec. Rec. F1
Li et al. (2020b) 87.16 86.59 86.88
Yu et al. (2020) 85.2 85.6 85.4
Yan et al. (2021) 83.16 86.38 84.74
Shen et al. (2021) 86.09 87.27 86.67
Li et al. (2022) 85.03 88.62 86.79
Zhu and Li (2022) 86.25 88.07 87.15
Shen et al. (2022) 86.27 88.60 87.42
DSpERT† 87.01 87.84 87.42
DSpERT‡ 85.73 88.19 86.93±0.49
GENIA
Model Prec. Rec. F1
Yu et al. (2020)* 81.8 79.3 80.5
Yan et al. (2021) 78.87 79.6 79.23
Shen et al. (2021)* 80.19 80.89 80.54
Li et al. (2022) 83.10 79.76 81.39
Shen et al. (2022)* 83.24 80.35 81.77
DSpERT† 82.31 81.49 81.90
DSpERT‡ 81.72 81.21 81.46±0.25
KBP 2017
Model Prec. Rec. F1
Li et al. (2020b) 82.33 77.61 80.97
Shen et al. (2021) 85.46 82.67 84.05
Shen et al. (2022) 85.67 83.37 84.50
DSpERT† 87.37 87.93 87.65
DSpERT‡ 87.00 87.33 87.16±0.50
Table 1: Results of English nested entity recognition. * means that the model is trained with both the training and development splits. † means the best score; ‡ means the average score of multiple independent runs; the subscript number is the corresponding standard deviation.

Table 2 presents the results on English flat NER datasets. The best F1F_{1} scores are 93.70% and 91.76% on CoNLL 2003 and OntoNotes 5, respectively. These scores are slightly higher than those reported by previous literature.

Appendix D further lists the category-wise F1F_{1} scores; the results show that DSpERT can consistently outperform the biaffine model, a classic and strong baseline, across most entity categories. Appendix E provides additional experimental results on Chinese NER, suggesting that the effectiveness of DSpERT is generalizable across languages.

Overall, DSpERT shows strong and competitive performance on both the nested and flat NER tasks. Given the long-term extensive investigation and experiments on these datasets by the NLP community, the seemingly marginal performance improvements are still notable.

CoNLL 2003
Model Prec. Rec. F1
Peters et al. (2018) 92.22±0.10
Devlin et al. (2019) 92.8
Li et al. (2020b) 92.33 94.61 93.04
Yu et al. (2020)* 93.7 93.3 93.5
Yan et al. (2021)* 92.61 93.87 93.24
Li et al. (2022) 92.71 93.44 93.07
Zhu and Li (2022) 93.61 93.68 93.65
Shen et al. (2022)* 93.29 92.46 92.87
DSpERT† 93.48 93.93 93.70
DSpERT‡ 93.39 93.88 93.64±0.06
OntoNotes 5
Model Prec. Rec. F1
Li et al. (2020b) 92.98 89.95 91.11
Yu et al. (2020) 91.1 91.5 91.3
Yan et al. (2021) 89.99 90.77 90.38
Li et al. (2022) 90.03 90.97 90.50
Zhu and Li (2022) 91.75 91.74 91.74
Shen et al. (2022) 91.43 90.73 90.96
DSpERT† 91.46 92.05 91.76
DSpERT‡ 90.87 91.25 91.06±0.26
Table 2: Results of English flat entity recognition. * means that the model is trained with both the training and development splits. † means the best score; ‡ means the average score of multiple independent runs; the subscript number is the corresponding standard deviation.

4.3 Ablation Studies

We perform ablation studies on three datasets, i.e., ACE 2004, GENIA and CoNLL 2003, covering flat and nested, common and domain-specific corpora.

Depth of Span Representations.

As previously highlighted, our core argument is that the deep span representations, which are computed throughout the span Transformer blocks, embrace deep and rich semantics and thus outperform the shallow counterparts.

To validate this point, Table 3 compares DSpERT to the models with a shallow setting, where the span representations are aggregated from the top token representations by max-pooling, mean-pooling, multiplicative attention or additive attention (See Appendix A for details). All the models are trained with the same recipe used in our main experiments. It shows that the 12-layer deep span representations achieve higher performance than its shallow counterparts equipped with any potential aggregating function, across all datasets.

We further run DSpERT with L~\tilde{L} (L~<L\tilde{L}<L) span Transformer blocks, where the initial aggregation happens at the (LL~)(L-\tilde{L})-th layer and the span Transformer corresponds to the top/last L~\tilde{L} Transformer blocks. These models may be thought of as intermediate configurations between fully deep span representations and fully shallow ones. As displayed in Table 3, the F1F_{1} score in general experiences a monotonically increasing trend when depth L~\tilde{L} increases from 2 to 12; this pattern holds for all three datasets. These results further strengthen our argument that the depth positively contributes to the quality of span representations.

Appendix F provides extensive ablation studies evaluating other components.

Depth ACE04 GENIA CoNLL03
Shallow agg. (L~\tilde{L} = 0)
   w/ max-pooling 82.22±0.64 79.44±0.20 92.99±0.32
   w/ mean-pooling 80.90±0.28 73.83±0.42 91.97±0.14
   w/ mul. attention 84.38±0.58 76.54±2.70 93.21±0.16
   w/ add. attention 83.73±0.52 76.23±3.27 93.05±0.04
DSpERT
   L~\tilde{L} = 2 87.87±0.13 80.66±0.36 93.30±0.09
   L~\tilde{L} = 4 87.88±0.41 80.88±0.39 93.38±0.08
   L~\tilde{L} = 6 87.81±0.13 81.01±0.22 93.40±0.13
   L~\tilde{L} = 8 88.00±0.22 81.13±0.25 93.48±0.08
   L~\tilde{L} = 10 88.00±0.21 81.12±0.14 93.51±0.10
   L~\tilde{L} = 12 88.05±0.18 81.46±0.25 93.64±0.06
Table 3: The effect of depth. The underlined specification is the one used in our main experiments. All the results are average scores of five independent runs, with subscript standard deviations.

4.4 Effect on Long-Span Entities

The recognition of long-span entities is a long-tail and challenging problem. Taking ACE 2004 as an example, the ground-truth entities longer than 10 tokens only account for 2.8%, and the maximum length reaches 57. Empirical evidence also illustrates that existing NER models show relatively weak performance on long entities (e.g., Shen et al., 2021; Yuan et al., 2022).

Figure 2 presents the F1F_{1} scores grouped by different span lengths. In general, the models based on shallow span representations perform relatively well on the short entities, but struggle for the long ones. However, DSpERT show much higher F1F_{1} scores on the long entities, without any performance sacrifice on the short ones. For ACE 2004, DSpERT outperforms its shallow counterpart by 2%–12% absolute F1F_{1} score on spans shorter than 10, while this difference exceeds 30% for spans longer than 10. Similar patterns are observed on GENIA and CoNLL 2003.

Conceptually, a longer span contains more information, so it would be more difficult to be encoded into a fixed-length vector, i.e., the span representation. According to our experimental results, the shallow aggregation fails to fully preserve the semantics in the original token representations, especially for long spans. The DSpERT, however, allows complicated interactions between tokens through multiple layers; in particular, longer spans experience more interactions. This mechanism amplifies the performance gain on long entities.

4.5 Effect on Nested Structures

Even in nested NER datasets, the nested entities are less than the flat ones (See Table 6). To deliberately investigate the performance on nested entities, we look into two subsets of spans that are directly related to nested structures: (1) Nested: the spans that are nested inside a ground-truth entity which covers other ground-truth entities; (2) Covering: the spans that cover a ground-truth entity which is nested inside other ground-truth entities.

For example, in sentence “Mr. John Smith graduated from New York University last year”, a location entity “New York” is nested in an organization entity “New York University”. A model is regarded to well handle nested structures if it can: (1) distinguish “New York” from other negative spans inside the outer entity “New York University”, i.e., those in the Nested subset; and (2) distinguish “New York University” from other negative spans covering the inner entity “New York”, i.e., those in the Covering subset.

Figure 3 depicts the F1F_{1} scores grouped by different nested structures. Consistent to a common expectation, nested structures create significant difficulties for entity recognition. Compared to the flat ones, a shallow span-based model encounters a substantial performance degradation on the nested structures, especially on the spans in both the Nested and Covering subsets. On the other hand, our deep span representations perform much better. For ACE 2004, DSpERT presents 2% higher absolute F1F_{1} score than the shallow model on flat spans, but achieves about 40% higher score on spans in both the Nested and Covering subsets. The experiments on GENIA report similar results, although the difference in performance gain becomes less substantial.

As previously emphasized, shallowly aggregated span representations are technically coupled if the spans overlap. This explains why such models perform poorly on nested structures. Our model addresses this issue by deep and multi-layered construction of span representations. Implied by the experimental results, deep span representations are less coupled and more easily separable for overlapping spans.

Refer to caption
(a) ACE 2004
Refer to caption
(b) GENIA
Refer to caption
(c) CoNLL 2003
Figure 2: F1F_{1} scores on spans of different lengths. All the results are average scores of five independent runs.
Refer to caption
(a) ACE 2004
Refer to caption
(b) GENIA
Figure 3: F1F_{1} scores on spans with different nested structures. “Nested” means the spans that are nested inside a ground-truth entity which covers other ground-truth entities; “Covering” means the spans that cover a ground-truth entity which is nested inside other ground-truth entities; “Both” means the spans that are both “Nested” and “Covering”; “Flat” means the spans that are neither “Nested” nor “Covering”. All the results are average scores of five independent runs.

5 Analysis of Pre-Logit Representations

For a neural classification model, the logits only relate to the pre-defined classification categories, while the pre-logit representations contain much richer information (Krizhevsky et al., 2012; Wu et al., 2018). Hence, the analysis of pre-logit representations has become a popular tool in machine learning research (Van der Maaten and Hinton, 2008; Müller et al., 2019; Chen et al., 2020).

However, such analysis is incompatible with most neural NER systems except for the span-based ones. Our DSpERT employs a standard classification head, which exposes the pre-logit representations and thus allows the analysis. In this section, we investigate the pre-logit span representations, i.e., 𝒛ij\boldsymbol{\mathit{z}}_{ij} in Eq. (6), providing more insights into why the deep span representations outperform the shallow counterparts.

5.1 Decoupling Effect on Overlapping Spans

The effectiveness of DSpERT on nested structures primarily attributes to its decoupling effect on the representations between overlapping spans. To support this argument, we compare the coupling strengths between span representations with different overlapping levels.

Specifically, we define the overlapping ratio α\alpha of two given spans as the proportion of the shared tokens in the spans, and then categorize the span pairs into three scenarios: non-overlapping (α=0\alpha=0), weakly overlapping (0<α0.50<\alpha\leq 0.5) and strongly overlapping (0.5<α<10.5<\alpha<1). Table 4 reports the cosine similarities of the representations between entities and their neighboring spans, categorized by overlapping ratios. In general, DSpERT and the shallow model have comparable similarity values on the non-overlapping spans; and DSpERT show slightly higher values on the overlapping spans. However, the shallow model produces significantly higher similarities for stronger overlapping levels. Hence, shallow models yield coupled representations for overlapping spans, while DSpERT can effectively decouple the representations and thus lead to the performance improvement, in particular on the nested entities.

ACE 2004
Shallow Deep Δ\Delta
Non-overlapping 0.45±0.05 0.40±0.04 -0.05
Weakly overlapping 0.65±0.02 0.41±0.04 -0.24
Strongly overlapping 0.76±0.01 0.44±0.04 -0.32
GENIA
Shallow Deep Δ\Delta
Non-overlapping 0.39±0.02 0.35±0.03 -0.04
Weakly overlapping 0.59±0.01 0.38±0.03 -0.21
Strongly overlapping 0.72±0.00 0.42±0.04 -0.30
CoNLL 2003
Shallow Deep Δ\Delta
Non-overlapping 0.31±0.02 0.40±0.02 +0.09
Weakly overlapping 0.58±0.01 0.46±0.02 -0.12
Strongly overlapping 0.69±0.01 0.52±0.02 -0.17
Table 4: Cosine similarities of the representations between entity spans and their neighboring spans. Non-/weakly/strongly overlapping means that the overlapping ratio is 0/0–0.5/0.5–1, respectively. All the metrics are first averaged within each experiment, and then averaged over five independent experiments, reported with subscript standard deviations.

5.2 2\ell_{2}-Norm and Cosine Similarity

We calculate the 2\ell_{2}-norm and cosine similarity of the span representations. As presented in Table 5, the deep span representations have larger 2\ell_{2}-norm than those of the shallow counterpart. Although the variance of representations inevitably shrinks during the aggregating process from the perspective of statistics, this result implies that deep span representations are less restricted and thus able to flexibly represent rich semantics. In addition, the deep span representations are associated with higher within-class similarities and lower between-class similarities, suggesting that the representations are more tightly clustered within each category, but more separable between categories. Apparently, this nature contributes to the high classification performance.

We further investigate the pre-logit weight, i.e., 𝗪\boldsymbol{\mathsf{W}} in Eq. (6). First, the trained DSpERT has a pre-logit weight with a smaller 2\ell_{2}-norm. According to a common understanding of neural networks, smaller norm implies that the model is simpler and thus more generalizable.

Second, as indicated by Müller et al. (2019), a typical neural classification head can be regarded as a template-matching mechanism, where each row vector of 𝗪\boldsymbol{\mathsf{W}} is a template.777Conceptually, if vector 𝒛ij\boldsymbol{\mathit{z}}_{ij} is most matched/correlated with the kk-th row vector of 𝗪\boldsymbol{\mathsf{W}}, then the kk-th logit will be most activated. Refer to Müller et al. (2019) for more details. Under this interpretation, each template “stands for” the overall direction of the span representations of the corresponding category in the feature space. As shown in Table 5, the absolute cosine values between the templates of DSpERT are fairly small. In other words, the templates are approximately orthogonal, which suggests that different entity categories are uncorrelated and separately occupy distinctive subareas in the feature space. This pattern, however, is not present for the shallow models.

ACE 2004 GENIA CoNLL 2003
Shallow Deep Shallow Deep Shallow Deep
Pre-logit representations
   2\ell_{2}-norm 9.58±0.17 14.55±0.33 9.89±0.18 11.69±0.52 9.08±0.17 13.49±0.50
   Cosine within pos class 0.70±0.02   0.80±0.01 0.78±0.01   0.83±0.01 0.79±0.01   0.88±0.00
   Cosine between pos vs. pos 0.45±0.02   0.37±0.01 0.54±0.02   0.29±0.01 0.35±0.02   0.32±0.01
   Cosine between pos vs. neg 0.48±0.03   0.35±0.03 0.55±0.01   0.35±0.02 0.39±0.01   0.37±0.02
Pre-logit weight/templates 𝗪\boldsymbol{\mathsf{W}}
   2\ell_{2}-norm 1.93±0.28   1.57±0.03 1.84±0.08   1.39±0.02 1.71±0.04   1.39±0.01
   Abs cosine 0.17±0.12   0.10±0.06 0.31±0.11   0.10±0.10 0.23±0.08   0.05±0.03
   Abs cosine between pos vs. pos 0.13±0.10   0.10±0.07 0.27±0.11   0.10±0.11 0.17±0.04   0.05±0.04
   Abs cosine between pos vs. neg 0.31±0.06   0.10±0.05 0.40±0.06   0.09±0.04 0.32±0.04   0.05±0.03
Table 5: 2\ell_{2}-norm and cosine similarity of pre-logit representations and templates. “pos” means the positive types (i.e., entity types); “neg” means the negative type (i.e., non-entity type). / indicates that DSpERT presents a metric higher/lower than its shallow counterpart. All the metrics are first averaged within each experiment, and then averaged over five independent experiments, reported with subscript standard deviations.
Refer to caption
(a) ACE 2004, Shallow
Refer to caption
(b) GENIA, Shallow
Refer to caption
(c) CoNLL 2003, Shallow
Refer to caption
(d) ACE 2004, Deep
Refer to caption
(e) GENIA, Deep
Refer to caption
(f) CoNLL 2003, Deep
Figure 4: PCA visualization of pre-logit span representations of entities in the testing set.

5.3 Visualization

Figure 4 visualizes the span representations dimension-reduced by principal component analysis (PCA). The results are quite impressive. The representations by shallow aggregation are scattered over the plane. Although they are largely clustered by categories, the boundaries are mixed and ambiguous. In contrast, the deep span representations group by relatively clear and tight clusters, corresponding to the ground-truth categories. Except for some outliers, each pair of the clusters can be easily separable in this projected plane. This is also consistent to the aforementioned finding that deep span representations have high within-class similarities but low between-class similarities. As a linear dimensionality reduction, PCA results indicate whether and how the features are linearly separable. Note that the pre-logit representations are the last ones before the logits, so the linear separability is crucial to the classification performance.

6 Discussion and Conclusion

Neural NLP models have been rigidly adhere to the paradigm where an encoder produces token-level representations, and a task-specific decoder receives these representations, computes the loss and yields the outputs (Collobert et al., 2011). This design works well on most, if not all, NLP tasks; and it may also deserve a credit for facilitating NLP pretraining (Peters et al., 2018; Devlin et al., 2019), since such common structure in advance bridges the pretraining and downstream phases.

However, this paradigm may be sub-optimal for specific tasks. In span-based NER (or information extraction from a broader perspective), the smallest modeling unit should be spans instead of tokens; and thus the span representations should be crucial. This originally motivates our DSpERT. In addition, DSpERT also successfully shows how to exploit the pretrained weights beyond the original Transformer structure, i.e., adapting the weights from computing token representations for span representations. We believe that adding span representation learning in the pretraining stage will further contribute positively.

In conclusion, deep and span-specific representations can significantly boost span-based neural NER models. Our DSpERT achieves SOTA results on six well-known NER benchmarks; the model presents pronounced effect on long-span entities and nested structures. Further analysis shows that the resulting deep span representations are well structured and easily separable in the feature space.

7 Limitations

To some extent, DSpERT pursues performance and interpretability over computational efficiency. The major computational cost of a Transformer encoder is on the multihead attention module and FFN. For a TT-length input and a dd-dimensional Transformer encoder, the per-layer complexities of the multihead attention and FFN are of order O(T2d)O(T^{2}d) and O(Td2)O(Td^{2}), respectively. When the maximum span size KTK\ll T, our span Transformer brings additional O(K2Td)O(K^{2}Td) complexity on the attention module, and O(KTd2)O(KTd^{2}) complexity on the FFN. Empirically, training a DSpERT consumes about five times the time for a shallow model of a same scale. However, this issue can be mitigated if we use fewer layers for the span Transformer (Subsection 4.3).

As noted, we empirically choose the maximum span size KK such that it covers most entities in the training and development splits. From the perspective of F1F_{1} score, this heuristic works well, and DSpERT performs favourably on long-span entities as long as they are covered. However, the entities with extreme lengths beyond KK will be theoretically irretrievable.

Acknowledgements

This work is supported by Zhejiang Provincial Natural Science Foundation of China (No. LQ23F020005), National Natural Science Foundation of China (No. 62106248). We thank the anonymous reviewers for their insightful and constructive comments.

References

  • Bengio et al. (2013) Yoshua Bengio, Aaron Courville, and Pascal Vincent. 2013. Representation learning: A review and new perspectives. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(8):1798–1828.
  • Chen et al. (2020) Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. 2020. A simple framework for contrastive learning of visual representations. In International Conference on Machine Learning, pages 1597–1607. PMLR.
  • Chiu and Nichols (2016) Jason P.C. Chiu and Eric Nichols. 2016. Named entity recognition with bidirectional LSTM-CNNs. Transactions of the Association for Computational Linguistics, 4:357–370.
  • Collobert et al. (2011) Ronan Collobert, Jason Weston, Léon Bottou, Michael Karlen, Koray Kavukcuoglu, and Pavel Kuksa. 2011. Natural language processing (almost) from scratch. Journal of Machine Learning Research, 12(ARTICLE):2493–2537.
  • Devlin et al. (2019) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186, Minneapolis, Minnesota. Association for Computational Linguistics.
  • Dozat and Manning (2017) Timothy Dozat and Christopher D Manning. 2017. Deep biaffine attention for neural dependency parsing. In Proceedings of 5th International Conference on Learning Representations. OpenReview.net.
  • Eberts and Ulges (2020) Markus Eberts and Adrian Ulges. 2020. Span-based joint entity and relation extraction with Transformer pre-training. In Proceedings of the 24th European Conference on Artificial Intelligence, Santiago de Compostela, Spain.
  • Hochreiter and Schmidhuber (1997) Sepp Hochreiter and Jürgen Schmidhuber. 1997. Long short-term memory. Neural Computation, 9(8):1735–1780.
  • Huang et al. (2015) Zhiheng Huang, Wei Xu, and Kai Yu. 2015. Bidirectional lstm-crf models for sequence tagging. arXiv preprint arXiv:1508.01991.
  • Ji et al. (2017) Heng Ji, Xiaoman Pan, Boliang Zhang, Joel Nothman, James Mayfield, Paul McNamee, and Cash Costello. 2017. Overview of TAC-KBP2017 13 languages entity discovery and linking. In Proceedings of the 2017 Text Analysis Conference, Gaithersburg, Maryland, USA. NIST.
  • Ju et al. (2018) Meizhi Ju, Makoto Miwa, and Sophia Ananiadou. 2018. A neural layered model for nested named entity recognition. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 1446–1459, New Orleans, Louisiana. Association for Computational Linguistics.
  • Katiyar and Cardie (2018) Arzoo Katiyar and Claire Cardie. 2018. Nested named entity recognition revisited. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 861–871, New Orleans, Louisiana. Association for Computational Linguistics.
  • Kim et al. (2003) J-D Kim, Tomoko Ohta, Yuka Tateisi, and Jun’ichi Tsujii. 2003. GENIA corpus – a semantically annotated corpus for bio-textmining. Bioinformatics, 19(suppl_1):i180–i182.
  • Krizhevsky et al. (2012) Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. 2012. Imagenet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems, volume 25.
  • Lample et al. (2016) Guillaume Lample, Miguel Ballesteros, Sandeep Subramanian, Kazuya Kawakami, and Chris Dyer. 2016. Neural architectures for named entity recognition. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 260–270, San Diego, California. Association for Computational Linguistics.
  • Li et al. (2022) Jingye Li, Hao Fei, Jiang Liu, Shengqiong Wu, Meishan Zhang, Chong Teng, Donghong Ji, and Fei Li. 2022. Unified named entity recognition as word-word relation classification. In Proceedings of the AAAI Conference on Artificial Intelligence.
  • Li et al. (2020a) Xiaonan Li, Hang Yan, Xipeng Qiu, and Xuanjing Huang. 2020a. FLAT: Chinese NER using flat-lattice transformer. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 6836–6842, Online. Association for Computational Linguistics.
  • Li et al. (2020b) Xiaoya Li, Jingrong Feng, Yuxian Meng, Qinghong Han, Fei Wu, and Jiwei Li. 2020b. A unified MRC framework for named entity recognition. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 5849–5859, Online. Association for Computational Linguistics.
  • Li et al. (2020c) Xiaoya Li, Xiaofei Sun, Yuxian Meng, Junjun Liang, Fei Wu, and Jiwei Li. 2020c. Dice loss for data-imbalanced NLP tasks. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 465–476, Online. Association for Computational Linguistics.
  • Li et al. (2020d) Yangming Li, Shuming Shi, et al. 2020d. Empirical analysis of unlabeled entity problem in named entity recognition. In International Conference on Learning Representations.
  • Lin et al. (2019) Hongyu Lin, Yaojie Lu, Xianpei Han, and Le Sun. 2019. Sequence-to-nuggets: Nested entity mention detection via anchor-region networks. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 5182–5192, Florence, Italy. Association for Computational Linguistics.
  • Liu et al. (2019) Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. RoBERTa: A robustly optimized BERT pretraining approach. arXiv preprint arXiv:1907.11692.
  • Loshchilov and Hutter (2018) Ilya Loshchilov and Frank Hutter. 2018. Decoupled weight decay regularization. In International Conference on Learning Representations.
  • Lu and Roth (2015) Wei Lu and Dan Roth. 2015. Joint mention extraction and classification with mention hypergraphs. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pages 857–867, Lisbon, Portugal. Association for Computational Linguistics.
  • Ma et al. (2020) Ruotian Ma, Minlong Peng, Qi Zhang, Zhongyu Wei, and Xuanjing Huang. 2020. Simplify the usage of lexicon in Chinese NER. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 5951–5960, Online. Association for Computational Linguistics.
  • Ma and Hovy (2016) Xuezhe Ma and Eduard Hovy. 2016. End-to-end sequence labeling via bi-directional LSTM-CNNs-CRF. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1064–1074, Berlin, Germany. Association for Computational Linguistics.
  • Müller et al. (2019) Rafael Müller, Simon Kornblith, and Geoffrey E Hinton. 2019. When does label smoothing help? In Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc.
  • Ouchi et al. (2020) Hiroki Ouchi, Jun Suzuki, Sosuke Kobayashi, Sho Yokoi, Tatsuki Kuribayashi, Ryuto Konno, and Kentaro Inui. 2020. Instance-based learning of span representations: A case study through named entity recognition. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 6452–6459, Online. Association for Computational Linguistics.
  • Pascanu et al. (2013) Razvan Pascanu, Tomas Mikolov, and Yoshua Bengio. 2013. On the difficulty of training recurrent neural networks. In International Conference on Machine Learning, pages 1310–1318. PMLR.
  • Peng and Dredze (2015) Nanyun Peng and Mark Dredze. 2015. Named entity recognition for Chinese social media with jointly trained embeddings. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pages 548–554, Lisbon, Portugal. Association for Computational Linguistics.
  • Peters et al. (2018) Matthew E. Peters, Mark Neumann, Mohit Iyyer, Matt Gardner, Christopher Clark, Kenton Lee, and Luke Zettlemoyer. 2018. Deep contextualized word representations. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 2227–2237, New Orleans, Louisiana. Association for Computational Linguistics.
  • Pradhan et al. (2013) Sameer Pradhan, Alessandro Moschitti, Nianwen Xue, Hwee Tou Ng, Anders Björkelund, Olga Uryupina, Yuchen Zhang, and Zhi Zhong. 2013. Towards robust linguistic analysis using OntoNotes. In Proceedings of the Seventeenth Conference on Computational Natural Language Learning, pages 143–152, Sofia, Bulgaria. Association for Computational Linguistics.
  • Shen et al. (2021) Yongliang Shen, Xinyin Ma, Zeqi Tan, Shuai Zhang, Wen Wang, and Weiming Lu. 2021. Locate and label: A two-stage identifier for nested named entity recognition. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 2782–2794, Online. Association for Computational Linguistics.
  • Shen et al. (2022) Yongliang Shen, Xiaobin Wang, Zeqi Tan, Guangwei Xu, Pengjun Xie, Fei Huang, Weiming Lu, and Yueting Zhuang. 2022. Parallel instance query network for named entity recognition. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 947–961, Dublin, Ireland. Association for Computational Linguistics.
  • Sohrab and Miwa (2018) Mohammad Golam Sohrab and Makoto Miwa. 2018. Deep exhaustive model for nested named entity recognition. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 2843–2849, Brussels, Belgium. Association for Computational Linguistics.
  • Szegedy et al. (2016) Christian Szegedy, Vincent Vanhoucke, Sergey Ioffe, Jon Shlens, and Zbigniew Wojna. 2016. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 2818–2826.
  • Tjong Kim Sang and De Meulder (2003) Erik F. Tjong Kim Sang and Fien De Meulder. 2003. Introduction to the CoNLL-2003 shared task: Language-independent named entity recognition. In Proceedings of the Seventh Conference on Natural Language Learning at HLT-NAACL 2003, pages 142–147.
  • Van der Maaten and Hinton (2008) Laurens Van der Maaten and Geoffrey Hinton. 2008. Visualizing data using t-SNE. Journal of Machine Learning Research, 9(11).
  • Vaswani et al. (2017) Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Advances in Neural Information Processing Systems, pages 5998–6008.
  • Wang et al. (2021) Xinyu Wang, Yong Jiang, Nguyen Bach, Tao Wang, Zhongqiang Huang, Fei Huang, and Kewei Tu. 2021. Automated concatenation of embeddings for structured prediction. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 2643–2660, Online. Association for Computational Linguistics.
  • Wu et al. (2021) Shuang Wu, Xiaoning Song, and Zhenhua Feng. 2021. MECT: Multi-metadata embedding based cross-transformer for Chinese named entity recognition. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 1529–1539, Online. Association for Computational Linguistics.
  • Wu et al. (2018) Zhirong Wu, Yuanjun Xiong, Stella X Yu, and Dahua Lin. 2018. Unsupervised feature learning via non-parametric instance discrimination. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 3733–3742.
  • Yamada et al. (2020) Ikuya Yamada, Akari Asai, Hiroyuki Shindo, Hideaki Takeda, and Yuji Matsumoto. 2020. LUKE: Deep contextualized entity representations with entity-aware self-attention. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 6442–6454, Online. Association for Computational Linguistics.
  • Yan et al. (2021) Hang Yan, Tao Gui, Junqi Dai, Qipeng Guo, Zheng Zhang, and Xipeng Qiu. 2021. A unified generative framework for various NER subtasks. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 5808–5822, Online. Association for Computational Linguistics.
  • Yu et al. (2020) Juntao Yu, Bernd Bohnet, and Massimo Poesio. 2020. Named entity recognition as dependency parsing. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 6470–6476, Online. Association for Computational Linguistics.
  • Yuan et al. (2022) Zheng Yuan, Chuanqi Tan, Songfang Huang, and Fei Huang. 2022. Fusing heterogeneous factors with triaffine mechanism for nested named entity recognition. In Findings of the Association for Computational Linguistics: ACL 2022, pages 3174–3186, Dublin, Ireland. Association for Computational Linguistics.
  • Zhang and Yang (2018) Yue Zhang and Jie Yang. 2018. Chinese NER using lattice LSTM. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1554–1564, Melbourne, Australia. Association for Computational Linguistics.
  • Zhu and Li (2022) Enwei Zhu and Jinpeng Li. 2022. Boundary smoothing for named entity recognition. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 7096–7108, Dublin, Ireland. Association for Computational Linguistics.

Appendix A (Shallowly) Aggregating Functions

Given token representations 𝗛T×d\boldsymbol{\mathsf{H}}\in\mathbb{R}^{T\times d}, a model can shallowly aggregate them in corresponding positions to construct span representations. Formally, the span representation of (i,j)(i,j) can be built by:

Max-pooling.

Applying max-pooling to 𝗛[i:j]\boldsymbol{\mathsf{H}}_{[i:j]} over the first dimension.

Mean-pooling.

Applying mean-pooling to 𝗛[i:j]\boldsymbol{\mathsf{H}}_{[i:j]} over the first dimension.

Multiplicative Attention.

Computing

softmax(𝒖𝖳tanh(𝗪𝗛[i:j]𝖳))𝗛[i:j],\operatorname{softmax}\left(\boldsymbol{\mathit{u}}^{\mathsf{T}}\tanh\left(\boldsymbol{\mathsf{W}}\boldsymbol{\mathsf{H}}_{[i:j]}^{\mathsf{T}}\right)\right)\boldsymbol{\mathsf{H}}_{[i:j]},

where 𝗪\boldsymbol{\mathsf{W}} and 𝒖\boldsymbol{\mathit{u}} are learnable parameters.

Additive Attention.

Computing

softmax(𝒖𝖳tanh(𝗪(𝗛[i:j]𝒗)𝖳))𝗛[i:j],\operatorname{softmax}\left(\boldsymbol{\mathit{u}}^{\mathsf{T}}\tanh\left(\boldsymbol{\mathsf{W}}\left(\boldsymbol{\mathsf{H}}_{[i:j]}\oplus\boldsymbol{\mathit{v}}\right)^{\mathsf{T}}\right)\right)\boldsymbol{\mathsf{H}}_{[i:j]},

where 𝗪\boldsymbol{\mathsf{W}}, 𝒖\boldsymbol{\mathit{u}} and 𝒗\boldsymbol{\mathit{v}} are learnable parameters; \oplus means concatenation over the second dimension, and vector 𝒗\boldsymbol{\mathit{v}} should be repeated for jij-i times before the concatenation.

In general, either multiplicative or additive attention computes normalized weights over the sequence dimension of span (i,j)(i,j), where the weights are dependent on the values of 𝗛[i:j]\boldsymbol{\mathsf{H}}_{[i:j]}; and then applies the weights to 𝗛[i:j]\boldsymbol{\mathsf{H}}_{[i:j]}, resulting in weighted average values.

Appendix B Datasets

ACE04 ACE05 GENIA KBP17 CoNLL03 OntoNotes 5
#Sent. All 8,507 9,341 18,546 15,358 22,137 76,714
Train 6,799 7,336 15,023 10,546 14,987 59,924
Dev. 829 958 1,669 545 3,466 8,528
Test 879 1,047 1,854 4,267 3,684 8,262
#Type 7 7 5 5 4 18
#Token 173,796 176,575 471,264 300,345 302,811 1,388,955
#Entity 27,749 30,931 56,046 45,714 35,089 104,151
#Nested 7,832 7,460 5,431 7,479
Table 6: Descriptive statistics of datasets. #Sent. denotes the number of sentences; #Type denotes the number of entity types; #Token denotes the number of tokens; #Entity denotes the number of entities; #Nested denotes the number of entities that are nested in other entities.
ACE04 ACE05 GENIA KBP17 CoNLL03 OntoNotes 5
PLM RoBERTa-base
Maximum span size 25 25 18 16 10 16
Initial aggregation Max Max Max Max Max Max
LSTM hidden size 400 400 400 400 400 400
LSTM layers 1 1 1 1 1 1
FFN hidden size 300 300 300 300 300 300
FFN layers 1 1 1 1 1 1
Boundary smoothing ε\varepsilon 0.1 0.1 0.1 0.1 0.1 0.1
Number of epochs 50 50 30 50 50 30
Learning rate
   Pretrained weights 2e-5 2e-5 4e-5 2e-5 2e-5 2e-5
   Other weights 2e-3 2e-3 2e-3 2e-3 2e-3 2e-3
Batch size 48 48 16 48 48 16
Number of parameters (M) 212.9 (w/o weight sharing); or 126.3 (w/ weight sharing)
Training time (hours on A6000) 13.9 15.0 8.3 16.2 3.3 15.3
Table 7: Hyperparameters and computational costs of main experiments.

ACE 2004 and ACE 2005

are two English nested NER datasets, either of which contains seven entity types, i.e., Person, Organization, Facility, Location, Geo-political Entity, Vehicle, Weapon. Our data processing and splits follow Lu and Roth (2015).

GENIA

(Kim et al., 2003) is a nested NER corpus on English biological articles. Our data processing follows Lu and Roth (2015), resulting in five entity types (DNA, RNA, Protein, Cell line, Cell type); data splits follow Yan et al. (2021) and Li et al. (2022).

KBP 2017

(Ji et al., 2017) is an English nested NER corpus including text from news, discussion forum, web blog, tweets and scientific literature. It contains five entity categories, i.e., Person, Geo-political Entity, Organization, Location, and Facility. Our data processing and splits follow Lin et al. (2019) and Shen et al. (2022).

CoNLL 2003

(Tjong Kim Sang and De Meulder, 2003) is an English flat NER benchmark with four entity types, i.e., Person, Organization, Location and Miscellaneous. We use the original data splits for experiments.

OntoNotes 5

is a large-scale English flat NER benchmark, which has 18 entity types. Our data processing and splits follow Pradhan et al. (2013).

Table 6 presents the descriptive statistics of the datasets.

Appendix C Implementation Details

Hyperparameters.

We choose RoBERTa (Liu et al., 2019) as the PLM to initialize the weights in the Transformer blocks and span Transformer blocks. The PLMs used in our main experiments are all of the base size (768 hidden size, 12 layers).

For span Transformer blocks, the maximum span size KK is specifically determined for each dataset. In general, a larger KK would improve the recall performance (entities longer than KK will never be recalled), but significantly increase the computation cost. We empirically choose KK such that it covers most entities in the training and development splits. For example, most entities are short in CoNLL 2003, so we use K=10K=10; while entities are relatively long in ACE 2004 and ACE 2005, so we use K=25K=25. We use max-pooling as the initial aggregating function.

We find it beneficial to additionally include a BiLSTM (Hochreiter and Schmidhuber, 1997) before passing the span representations to the entity classifier. The BiLSTM has one layer and 400 hidden states, with dropout rate of 0.5. In the entity classifier, the FFN has one layer and 300 hidden states, with dropout rate of 0.4, and the activation is ReLU (Krizhevsky et al., 2012). In addition, boundary smoothing (Zhu and Li, 2022) with ε=0.1\varepsilon=0.1 is applied to loss computation.

We train the models by the AdamW optimizer (Loshchilov and Hutter, 2018) for 50 epochs with the batch size of 48. Gradients are clipped at 2\ell_{2}-norm of 5 (Pascanu et al., 2013). The learning rates are 2e-5 and 2e-3 for pretrained weights and randomly initialized weights, respectively; a scheduler of linear warmup is applied in the first 20% steps followed by linear decay. For some datasets, a few hyperparameters are further tuned and thus slightly different from the above ones.

The experiments are run on NVIDIA RTX A6000 GPUs. More details on the hyperparameters and computational costs are reported in Table 7.

Evaluation.

An entity is considered correctly recognized if its predicted type and boundaries exactly match the ground truth.

The model checkpoint with the best F1F_{1} score throughout the training process on the development split is used for evaluation. The evaluation metrics are micro precision rate, recall rate and F1F_{1} score on the testing split. Unless otherwise noted, we run each experiment for five times and report the average metrics with corresponding standard deviations.

Appendix D Categorical Results

Table 8 lists the category-specific results on ACE 2004, GENIA and CoNLL 2003. As a strong baseline, the classic biaffine model (Yu et al., 2020) is re-implemented with PLM and hyperparameters consistent with our DSpERT; note that our re-implementation achieves higher performances than the scores reported in the original paper. The categorical results show that DSpERT outperforms the biaffine model across almost all the categories of the three datasets, except for Geo-political Entity and Vehicle from ACE 2004.

ACE 2004
Biaffine DSpERT Δ\Delta
PER 91.35±0.10 91.50±0.16 +0.14
ORG 82.49±0.54 83.59±0.33 +1.10
FAC 71.18±1.30 72.28±1.43 +1.10
LOC 73.38±1.55 75.22±0.86 +1.84
GPE 89.30±0.36 89.16±0.44 -0.15
VEH 87.18±2.66 84.35±3.58 -2.83
WEA 70.87±4.18 79.88±1.41 +9.00
Overall 87.64±0.19 88.05±0.18 +0.41
GENIA
Biaffine DSpERT Δ\Delta
DNA 77.50±0.44 77.75±0.56 +0.26
RNA 83.75±0.68 85.02±1.31 +1.27
Protein 84.00±0.36 84.57±0.26 +0.57
Cell Line 74.60±0.48 76.39±0.96 +1.79
Cell Type 75.89±0.46 76.10±0.48 +0.22
Overall 80.93±0.24 81.46±0.25 +0.53
CoNLL 2003
Biaffine DSpERT Δ\Delta
PER 96.74±0.23 96.94±0.11 +0.20
ORG 92.92±0.13 93.09±0.14 +0.17
LOC 94.83±0.18 94.96±0.13 +0.12
MISC 83.83±0.28 84.52±0.56 +0.69
Overall 93.37±0.10 93.64±0.06 +0.27
Table 8: Categorical F1F_{1} scores by biaffine model and DSpERT. All the results are average scores of five independent runs, with subscript standard deviations.

Appendix E Results on Chinese NER

Table 9 shows the experimental results on two Chinese flat NER datasets: Weibo NER (Peng and Dredze, 2015) and Resume NER (Zhang and Yang, 2018). DSpERT achieves 72.64% and 96.72% best F1F_{1} scores on the two benchmarks; they are also quite close to the recently reported SOTA results.

Weibo NER
Model Prec. Rec. F1
Ma et al. (2020) 70.50
Li et al. (2020a) 68.55
Wu et al. (2021) 70.43
Li et al. (2022) 70.84 73.87 72.32
Zhu and Li (2022) 70.16 75.36 72.66
DSpERT† 74.56 70.81 72.64
DSpERT‡ 71.09 71.58 71.30±0.86
Resume NER
Model Prec. Rec. F1
Ma et al. (2020) 96.08 96.13 96.11
Li et al. (2020a) 95.86
Wu et al. (2021) 95.98
Li et al. (2022) 96.96 96.35 96.65
Zhu and Li (2022) 96.63 96.69 96.66
DSpERT† 96.69 96.75 96.72
DSpERT‡ 96.44 96.58 96.51±0.17
Table 9: Results of Chinese flat entity recognition. † means the best score; ‡ means the average score of multiple independent runs; the subscript number is the corresponding standard deviation.

Appendix F Additional Ablation Studies

Weight Sharing.

As described, the span Transformer shares the same structure with the Transformer, but their weights are independent and separately initialized from the PLM. A straightforward idea is to tie the corresponding weights between these two modules, which can reduce the model parameters and conceptually performs as a regularization technique.

As reported in Table 10, sharing the weights results in higher F1F_{1} scores on ACE 2004, but lower scores on GENIA and CoNLL 2003; the performance differences are largely insignificant. Note that with weight sharing, the span Transformer fully reuses the weights of the standard Transformer, thus requires no additional parameters. This suggests that the performance improvements of DSpERT are from better span representations instead of increased model size.

In addition to the performance, weight sharing has very limited effect on reducing training and inference cost — both the forward and backward computations remain almost unchanged. However, it does effectively halve the parameter number; hence, weight sharing would be a preferred option when the storage space is limited.

Weight Sharing ACE04 GENIA CoNLL03
× 88.05±0.18 81.46±0.25 93.64±0.06
88.11±0.23 81.19±0.26 93.43±0.17
Table 10: The effect of weight sharing. The underlined specification is the one used in our main experiments. All the results are average scores of five independent runs, with subscript standard deviations.

Initial Aggregation Functions.

We test different functions for initial aggregation in the span Transformer. As shown in Table 11, max-pooling outperforms all other alternatives, although the performance differences are limited. This result is different from that in the shallow setting, where max-pooling underperforms multiplicative and additive attentions (Table 3).

One possible explanation is that, the span Transformer blocks with weights initialized from the PLM have already performed very sophisticated multi-head attentions through multiple layers, so one extra attention layer with randomly initialized weights cannot take positive effect further, or even plagues the model. Max-pooling, on the other hand, is parameter-free and thus more advantageous in this case.

Pretrained Language Models.

Table 11 also lists the results by alternative PLMs. In general, BERT-base and BERT-large (Devlin et al., 2019) underperforms RoBERTa-base (Liu et al., 2019), while RoBERTa-large can further improve the performance by 0.54, 0.52 and 0.06 percentage F1F_{1} scores on ACE 2004, GENIA and CoNLL 2003, respectively. These results are consistent with Zhu and Li (2022), confirming the superiority of RoBERTa in NER tasks.

Since GENIA is a biological corpus, some previous studies use BioBERT on this benchmark (Shen et al., 2021, 2022). We also test BioBERT with DSpERT on GENIA. The results show that BioBERT can achieve performance competitive to RoBERTa.

BiLSTM and Boundary Smoothing.

As presented in Table 11, removing the BiLSTM layer will result in a drop of 0.2–0.4 percentage F1F_{1} scores. In addition, replacing boundary smoothing (Zhu and Li, 2022) with the standard cross entropy loss will reduce the F1F_{1} scores by similar magnitudes.

ACE04 GENIA CoNLL03
Initial agg.
   w/ max-pooling 88.05±0.18 81.46±0.25 93.64±0.06
   w/ mean-pooling 87.77±0.31 81.31±0.14 93.58±0.11
   w/ mul. attention 87.67±0.29 81.36±0.17 93.55±0.09
   w/ add. attention 87.90±0.11 81.20±0.12 93.56±0.12
PLM
   w/ RoBERTa-b 88.05±0.18 81.46±0.25 93.64±0.06
   w/ RoBERTa-l 88.59±0.27 81.98±0.41 93.70±0.12
   w/ BERT-b 86.38±0.20 79.13±0.16 91.90±0.08
   w/ BERT-l 87.73±0.30 79.27±0.20 92.79±0.14
   w/ BioBERT-b 81.52±0.26
   w/ BioBERT-l 81.78±0.26
Others
   w/o BiLSTM 87.71±0.05 81.29±0.33 93.42±0.11
   w/o BS 87.66±0.26 81.02±0.27 93.45±0.17
Table 11: Results of ablation studies. “b” and “l” mean the PLM sizes of base and large, respectively; for large PLM, span Transformer has 12 layers. “BS” means boundary smoothing. The underlined specification is the one used in our main experiments. All the results are average scores of five independent runs, with subscript standard deviations.