Learning Semantic Role Labeling from Compatible Label Sequences
Abstract
Semantic role labeling (SRL) has multiple disjoint label sets, e.g., VerbNet and PropBank. Creating these datasets is challenging, therefore a natural question is how to use each one to help the other. Prior work has shown that cross-task interaction helps, but only explored multitask learning so far. A common issue with multi-task setup is that argument sequences are still separately decoded, running the risk of generating structurally inconsistent label sequences (as per lexicons like Semlink). In this paper, we eliminate such issue with a framework that jointly models VerbNet and PropBank labels as one sequence. In this setup, we show that enforcing Semlink constraints during decoding constantly improves the overall F1. With special input constructions, our joint model infers VerbNet arguments from given PropBank arguments with over F1. For learning, we propose a constrained marginal model that learns with knowledge defined in Semlink to further benefit from the large amounts of PropBank-only data. On the joint benchmark based on CoNLL05, our models achieve state-of-the-art F1’s, outperforming the prior best in-domain model by (VerbNet) and (PropBank). For out-of-domain generalization, our models surpass the prior best by (VerbNet) and (PropBank).
1 Introduction
Semantic Role Labeling (SRL, Palmer et al., 2010) aims to understand the role of words or phrases in a sentence. It has facilitated other natural language processing tasks including question answering FitzGerald et al. (2018), sentiment analysis Marasović and Frank (2018), information extraction Solawetz and Larson (2021), and machine translation Rapp (2022).
Semantic role labeling can take various forms, each associated with different datasets. Predicates can be coarsely divided into PropBank Palmer et al. (2005) senses, each with a core set of numbered semantic arguments (e.g., Arg0–Arg5). There are also modifier arguments (e.g., ArgMLoc) typically representing information such as the location, purpose, manner or time of an event. Alternatively, predicates can also be hierarchically clustered into VerbNet Schuler (2005) classes according to similarities in their syntactic behavior. Each class admits a set of thematic roles (e.g. Agent, Theme) whose interpretations are consistent with all predicates within the class.
As a modeling problem, SRL requires associating argument types and phrases with respect to an identified predicate. The two labeling tasks (i.e., VerbNet SRL and PropBank SRL) are closely related; but they differ in their treatment of predicates and have disjoint label sets. Learning jointly can improve data efficiency across the different labeling SRL tasks.
A common formulation used to instantiate this idea in prior work is multitask learning (e.g. Strzyz et al., 2019; Gung and Palmer, 2021): each label set is treated as a separate labeling task, and sometimes also modeled with inter-task feature interaction or consistency losses. While multitask learning often works well in such cases, the loss formulation represents a conservative view over label compatibilities of different tasks. At prediction time, subtask modules still run independently and are not constrained by each other. Consequently, decoded labels may violate structural constraints with respect to each other. In such settings, constrained inference (e.g., Fürstenau and Lapata, 2012; Greenberg et al., 2018) has been found helpful. However, this raises the question of how to involve such inference during learning for better data efficiency. Furthermore, given the wider availability of PropBank-only data (e.g., Pradhan et al., 2013, 2022), how to efficiently benefit from such data also remains a question.
In this paper, we argue that the two disjoint but compatible labeling tasks can be more effectively modeled as one task using their compatibility structures that are already explicitly defined in the form of Semlink Stowe et al. (2021). Semlink offers mappings between various semantic ontologies including PropBank and VerbNet. Gung and Palmer (2021) devised a deterministic conversion from PropBank label sequences to VerbNet ones using only the unambiguous mappings in Semlink. This conversion gives a test bed that has half of the predicates in CoNLL05 SRL dataset Carreras and Màrquez (2005) with both VerbNet and PropBank jointly labeled.
Given this setting, we propose a simple and effective joint CRF model for the VerbNet SRL and PropBank SRL tasks. In addition to the joint CRF, we propose an inference constraint that uses compatible label structures defined in Semlink, and show that our constrained inference achieves higher overall SRL F1—the average of VerbNet and ProbBank F1 scores—than the current state-of-the-art. Indeed, when PropBank labels are observed, it achieves over F1 on the VerbNet SRL, suggesting the possibility of an automated annotation helper. We show that our formulation naturally extends to a constrained marginal model that learns from the more abundant PropBank-only data in a semi-supervised setting. When learning and predicting with constraints, it achieves even better SRL F1 in out-of-domain generalization.111The code for reproducing our experiments: https://github.com/utahnlp/marginal_srl_with_semlink
2 Joint Task of Semantic Role Labeling
We consider modeling VerbNet (VN) and PropBank (PB) SRL as a joint labeling task. Given a sentence , we want to identify a set of predicates (e.g., verbs), and for each predicate, generate two sequences of labels, one for VerbNet arguments , the other for PropBank arguments . With respect to VN parsing, a predicate is associated with a VerbNet class that represents a group of verbs with shared semantic and syntactic behavior, thereby scoping a set of thematic roles for the class. Similarly, the predicate is associated with a PropBank sense tag that defines a set of PB core arguments along with their modifiers. A example is shown in Tab. 1.
Admissible alignments (-) | ||
---|---|---|
price-54.4 | 01 | Agent-Arg0; Theme-Arg1; Value-Arg2 |
price-54.4 | 02 | Agent-Arg0; Theme-Arg1; Value-Arg2 |
admire-31.2 | 02 | Experiencer-Arg0; Stimulus-Arg1 |
We treat predicate classification and argument labeling as separate tasks and focus on the latter.222Prior work (e.g., Täckström et al., 2015) has shown that the predicate disambiguation can be modeled with high accuracy using standalone classifiers. Assuming predicates and their associated VN classes and PB senses are given along with , we can write the prediction problem as:
(1) |
2.1 VerbNet Completion
There is a much larger amount of PropBank-only data (e.g., Pradhan et al., 2013, 2022) than jointly labeled data. Inferring VerbNet labels from observed PropBank labels, therefore, is a realistic use case. This corresponds to the modeling problem:
(2) |
We refer to this scenario as completion mode. In this paper, we will focus on the joint task defined in Eq. 1 while also generalizing our approach to address the completion task in Eq. 2.
2.2 Multitask learning and Its Limitations
When predicting multiple label sequences for SRL, a common approach is multitask learning using dedicated classifiers for each task that operate on a shared representation. The current state-of-the-art model Gung and Palmer (2021) used an LSTM stacked on top of BERT Devlin et al. (2019) to model both PropBank and VerbNet. While each set of the semantic roles is modeled jointly with VerbNet predicates, the argument labeling of the two subtasks is still kept separate.
Separate modeling of VerbNet SRL and PropBank SRL has a clear disadvantage: subtask argument labels might disagree in three ways: 1) in terms of the BIO tagging scheme—e.g., a word having a B-* VN label and a I-* PropBank label, or 2) assigning semantically invalid label pairs—e.g., an ArgM-Loc being called a Theme, or 3) violating Semlink constraints. In Sec. 6, we show that a model with separate task classifiers, while having a close to state-of-the-art F1, can have a fair amount of argument assignment errors with respect to Semlink, especially for out-of-domain inputs.
3 A Joint CRF Model
To eliminate the errors discussed in Sec. 2.2, we propose to model the disjoint SRL tasks using a joint set of labels. This involves converting multitask modeling into a single sequence labeling task whose labels are pairs of PB and VN labels. Doing so not only eliminates the BIO inconsistency, but also exposes an interface for injecting Semlink constraints.
Our model uses RoBERTa Liu et al. (2019) as the backbone to handle textual encoding, similarly to the SRL model of Li et al. (2020). At a high level, we use a stack of linear layers with GELU activations Hendrycks and Gimpel (2016) to encode tokens to be classified for a predicate. For the problem of predicting arguments of a predicate , we have an encoding vector for the -th word in the input text .
(3) | ||||
(4) |
Here, map sums up word-piece embeddings to form a sequence of word-level embeddings, the functions and are both linear layers, and denotes a two-layer network with GELU activations in the hidden layer. We use a dedicated module of the form in Eq. 4 for the VN and PB subtasks. This gives us a sequence of vectors for VN and a sequence of vectors for PB.
Next, we project the VN and PB feature sequences into a label space:
(5) |
Here, is another two-layer GELU network followed by a linear projection that outputs scores, corresponding to VN-PB label pairs. The final result denotes a sequence of VN-PB label scores for a specific predicate . In addition, we use a CRF as a standard first-order sequence model over (treating it as emission scores), and use Viterbi decoding for inference. The training objective is to maximize:
(6) |
where denotes the scoring function for a label sequence that adds up the emission and the transition scores, and the term denotes the partition that sums exponentiated scores over all label sequences. The term denotes the label sequence that has VN and PB jointly labeled. We will refer to this model as the joint CRF, and the label sequence as the joint labels.
Reduced Joint Label Space.
We use the cross-product the two label sets, prefixed with a BIO prefix333For example, the pair (B-Theme, B-Arg1).. A brute-force cross product leads to a label space. In practice, it is important to keep the joint label space at a small scale for efficient computation, especially for the CRF module. Therefore, we condense it by first disallowing pairs of the form (B-*, I-*) and predicate-to-argument pairs. The former enforces that the VerbNet arguments do not start within ProbBank arguments, while the latter ensures that the predicate is not part of any argument. Next, we observe the co-occurence pattern of VN and PB arguments, disabling semantically invalid pairs such as (Theme, ArgM-Loc)444VerbNet arguments typically align to PropBank core arguments, not modifiers; thus pairs like (Actor, ArgM-Cau) can be filtered out.. This reduces the label space by an order of magnitude (from to ).
Input Construction using Predicates.
We take inspiration from prior work (e.g. Zhou and Xu, 2015; He et al., 2017; Zhang et al., 2022) to explicitly put predicate features as part of the input to augment textual information. At the same time, we also seek to maintain a simple construction that can be easily adapted to a semi-supervised setting (i.e. compatible with PropBank-only data). To this end, we propose a simple solution that appends the PropBank senses of potential predicates to the original sentence :
where denotes the input words, and denotes the senses of the predicates. In practice, we use the PropBank roleset IDs which consist of a pair of (lemma, sense)—e.g., run.01. Our models only take the encodings for after the RoBERTa encoder and ignore the rest. We consider this design to be more efficient than prior work (e.g. Gung and Palmer, 2021; Zhang et al., 2022) that dedicated text feature for each predicate. In our setup, the argument labeling for different predicates shares the same input, thus no need to run encoding multiple times for multiple predicates.
4 Semi-supervised Learning with PropBank-only Data
Compared to data with both VerbNet and PropBank fully annotated, there is more data with only PropBank labeled. The Semlink corpus helps in unambiguously mapping of the CoNLL05 data Gung and Palmer (2021). Therefore, a natural question is: can we use PropBank-only data to improve the joint task?
Here, we explore model variants based on the joint CRF architecture described in Sec. 3. We will focus on modeling for PropBank-only sentences.
4.1 Separate Classifiers for VN and PB
4.2 Dedicated PropBank Classifier
Another option is to retain the joint CRF for the jointly labeled dataset and use an additional dedicated CRF for PB-only sentences. Note that this setup is different from the model in Sec. 4.1. As before, we let these two to share the same encoder and, for Eq. 4-6, they have dedicated trainable weights.
During inference, we rely on the Viterbi decoding associated with the joint CRF module to make predictions. In our preliminary experiments, the joint CRF and the dedicated PropBank CRF achieve similar F1 on PropBank arguments.
4.3 Marginal CRF
For partially labeled sequences, we take inspiration from Greenberg et al. (2018) to maximize the marginal distribution of those observed labels. In our joint CRF, the marginalization assumes uniform distribution over VN arguments that are paired with observed PB arguments. The learning objective is to maximize the probabilities of such label sequences as a whole:
(7) | |||
where denotes a potential joint label sequence with only PropBank arguments observed for a predicate . Scores of such label sequences are aggregated by the Lse operator. Note that the marginal CRF and the joint CRF (Eq. 6) use the same model architecture, just with a different loss.
4.4 Marginal Model with Semlink (Marginal)
The log marginal probability in Eq. 7 assumes uniform distribution over a large label space. It included any arbitrary VerbNet arguments paired to the observed PropBank labels. In practice, we can narrow it down to only legitimate VN-PB argument pairs defined in Semlink. Such legitimate space is uniquely determined by a VerbNet class and PropBank sense . We will refer to label sequences that comply with this space as , and apply it on Eq. 7:
(8) |
Note that this formulation essentially changes the global optimization into a local version which implicitly requires using Semlink at inference time. We will present the details of in Sec. 5. Intuitively, it zeros out losses associated with joint label traces that violate Semlink constraints. During training, we found that it is important to apply this constraint to both B-* and I-* labels.
Where to apply ?
Technically, the summation over reduced label space can be applied at different places, such as the partition in Eq. 6. We will report performances on this setting in Sec. 7.2. In short, plugging the label filter to the joint CRF (therefore jointly labeled data) has little impact on F1 scores, thus we reserve it for the PropBank-only data (as in Eq. 8).
5 Inference with Semlink
Here we discuss the implementation of the . Remember that each pair of VerbNet class and PropBank sense uniquely determines a set of joint argument labels for the predicate . For brevity, let us denote this set as (e.g., Tab. 1). Eventually, we want the Viterbi-decoded label sequence to comply with . That is,
(9) |
where denotes the location in a sentence, is the joint label sequence, consisting of pairs. The constraint in Eq. 9 translates as: if a VerbNet argument is present in the predicate ’s Semlink entry, we prevent it from aligning to any PropBank arguments not defined in Semlink; and the same applies to PropBank arguments.
This constraint can be easily implemented by a masking operation on the emission scores of the joint CRF, thus can be used at both training and inference time. During inference, it effectively ignores those label sequences with Semlink violations during Viterbi decoding:
(10) |
VerbNet Label Completion.
For models based on our joint CRF, we mask out joint labels that are not defined in during inference, similar to Eq. 8. For models with separate VN and PB classifiers (in Sec. 4.1), we enforce the VN’s Viterbi decoding to only search arguments that are compatible with the gold in Semlink. Furthermore, we always use the constraint (Eq. 9) in the completion mode.
Train | Inf | Dev | WSJ | Brown | |||||||
Data | Fine-tune | Seml | VN | PB | VN | PB | VN | PB | |||
Joint | IWCS2021 (gold & ) | - | - | - | 88.2 | 88.7 | - | 83.0 | 82.8 | - | |
Multitask | ✗ | 89.51.09 | 87.91.17 | 3.68.06 | 90.74.23 | 89.33.15 | 3.43.50 | 83.691.54 | 81.73.98 | 8.71.74 | |
Joint | ✓ | 90.20.22 | 87.58.18 | 0 | 91.42.11 | 89.12.16 | 0 | 85.901.11 | 82.56.52 | 0 | |
Semi | Multitask | ✗ | 89.56.09 | 88.30.42 | 4.82.35 | 90.45.09 | 89.34.74 | 4.08.37 | 84.28.85 | 82.51.70 | 10.48.16 |
Joint+CRF | ✓ | 90.73.32 | 88.42.11 | 0 | 91.52.36 | 89.30.29 | 0 | 85.72.13 | 82.33.31 | 0 | |
Marginal | ✓ | 90.91.27 | 88.68.11 | 0 | 91.74.22 | 89.51.28 | 0 | 85.391.74 | 82.461.06 | 0 | |
Marginal | ✓ | 90.87.32 | 88.59.24 | 0 | 91.55.28 | 89.49.18 | 0 | 86.39.17 | 83.04.37 | 0 |
6 Experiments
In this section, we aim to verify whether the compatibility structure between VerbNet and PropBank (in the form of Semlink) has a positive impact on their sequence labeling performances.
Train | Inf | Dev | WSJ | Brown | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|
Data | Fine-tune | Seml | VN | PB | VN | PB | VN | PB | |||
Joint | Joint | ✗ | 89.28 | 87.70 | 1.58 | 90.73 | 89.10 | 1.59 | 84.43 | 82.48 | 5.99 |
Joint | ✓ | 90.20 | 87.58 | 0 | 91.42 | 89.12 | 0 | 85.90 | 82.56 | 0 | |
Semi | Joint+CRF | ✗ | 89.66 | 88.31 | 2.74 | 90.61 | 89.21 | 2.37 | 83.53 | 81.87 | 7.59 |
Joint+CRF | ✓ | 90.73 | 88.42 | 0 | 91.52 | 89.30 | 0 | 85.72 | 82.33 | 0 | |
Marginal | ✗ | 89.79 | 88.50 | 2.92 | 90.73 | 89.30 | 2.39 | 83.23 | 81.88 | 6.74 | |
Marginal | ✓ | 90.91 | 88.68 | 0 | 91.74 | 89.51 | 0 | 85.39 | 82.46 | 0 | |
Marginal | ✗ | 89.53 | 88.55 | 2.29 | 90.70 | 89.46 | 1.98 | 83.64 | 82.71 | 7.77 | |
Marginal | ✓ | 90.87 | 88.59 | 0 | 91.55 | 89.49 | 0 | 86.39 | 83.04 | 0 |
6.1 Data
We follow the prior state-of-the-art Gung and Palmer (2021) in extracting VerbNet labels from the CoNLL05 dataset using the Semlink corpus. We use the same version of Semlink to extract the data for training and evaluation. Therefore, our F1 scores are directly comparable with theirs (denoted as Iwcs2021 in Table 2) The resulting dataset accounts for about of the CoNLL05 predicates, across training, development and test sets (including WSJ and Brown). We will refer to this data as the joint data column in Table 2. For semi-supervised learning, we incorporate the rest of the PropBank-only predicates in the CoNLL05 training split. For development and testing, we use the splits in the joint dataset for fair comparison with prior work.
6.2 Training and Evaluation
We adopt the same fine-tuning strategy as in Li et al. (2020)—we fine-tune twice since this generally outperforms fine-tuning only once, even with the same number of total epochs. In the first round, we fine-tune our model for epochs. In the second round, we restart the optimizer and learning rate scheduler, and fine-tune for epochs. In both rounds, checkpoints with the highest average VN/PB development F1 are saved. For different model variants, we report the average F1 from models trained with random seeds.
For the Semlink constraint, we use the official mapping between VN/PB arguments555https://github.com/cu-clear/semlink/blob/master/instances/semlink-2. When involved in constrained inference, we use the gold VerbNet classes and PropBank senses .
For evaluation, in addition to the standard VN and PB F1 scores, we also report the percent of predicates with predictions that are inconsistent with Semlink, denoted by .
6.3 Performance on Semlink Extracted CoNLL05
We want to compare models trained on the joint dataset and variants (in Sec. 4) on the semi-supervised setup. Table 2 presents their performances along with Semlink violation rates in model predictions. Note that the ground truth joint data bears no violation at all (i.e., ).
Multitask involves Semlink violations.
Firstly, we show the limitations of multitask learning. While the architecture is simple, the testing scores mostly outperform the previous state-of-the-art, except the Brown PropBank F1. However, there is a fair percentage of predicates having structurally wrong predictions, especially in the Brown test set. With semi-supervised learning, VN and PB F1s are improved on the Brown set while slightly lowered on the WSJ VN. This also comes with a degraded Semlink error rate ( on WSJ and on Brown). While Semlink inconsistency is not reported in Gung and Palmer (2021), we believe that, due to the nature of multitask learning, Semlink errors are inevitable.
Joint CRF outperforms multitask learning.
A direct comparison is between the Multitask v.s. Joint. Our joint CRF obtains higher overall SRL F1 across the WSJ and the Brown sets. A similar observation applies to the semi-supervised setting where Multitask compares to Joint+CRF. Most of such improvements are from the use of inference-time Semlink constraints.
Inference with Semlink improves SRL.
In Table 3, we do side-to-side comparison of using versus not using the Semlink structure during inference. We do so for each modeling variant. With constrained inference, models no longer have Semlink structural violations (). And this results in a clear trend where using Semlink systematically improves the F1 scores. We hypothesize this is due to the reduced search space which makes the decoding easier. Likely due to the higher granularity of VerbNet argument types compared to PropBank, a majority of the improvements are on the VN F1s.
Does semi-supervised learning make a difference?
The answer is that it depends. For Multitask, using PropBank-only data traded off a bit on the overall WSJ F1 but improved the out-of-domain performances. Accompanied with this trade-off is the slightly higher inconsistency rate . The Joint+CRFPB model tells an opposite story that the partially labeled data is favorable on the in-domain test but not so in the out-of-domain test. This observation is also consistent with both the Marginal CRF and constrained Marginal model (Marginal). Furthermore, when performance improves, the margins on VN and PB are fairly distributed. Finally, we should note that, neither the Joint+CRF nor the Marginal have a better Brown F1 than the Joint, meaning that they did not use the PB-only data efficiently.
Impact of marginal CRF.
We compare the Joint+CRF to Marginal to see how a single CRF handles partially labeled data. The latter outperforms the former consistently by on the in-domain test set but performed slightly worse on the out-of-domain Brown set. Comparing to the Joint model, it seems that naively applying marginal CRF leads to even worse generalization.
Constrained marginal model improves generalization.
We want to see if our constrained model can help learning. Modeling PropBank-only data with a separate classifier (i.e., the Multitask and Joint+CRF) failed to do so (although they indeed work better on the in-domain WSJ). In contrast, our constrained Marginal apparently learns from the PropBank-only data more efficiently, achieving strong in-domain performance and substantially better out-of-domain generalization. This suggests that learning with constraints works better with partially labeled data. Interestingly though, it seems that the constraint is optional for fully annotated data since the Marginal only enables the constraint on PB-only data. We verify this phenomenon in Sec. 7.2 with an ablation study.
Statistical Significance of Constrained Inference.
We measure statistical significance using a t-test implemented by Dror et al. (2018) on predictions from models in Table 3. For each model, we compare inference with and without Semlink constraints (Sec. 5). For a fair comparison, we limit predictions from the model trained with the same random seed in each test, and apply the test for all random seeds (3 in total). We observe that the improvements on VerbNet F1’s are universally significant. The p-values are far less than across different testing data, models, and random seeds. This aligns to the observation in Table 3 that VN F1 has a substantial F1 boost while PB F1 improvements tend to be marginal.
To look closer, we examined the predictions of a Joint model on the Dev set ( predicates), and found that, after using Semlink during inference, wrongly predicted predicates in VN SRL were corrected (i.e., improved predicate-wise F1), and no predicates received a degraded F1. However, for PropBank SRL, there were predicates corrected by the constraint while became errors.
6.4 VerbNet Label Completion from PropBank
As discussed in Sec. 1, we also aim to address the realistic use case of VerbNet completion. Table 4 summarizes the performances for VerbNet argument prediction when gold PropBank arguments are given. In the completion mode, the Joint model performs generally better than all the semi-supervised models. This phenomenon is likely because the Joint model is optimized for the probability , while the semi-supervised models, in one way or the other, have a term for on PB-only data. The latter term does not explicitly boost model’s discriminative capability on the unique ground truth.
In addition to the input construction in Sec. 3, we propose a special construction for the VN completion mode by using the PB arguments as text input.:
where denotes the PropBank argument label for the -th word. For the predicate word, we use the predicate feature (i.e. lemma and sense). Compared to , this formulation makes the computation less efficient as the input is no longer shared across different predicates. However, it offers a more tailored input signal and delivers above F1 on both WSJ and Brown.
Data | Model | WSJ VN | Brown VN |
---|---|---|---|
Joint | Multitask | 83.14 | 86.26 |
Joint | 99.83 | 98.12 | |
Semi | Multitask | 93.08 | 85.71 |
Joint+CRF | 99.81 | 97.88 | |
Marginal | 99.76 | 97.45 | |
Marginal | 99.68 | 97.71 | |
Joint | Joint | 99.85 | 99.02 |
7 Analysis
We report statistical metrics in Sec. 7.1. In Sec. 7.2, we analyze the use of constrained learning on the jointly labeled data.
7.1 Variance of SRL Models
A majority of F1s in Tab. 2 vary in a small range. Models trained on joint-only data show higher variance on the out-of-domain Brown test set. Among the semi-supervised models, the Marginal models exhibit high F1 variance on the Brown set while the Marginal models work more stably.
7.2 Impact of Learning with Semlink Constraint on Joint Data
In Table 5, we use the form of constrained learning in Eq. 8 but apply it on the joint CRF loss over the jointly labeled training data. Note that the constraint term only affects the denominator part in Eq. 6. Overall, the effect of Semlink at training time seems small. On the WSJ test set, both VN and PB F1s are fairly close. The Brown test F1s have a drop, especially on VN, suggesting that constrained learning on the joint data is not needed.
Joint | WSJ | Brown | |||
---|---|---|---|---|---|
Semlink | VN | PB | VB | PB | |
1 | ✗ | 91.42 | 89.12 | 85.90 | 82.56 |
2 | ✓ | 91.42 | 89.21 | 85.41 | 82.46 |
7.3 Reliance on Gold Predicate Labels
While we focused on experiments that use given gold predicate labels (i.e., VN class and PB sense ), our models do not rely on them. Intuitively, there will be a performance degradation on argument labeling when predicate labels are not always accurate. To study this degradation curve, we experiment with randomly corrupted predicate classes/senses, to illustrate the performance dependency on predicate disambiguation. Specifically, we perturb the development split by randomly swapping a predicate’s VN class (or PB sense) with another one from the dataset and observe how well our best model performs. This setup essentially simulates a real-world testing scenario where predicate labels on either VN/PB are imperfect. Results in Tab. 6 suggest that F1 degradation happens in a smooth way due to error propagation.
corruption (%) | 0 | 5 | 10 | 20 | 30 |
---|---|---|---|---|---|
VN F1 | 91.23 | 89.45 | 83.64 | 82.90 | 79.02 |
PB F1 | 88.82 | 88.23 | 87.32 | 85.56 | 84.66 |
corruption (%) | 0 | 5 | 10 | 20 | 30 |
VN F1 | 91.23 | 90.35 | 89.55 | 87.18 | 84.94 |
PB F1 | 88.82 | 88.17 | 87.21 | 84.92 | 83.11 |
8 Discussion
The use of constraints in SRL has a long history that mostly focuses on the PropBank SRL task. Earlier work investigated inference with constraints (e.g. Punyakanok et al., 2004; Surdeanu et al., 2007; Punyakanok et al., 2008). Other work developed constrained models for learning (e.g. Chang et al., 2012) ; or incorporated constraints with emerging neural models (e.g. Riedel and Meza-Ruiz, 2008; Fürstenau and Lapata, 2012; Täckström et al., 2015; FitzGerald et al., 2015; Li et al., 2020).
VerbNet SRL, on the other hand, is often studied as a comparison or a helper for PropBank SRL Kuznetsov and Gurevych (2020). Yi et al. (2007) showed that the mapping between VerbNet and PropBank can be used to disambiguate PropBank labels. It has been shown that model performance on VerbNet SRL is affected more by predicate features than PropBank SRL Zapirain et al. (2008). In a sense, our observation that VerbNet F1 gains larger improvements from Semlink is also consistent with prior work. Beyond comparison work, Kazeminejad et al. (2021) explored the downstream impact of VerbNet SRL and showed promising uses in entity state tracking.
Multitask Learning
The closest work to this paper is Gung and Palmer (2021). Instead of modeling argument labels and predicate classes via multitasking, we adopted a simpler design and focused on joint SRL argument labeling. This comes with two benefits: 1) a focused design that models joint labels; 2) an easy extension for using marginal CRF for partially labeled data. Other technical differences include a generally better transformer (i.e., RoBERTa) instead of BERT Vaswani et al. (2017), simpler input construction, and our proposal of the completion mode.
Marginal CRF
Greenberg et al. (2018) explored the use of marginal CRF on disjoint label sequences in the biomedical domain. Disjoint label sequences are concatenated into one, thus requiring dedicated decoding to reduce inconsistency w.r.t. various structure patterns (e.g., aligned BIO pattern). In this paper, we took a step further by pairing label sequences to form a joint SRL task, allowing an easy interface for injecting decoding constraints.
Broader Impact
Recent advantages in large language models (LLM) have shown promising performance in semantic parsing tasks (e.g. Drozdov et al., 2022; Mekala et al., 2022; Yang et al., 2022). A well-established approach is via iterative prompting (e.g., Chain-of-Thought Wei et al. (2022)) and potentially using in-domain examples for prompt construction. While such work bears many technical differences from this work, there are advantages that can potentially be shared. For instance, our direct use of a knowledge base (Semlink in our case) allowed for a guarantee of violations; and LLM-based work is less reliant on training data. Another scenario is when treating semantic structures as explicit intermediate products, such as for language generation. Our joint modeling allows for accuracy in converting PropBank arguments to VN arguments. When using such labels for prompted inference, it can make fewer errors.
Conclusions
In this work, we presented a model that learns from compatible label sequences for the SRL task. The proposal includes a joint CRF design, extension for learning from partially labeled data, and reasoning and learning with Semlink constraints. On the VerbNet and PropBank benchmark based on CoNLL05, our models achieved state-of-the-art performance with especially strong out-of-domain generalization. For the newly proposed task of completing VerbNet arguments given PropBank labels, our models are near perfect, achieving over F1 scores.
9 Limitations
Towards fully end-to-end parser.
Our model architecture is on the track of end-to-end SRL parser, but it still assumes gold predicate positions and predicate attributes are given. A fully end-to-end parser can take sentences in raw text and output disjoint label sequences. While doing so can make computation less efficient (e.g., requiring substantially larger memory for training), it can bring users convenience.
Involving document context.
Gung and Palmer (2021) showed that using neighboring sentence prediction with transformer positively impacts parsing F1. In contrast, we assumed sentences in the corpus are independent.
Why does PropBank seem more difficult?
We hypothesized the reason to be less granularity in argument labels and more ambiguous label assignments. As mentioned in Sec. 3, prior work benefited from using dedicated label text/definition as an auxiliary input. We only used such features at the predicate level, implicitly trading off potential gains on PB F1 for more efficient computation.
Marginal model’s capacity at handling constraints.
In this paper, we focused on the Semlink constraint for compatible label sequences. There is a broad spectrum of SRL constraints in prior work Punyakanok et al. (2008), some of them do not easily fit in the marginalization formulation, such as the unique core role constraint.
Acknowledgement
We thank the reviewers for their helpful insights and comments. We gratefully acknowledge the support of NSF under grants #1801446 (SATC) and #1822877 (Cyberlearning), and of DARPA CwC (subcontracts from UIUC and SIFT), DARPA AIDA Award FA8750-18-2-0016 (RAMFIS), DARPA KAIROS FA8750-19-2-1004-A20-0047-S005 (RESIN, sub to RPI, PI: Heng Ji), as well as DTRA HDTRA1-16-1-0002/Project 1553695 (eTASC - Empirical Evidence for a Theoretical Approach to Semantic Components).
The views and conclusions contained herein are those of the authors and should not be interpreted as necessarily representing the official policies of any government agency.
References
- Carreras and Màrquez (2005) Xavier Carreras and Lluís Màrquez. 2005. Introduction to the CoNLL-2005 shared task: Semantic role labeling. In Proceedings of the Ninth Conference on Computational Natural Language Learning (CoNLL-2005), pages 152–164, Ann Arbor, Michigan. Association for Computational Linguistics.
- Chang et al. (2012) Ming-Wei Chang, Lev Ratinov, and Dan Roth. 2012. Structured Learning with Constrained Conditional Models. Machine learning, 88(3):399–431.
- Devlin et al. (2019) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186.
- Dror et al. (2018) Rotem Dror, Gili Baumer, Segev Shlomov, and Roi Reichart. 2018. The hitchhiker’s guide to testing statistical significance in natural language processing. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1383–1392, Melbourne, Australia. Association for Computational Linguistics.
- Drozdov et al. (2022) Andrew Drozdov, Nathanael Schärli, Ekin Akyürek, Nathan Scales, Xinying Song, Xinyun Chen, Olivier Bousquet, and Denny Zhou. 2022. Compositional semantic parsing with large language models. arXiv preprint arXiv:2209.15003.
- FitzGerald et al. (2018) Nicholas FitzGerald, Julian Michael, Luheng He, and Luke Zettlemoyer. 2018. Large-scale QA-SRL parsing. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2051–2060, Melbourne, Australia. Association for Computational Linguistics.
- FitzGerald et al. (2015) Nicholas FitzGerald, Oscar Täckström, Kuzman Ganchev, and Dipanjan Das. 2015. Semantic role labeling with neural network factors. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pages 960–970.
- Fürstenau and Lapata (2012) Hagen Fürstenau and Mirella Lapata. 2012. Semi-supervised semantic role labeling via structural alignment. Computational Linguistics, 38(1):135–171.
- Greenberg et al. (2018) Nathan Greenberg, Trapit Bansal, Patrick Verga, and Andrew McCallum. 2018. Marginal likelihood training of BiLSTM-CRF for biomedical named entity recognition from disjoint label sets. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 2824–2829, Brussels, Belgium. Association for Computational Linguistics.
- Gung and Palmer (2021) James Gung and Martha Palmer. 2021. Predicate representations and polysemy in VerbNet semantic parsing. In Proceedings of the 14th International Conference on Computational Semantics (IWCS), pages 51–62, Groningen, The Netherlands (online). Association for Computational Linguistics.
- He et al. (2017) Luheng He, Kenton Lee, Mike Lewis, and Luke Zettlemoyer. 2017. Deep semantic role labeling: What works and what’s next. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 473–483, Vancouver, Canada. Association for Computational Linguistics.
- Hendrycks and Gimpel (2016) Dan Hendrycks and Kevin Gimpel. 2016. Gaussian error linear units (gelus). arXiv preprint arXiv:1606.08415.
- Kazeminejad et al. (2021) Ghazaleh Kazeminejad, Martha Palmer, Tao Li, and Vivek Srikumar. 2021. Automatic entity state annotation using the VerbNet semantic parser. In Proceedings of the Joint 15th Linguistic Annotation Workshop (LAW) and 3rd Designing Meaning Representations (DMR) Workshop, pages 123–132, Punta Cana, Dominican Republic. Association for Computational Linguistics.
- Kuznetsov and Gurevych (2020) Ilia Kuznetsov and Iryna Gurevych. 2020. A matter of framing: The impact of linguistic formalism on probing results. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 171–182, Online. Association for Computational Linguistics.
- Li et al. (2020) Tao Li, Parth Anand Jawale, Martha Palmer, and Vivek Srikumar. 2020. Structured tuning for semantic role labeling. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 8402–8412, Online. Association for Computational Linguistics.
- Liu et al. (2019) Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692.
- Marasović and Frank (2018) Ana Marasović and Anette Frank. 2018. SRL4ORL: Improving opinion role labeling using multi-task learning with semantic role labeling. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 583–594, New Orleans, Louisiana. Association for Computational Linguistics.
- Mekala et al. (2022) Dheeraj Mekala, Jason Wolfe, and Subhro Roy. 2022. Zerotop: Zero-shot task-oriented semantic parsing using large language models. arXiv preprint arXiv:2212.10815.
- Palmer et al. (2005) Martha Palmer, Daniel Gildea, and Paul Kingsbury. 2005. The Proposition Bank: An annotated corpus of semantic roles. Computational Linguistics, 31(1):71–106.
- Palmer et al. (2010) Martha Palmer, Daniel Gildea, and Nianwen Xue. 2010. Semantic role labeling. Synthesis Lectures on Human Language Technologies, 3(1):1–103.
- Pradhan et al. (2022) Sameer Pradhan, Julia Bonn, Skatje Myers, Kathryn Conger, Tim O’gorman, James Gung, Kristin Wright-bettner, and Martha Palmer. 2022. PropBank comes of Age—Larger, smarter, and more diverse. In Proceedings of the 11th Joint Conference on Lexical and Computational Semantics, pages 278–288, Seattle, Washington. Association for Computational Linguistics.
- Pradhan et al. (2013) Sameer Pradhan, Alessandro Moschitti, Nianwen Xue, Hwee Tou Ng, Anders Björkelund, Olga Uryupina, Yuchen Zhang, and Zhi Zhong. 2013. Towards robust linguistic analysis using OntoNotes. In Proceedings of the Seventeenth Conference on Computational Natural Language Learning, pages 143–152, Sofia, Bulgaria. Association for Computational Linguistics.
- Punyakanok et al. (2008) Vasin Punyakanok, Dan Roth, and Wen-tau Yih. 2008. The importance of syntactic parsing and inference in semantic role labeling. Computational Linguistics, 34(2):257–287.
- Punyakanok et al. (2004) Vasin Punyakanok, Dan Roth, Wen-tau Yih, and Dav Zimak. 2004. Semantic role labeling via integer linear programming inference. In COLING 2004: Proceedings of the 20th International Conference on Computational Linguistics, pages 1346–1352, Geneva, Switzerland. COLING.
- Rapp (2022) Reinhard Rapp. 2022. Using semantic role labeling to improve neural machine translation. In Proceedings of the Thirteenth Language Resources and Evaluation Conference, pages 3079–3083, Marseille, France. European Language Resources Association.
- Riedel and Meza-Ruiz (2008) Sebastian Riedel and Ivan Meza-Ruiz. 2008. Collective semantic role labelling with Markov logic. In CoNLL 2008: Proceedings of the Twelfth Conference on Computational Natural Language Learning.
- Schuler (2005) KK Schuler. 2005. Verbnet: A broad-coverage, comprehensive verb lexicon. Ph. D. Thesis, University of Pennsylvania.
- Solawetz and Larson (2021) Jacob Solawetz and Stefan Larson. 2021. LSOIE: A large-scale dataset for supervised open information extraction. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, pages 2595–2600, Online. Association for Computational Linguistics.
- Stowe et al. (2021) Kevin Stowe, Jenette Preciado, Kathryn Conger, Susan Windisch Brown, Ghazaleh Kazeminejad, James Gung, and Martha Palmer. 2021. SemLink 2.0: Chasing lexical resources. In Proceedings of the 14th International Conference on Computational Semantics (IWCS), pages 222–227, Groningen, The Netherlands (online). Association for Computational Linguistics.
- Strzyz et al. (2019) Michalina Strzyz, David Vilares, and Carlos Gómez-Rodríguez. 2019. Sequence labeling parsing by learning across representations. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 5350–5357, Florence, Italy. Association for Computational Linguistics.
- Surdeanu et al. (2007) Mihai Surdeanu, Lluís Màrquez, Xavier Carreras, and Pere R Comas. 2007. Combination strategies for semantic role labeling. Journal of Artificial Intelligence Research, 29:105–151.
- Täckström et al. (2015) Oscar Täckström, Kuzman Ganchev, and Dipanjan Das. 2015. Efficient inference and structured learning for semantic role labeling. Transactions of the Association for Computational Linguistics.
- Vaswani et al. (2017) Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. 2017. Attention is All You Need. In Advances in Neural Information Processing Systems.
- Wei et al. (2022) Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, brian ichter, Fei Xia, Ed Chi, Quoc V Le, and Denny Zhou. 2022. Chain-of-thought prompting elicits reasoning in large language models. In Advances in Neural Information Processing Systems, volume 35, pages 24824–24837. Curran Associates, Inc.
- Wolf et al. (2019) Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, Rémi Louf, Morgan Funtowicz, et al. 2019. Transformers: State-of-the-art natural language processing. arXiv preprint arXiv:1910.03771.
- Yang et al. (2022) Jingfeng Yang, Haoming Jiang, Qingyu Yin, Danqing Zhang, Bing Yin, and Diyi Yang. 2022. SEQZERO: Few-shot compositional semantic parsing with sequential prompts and zero-shot models. In Findings of the Association for Computational Linguistics: NAACL 2022, pages 49–60, Seattle, United States. Association for Computational Linguistics.
- Yi et al. (2007) Szu-ting Yi, Edward Loper, and Martha Palmer. 2007. Can Semantic Roles Generalize Across Genres? In Human Language Technologies 2007: The Conference of the North American Chapter of the Association for Computational Linguistics; Proceedings of the Main Conference, pages 548–555, Rochester, New York. Association for Computational Linguistics.
- Zapirain et al. (2008) Beñat Zapirain, Eneko Agirre, and Lluís Màrquez. 2008. Robustness and generalization of role sets: PropBank vs. VerbNet. In Proceedings of ACL-08: HLT, pages 550–558, Columbus, Ohio. Association for Computational Linguistics.
- Zhang et al. (2022) Li Zhang, Ishan Jindal, and Yunyao Li. 2022. Label definitions improve semantic role labeling. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 5613–5620, Seattle, United States. Association for Computational Linguistics.
- Zhou and Xu (2015) Jie Zhou and Wei Xu. 2015. End-to-end learning of semantic role labeling using recurrent neural networks. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 1127–1137, Beijing, China. Association for Computational Linguistics.
Appendix A Appendix
A.1 Hyperparameters
As discussed in Sec. 6, we adopted a 2-stage fine-tuning for each of our models. In the first round, we fine-tune for epochs with initial learning rate of . In the second round, we fine-tune for another epochs with fresh start on the optimizer. Across the two stages, we applied a dropout layer with rate (which is preliminarily grid-searched from ) before each of the linear layers in Sec. 3. The RoBERTa transformer uses the built-in default configurations which is implemented by Wolf et al. (2019).
The learning rate () in the second round of fine-tuning varies by model. We preliminarily grid-searched from . Specifically, for each learning rate, we trained random models and chose the one with the highest overall SRL F1 on the Dev set. We put their values in Table 7.
Data | Model | |
Joint | Multitask | |
Joint | ||
Semi | Multitask | |
Joint+CRF | ||
Marginal | ||
Marginal |