Improving Human-Labeled Data through Dynamic Automatic Conflict Resolution
Abstract
This paper develops and implements a scalable methodology for (a) estimating the noisiness of labels produced by a typical crowdsourcing semantic annotation task, and (b) reducing the resulting error of the labeling process by as much as 20-30% in comparison to other common labeling strategies. Importantly, this new approach to the labeling process, which we name Dynamic Automatic Conflict Resolution (DACR), does not require a ground truth dataset and is instead based on inter-project annotation inconsistencies. This makes DACR not only more accurate but also available to a broad range of labeling tasks. In what follows we present results from a text classification task performed at scale for a commercial personal assistant, and evaluate the inherent ambiguity uncovered by this annotation strategy as compared to other common labeling strategies.
1 Introduction
Ground truth datasets for supervised learning are commonly collected through crowdsourcing annotation projects, either on public platforms (e.g. Yandex.Toloka, Amazon Mechanical Turk, Figure Eight) or on in-house counterparts. In both scenarios, the annotation project is conducted by first providing workers with annotation guidelines and then assigning each one with a set of annotation requests. Each unique annotation request is usually assigned to more than one worker, to increase the reliability of the results. Subsequently, some post-processing technique is used to create a consensus label that is recorded as the ground truth.
However, since human annotations are intrinsically noisy and subject to the influence of external factors, such a process is unlikely to produce accurate ground truth labels for 100% of the requests. Moreover, the practice of using a singular consensus label for multi-annotated requests mandates a one and only correct interpretation for every request in any given task, effectively denying the existence of potential ambiguities. This implicit denial of ambiguity, while acceptable for certain tasks such as arithmetics, can be problematic in tasks such as natural language understanding [Dumitrache et al., 2018a]. Perhaps because of these difficulties, estimating the latent accuracy of any annotation process has been a relatively under-examined subject.
In this paper, we first introduce a scalable methodology to estimate the latent inaccuracy of any crowdsourcing annotation process. The estimation method does not require any preexisting ground truth dataset, and is therefore immune to the Gold Question Inference type of attacks described by Checco et al. [Checco et al., 2018]. We construct an agent-based model to examine the impact of different annotation strategies on overall annotation accuracy. Inspired by the findings of our simulation experiments, we propose and implement a novel process of Dynamic Automatic Conflict Resolution (DACR), designed to improve the accuracy of the annotation output with limited incremental workload while collecting sufficient annotation input for requests with ambiguity. We present annotation results from a DACR project used in the context of a text classification task, implemented at scale for a commercial personal assistant.
The primary contributions of this paper are:
-
1.
latent inaccuracy estimation methodology: we propose a simple, robust, scalable methodology and theoretical framework to study the latent inaccuracies of any crowdsourcing annotation process;
-
2.
comparative study of different annotation strategies: we conduct an in-depth comparison of the performance of commonly used annotation strategies via agent-based simulation experiments;
-
3.
DACR, an ambiguity-preserving and accuracy-boosting annotation strategy: we introduce the Dynamic Automatic Conflict Resolution strategy to improve the accuracy of the annotation process;
-
4.
empirical analysis on a text classification task: we present statistical findings from a DACR annotation project performed at scale on a text classification task.
2 Related Work
This work relates to recent developments in three areas of research: (1) efficiency and accuracy of human annotation (in particular, inter-annotator agreement), (2) quality assurance in crowdsourcing, and (3) inherent ambiguity in natural language annotation tasks. We survey these in turn below.
2.1 Evaluating Human Annotation
Annotated natural language data is invaluable to machine learning techniques and algorithms. Two major questions arise as to how to optimize data collection task and how to measure the accuracy and reliability of the resulting data.
Many authors emphasize the importance of careful task design, guideline creation, and pre-testing as seen in [Compton et al., 2012, Compton et al., 2012, Pustejovsky and Stubbs, 2012]. Such tests assist in improving inter-annotator reliability — the extent to which the results of an annotation task represent replicable unbiased judgments. Several reliability estimates correct for this, see especially Krippendorff’s [Krippendorff, 2004].
Once data is collected, it is standard practice to take Inter-Annotator Agreement (IAA) to provide an estimate of the correctness or validity of human annotations ([Smyth et al., 1995]). IAA is distinct from the reliability metric: the former cannot account for biases that might be shared by a group and hence lead to high agreement across annotators. There is no consensus on how to measure IAA, but one common metric is the Kappa () statistic (or some variant), which measures pairwise agreement among a set of annotators making category judgments and correcting for expected chance agreement:
,
where is the proportion of times that the annotators agree and is the proportion of times that we would expect them to agree by chance.
This measure has been shown to be effective, but it also has certain limitations [Eugenio and Glass, 2004, Pustejovsky and Stubbs, 2012, Artstein, 2017]. See also [Craggs and Wood, 2005, Passonneau and Carpenter, 2014, Kara et al., 2018] for some less commonly used measures.
In addition to the ongoing debate about the accuracy and correctness of the and statistics as proxies for inter-annotator reliability and agreement respectively, the metrics provide only a partial picture of human-annotated data. The statistic is unable to account for the level of expertise of the annotators, for example. In a meta study, [Bayerl and Paul, 2011] identify several factors that may influence IAA: annotation domain, number of categories in a coding scheme, number of annotators in a project, whether annotators received training, the intensity of annotator training, the annotation purpose, and the method used for calculating IAA. One of their key conclusions is that complex annotation tasks should be simplified to use fewer grading categories. [Bayerl and Paul, 2011] additionally suggest using expert annotators who receive specialized training and collecting a large number of annotations for critical tasks.
The use of expert annotators versus naive ones is a contested issue. While some authors suggest using experts or adjudicators [Kowtko et al., 1993], Snow et al. [Snow et al., 2008] argue that many labeling tasks can be effectively designed and implemented using non-expert annotators recruited via crowd-sourcing tools.
2.2 Quality Assurance in Crowdsourcing
Many solutions exist to resolve the low-quality contributions that come either from deliberate scamming or poor performance [Ipeirotis et al., 2010, Dow et al., 2012, Ipeirotis et al., 2014, Gadiraju et al., 2015]. The most commonly adopted technique for quality assurance relies on a set of questions with known correct answers, often referred to as gold standard test questions, which are randomly injected into the workers’ sessions in order to measure their performance [Checco et al., 2018, Sun and Dance, 2012, El Maarry et al., 2014]. In addition, since the original Dawid-Sekene [Dawid and Skene, 1979] crowdsourcing model based on maximum likelihood estimation, several improvements have been proposed: the Fast Dawid-Skene model for improved speed of convergence [Sinha et al., 2018], the effort to extend to multiclass labeling from binary labeling [Li and Yu, 2014], and the Minmax Entropy method [Zhou et al., 2012] for accuracy improvements. There are also other comparable models [Raykar et al., 2010, Welinder et al., 2010, Karger et al., 2011], but none could outperform Dawid-Skene [Liu et al., 2012].
However, such techniques share several clear limitations:
-
1.
Scalability Challenge: the curation of gold standard test questions, or the initialization for ground truth in maximum likelihood estimation methods, often demands significant effort and resources (as many as 30 identical batch of graders per question assignment), and therefore such question sets are usually small in size [Oleson et al., 2011];
-
2.
Measurement by Proxy: the gold standard test questions only measure the annotator performances on the given question set, and the results do not necessarily reflect their performances on the actual tasks; further, the question set may differ significantly in terms of style or difficulty from the requests in the actual annotation task, leading to greater deviation from the intended measurement;
-
3.
Vulnerability to Exploitation: the inherent limit on the size of the gold standard test question set makes the method prone to a coordinated attack from even a small group of annotation workers, once they can collude and identify the gold questions; Checco et al. [Checco et al., 2018] have proposed an attack mechanism targeting the gold question-based quality assurance method.
2.3 Inter-Annotator Agreement and Ambiguity
This work is in part a continued discussion on inter-annotator (dis)agreement and its implications. As suggested by many [Dumitrache et al., 2018b, Cheatham and Hitzler, 2013, Plank et al., 2014, Hollenstein et al., 2016], the lack of consensus among a group of annotators on a specific task can be attributed to one or more of the following causes: inherent uncertainty in the relevant domain knowledge, true natural language ambiguity, and annotation error. Here we do not attempt to provide a comprehensive interpretation of inter-annotator disagreement. Instead, we propose an annotation strategy that efficiently preserves material inter-annotator disagreement while reducing the impact of annotation errors on consensus extraction. We then share findings from a text classification task which uses this strategy and examine some specific cases of inter-annotator disagreement. We show that the DACR strategy can assist in identifying material disagreements — ones which stem from issues with domain knowledge or the ontology used in the classification task — while reducing disagreements related to simple human error.
3 Latent Inaccuracy Estimation Method (LIEM)
This section presents a mathematical formula for measuring the inaccuracy of a human annotation task.
3.1 LIEM: a brief description
The LIEM treats a given annotation process as a black box system with the following assumptions:
-
•
Singular ground truth label: each request has a singular ground truth label;
-
•
Absence of systematic confusion: the annotation process yields incorrect labels from a discrete uniform distribution and doesn’t contain systematic confusions;
-
•
Memoryless behavioral stability: the annotation process is not affected by a priori events, and exhibits stable behavior over time.
Given these assumptions, we may assign labels to a set of requests using the same predetermined annotation process in two duplicate annotation projects, and compute the proportion of requests that received identical labels, . The latent accuracy may then be expressed as:
,
with decreasing margin of error for larger values of .
3.2 LIEM: a detailed derivation
In the remainder of this section we explicate this result in detail. To start, consider a project with an unspecified annotation task for a set of requests . Let us assume each request has a singular ground truth label , where is a finite set of all permissible labels in the annotation task.111Although we acknowledge potential ambiguity and existence of multiple correct labels, this stipulation allows us to study the impact of annotation inaccuracy for the prevalent choice of singular ground truth label in semantic annotation tasks.
3.3 Expected performance of a black box system
Each request undergoes a predetermined annotation process, which returns a label . In the crowdsourcing context, the returned label is extracted from the answer(s) provided by one or more randomly selected annotation workers. Therefore, if we treat the predetermined annotation process as a black box system and focus on its input-output relationship, the system is expected to behave inconsistently in terms of its performance (that is, its ability to yield the correct answer such that ) due to (1) varying difficulty levels of each request and (2) non-uniformity of performance both amongst annotation workers and for any individual worker over different requests.
Let be a continuous random variable over associated with the performance of the predetermined annotation process. The performance of the system for each request is defined as the likelihood of yielding the correct label, i.e. . Accordingly, we consider the performance of the system, , for each request at time to be an independent random draw from . We denote the expected performance of the system as , the value we intend to estimate.
3.4 Central limit theorem and Bhatia–Davis inequality
Next, we briefly discuss two important concepts of probability: the Central Limit Theorem (CLT) and Bhatia-Davis inequality. More complete introductions can be found in [Barron and others, 1986, Bhatia and Davis, 1993]. Without loss of generality, we may treat the performances of a black box system over a set of annotation requests (where ) as a random sample of size from the given distribution of , with expectation and variance . Such a set of random variables are independent, identically distributed. Conveniently, then, the sample mean:
,
is the aggregated distribution of the proportion of correctly labeled requests in . Based on CLT, we conclude that the mean of a sufficiently large random sample from an arbitrary distribution to have approximately normal distribution such that .222The larger the sample size, the better the CLT approximation becomes. With the help of the Bhatia-Davis inequality, we may write the upper bound on the variance of any bounded distribution on the real line as:
,
where , are the maximum and minimum of the distribution. For with we have:
.
The method to estimate the accuracy of the predetermined annotation process is trivial in the case when is a gold standard question set so that the proportion of correctly labeled requests is directly observable. Since the premise of this work is to estimate accuracy without relying on gold questions, we continue on to next section.
3.5 Consistency as a reliable proxy
Turning back to our labeling task, in order to find a reliable proxy variable, we examine the annotation process for a single request in detail. Given the finite set of all permissible labels in the annotation task, , for request with ground truth label , the likelihood of receiving the correct label such that is previously defined as the performance of the annotation process on , i.e. .
Accordingly, the probability of not receiving a correct label is . Now, let us assume that when the annotation process (treated as a black box system) does not yield the correct label, it simply returns one of the remaining labels in the sample space, i.e. , with equal probability333[Giancola et al., 2018] suggest that mislabeling is not necessarily follow a discrete uniform distribution, and systematic confusions (aptly named ”class permutations”) are observed among the annotation workers on several tasks including text clustering, rare class simulation, and dense image segmentation. However, their findings from the text clustering task also suggest that such systematic confusions are relatively less prevalent in the given task. We assume a discrete uniform distribution as a necessary approximation to avoid intractability..
Consider a request undergoing the said annotation process twice, with the system’s performances at , yielding labels respectively. With the help of Bayes’ theorem, we may write the probability of receiving two identical labels as:
,
from which we calculate:
,
where .
Further, with reasonably good faith in the system performances such that and a sufficiently large sample space (e.g. ), we may allow for further approximation given a significantly smaller second term444Essentially, the chances of receiving identical labels from two independent annotation sessions on the same request are negligible when both labels are incorrect. as and . This approximation enables us to simplify the expression for to:
,
based on which we may express the probability distribution of receiving identical labels for a request in two independent tries from a predetermined annotation process as a product distribution,
,
with its mean and variance .555The mathematical details to derive the variance of the product of two independent random variables can be found in [Goodman, 1960]
3.6 Duplicate annotation: measuring consistency
To measure consistency at scale, we set up two annotation projects with an identical set of requests . This yields two sets of labels, . We may then treat the consistency observations as a random sample of size from a distribution , with expectation and variance . Assuming that they are independent, identically distributed, according to Central Limit Theorem the distribution of sample mean converges to a normal distribution when the sample size is large enough:
,
where by Bhatia–Davis inequality, we know , so that for . Accordingly, we may establish a (loose) upper bound for the variance of as:
,
from which we find that as , i.e. by law of large numbers, the observed sample mean becomes closer to as the sample size increases.
Meanwhile, the established upper bound helps us to properly capture the margin of error in estimating for a not-so-large size of . For example, with we have ; to maintain a confidence level of a three-sigma effect (99.73% confidence) we would estimate to be .
3.7 LIEM: a summary
In sum, the Latent Inaccuracy Estimation Method treats a given annotation process as a black box system and assumes that (a) each request has a singular ground truth label; (b) the annotation process yields incorrect labels from a discrete uniform distribution and doesn’t contain systematic confusions; and (c) the annotation process is not affected by a priori events, and exhibits stable behavior over time.
Therefore, if we assign labels to a set of requests using the same annotation process in two duplicate projects, we may compute the proportion of requests that receive identical labels, . The latent accuracy of the annotation process may then be expressed as: , with decreasing margin of error for larger values of .
4 A Look into Different Annotation Strategies
With this result in hand, we examine three commonly used annotation strategies and evaluate their respective performances through agent based simulations. We compare these results to our novel methodology, Dynamic Automatic Conflict Resolution (DACR), which we introduce immediately below. We establish that DACR outperforms other common approaches to human annotation. In a subsequent section we will show that DACR indeed outperforms these other approaches in an actual annotation task.
4.1 Annotation Strategies
We compare the following three commonly used annotation strategies:
-
1.
One-grader: each request is assigned to a randomly selected grader and the answer label is recorded;
-
2.
Double Graded, Conflict Resolved (DG, CR): each request is first assigned to two randomly selected graders; if the two agree, the consensus answer label is recorded; otherwise (in the case of a conflict), the request is escalated to an expert annotator with higher expected performance, and the expert’s answer label is recorded as the final answer;
-
3.
N-Graded: each request is assigned to N randomly selected graders; the answer label with a simple majority (greater than ) is recorded as the final answer. If no answer label with a simple majority exists, the request is assigned an in-conflict label.
Against these approaches, we propose a novel approach: Dynamic Automatic Conflict Resolution (DACR) with maximum grades: each request is first assigned to two randomly selected annotation workers; if the two agree, the consensus answer label is recorded; otherwise (in the case of a conflict), the request is iteratively assigned to one additional worker until (a) a simple majority answer label (from all historical answers) is produced, or (b) the maximum grades is reached and the annotation process ends (regardless of the presence of a majority label). A separate process can then apply to determine how to resolve in-conflict requests.
Parameter | Initialization | Count |
---|---|---|
Annotation Workers | 100 | |
Expert Annotators | 20 | |
Request Difficulty Levels | Varied |
4.2 Simulation Configurations
Our simulation experiments vary the following parameters of the crowdsourcing annotation task: (1) the quality of the annotation workers, (2) the quality of expert annotators (if any), (3) the difficulty levels of the requests in the annotation task, and (4) the annotation strategy in use.
As detailed in Table 1, the parameters are initialized666We also simulated with varying parameter values and the subsequent findings are established consistently. with reasonable estimates from empirical observations. Accordingly, the annotation act is abstracted as follows: when an annotator with capability annotates a request with difficulty level , the correctness of the answer label is generated as the outcome of a single Bernoulli trial, with , and .
4.3 Experiment design and findings
Our simulation experiments investigate (1) the performance and efficiency of the competing annotation strategies outlined above, and (2) the validity of the proposed LIEM approach in estimating the expected performance of a given annotation strategy.
We ran 100 simulated annotation projects for each annotation strategy; each project consisted of the same batch of randomly initialized 10,000 requests. We plot the distribution of the annotation accuracy over the projects for each strategy in Figure 1.

We conclude in Table 2 that the DACR strategy is consistently ranked top 2 compared to its competitors along all three metrics listed. We furthermore find that the DACR strategy requires a significantly lower number of grades to achieve comparable performance to that produced by a 7-graded strategy (63.6% lower than this top competitor), and achieves improved performance with a small incremental number of grades compared to the strategy (9.0% higher accuracy compared to this top competitor).
Strategy | Avg. Grades | ||
---|---|---|---|
One-grader | 80.9% | 10,000 | |
DG, CR | 94.0% | 23,382 | |
N-grader (N=5) | 93.4% | 50,000 | |
N-grader (N=7) | 95.4% | 70,000 | |
DACR (Max = 5) | 94.9% | 25,495 |
To validate the LIEM approach, we exhaustively compute the inter-project consistency rates and the corresponding pair’s mean accuracy, for each annotation strategy. The data points are plotted in Figure 2, where we conclude with simple curve fitting that LIEM’s proposed estimation method, , is a reasonably accurate approach to find the latent accuracy for a variety of annotation strategies.

5 Exploring Ambiguity with DACR
As a part of a greater effort to build an intelligent virtual assistant, we have long been working on a text classification task. The text classifier predicts user intent for a given utterance according to a predetermined taxonomy, as in Table 3. In our task, over 60 possible intent labels are available and we stipulate that only one correct intent label exists for any given utterance. In the remainder of this paper, we introduce the results from a DACR project implemented at scale for a commercially used assistant.
User Utterance | Intent Class |
---|---|
Davos local time | Time |
Status for AA 166 | Flight |
Compute 37 times 30 | Math |
Kesha album Rainbow | Music |
5.1 Results from DACR
Having examined DACR’s (superior) performance via simulation experiments, we deployed it to our internal crowdsourcing platform for annotation on a text classification task. The project consisted of 60,000 randomly sampled utterances and used the DACR annotation strategy, with the minimum grade count set to 3 and the maximum set to 9. At the completion of the project, we recorded an average of 3.74 grades per utterance. Additionally, 5.00% of the utterances did not receive a consensus label even when the maximum grade count of 9 had been reached (we will refer to such cases as “in-conflict”).777In simulation, we record 2.55 average grades per utterance with and , and an in-conflict rate of 5.1%.
To determine whether such in-conflict requests are a result of inherent ambiguity or poor performance of annotation workers, we construct a confusion matrix based on the inter-annotator disagreements over the set of requests with no consensus label. As we observe from Figure 3, although we omit the label names to preserve privacy — there is a clear pattern where a small number of specific intent labels are significantly more likely to be confused with a small set of select other labels.888Notice the existence of highlighted blocks in Figure 3. Such highlighted blocks correspond to unusually high levels of inter-category confusion, and are to be rarely observed in the confusion matrix if no systematic bias exists. The confusion matrix is normalized by adjusting for background frequency of each category.
Accordingly, the observed systematic bias indicates the presence of one or several of following factors: (1) ambiguity or vagueness in the annotation guidelines, (2) inherent ambiguity in the ontology design, or (3) inherent ambiguity in a few select types of utterances.

An examination of the in-conflict cases has led us to conclude that in this study, (2) and (3) were sizable contributing factors. As a result, we have been able to undertake a series of directed efforts aimed at educating the annotators on the one hand, and better-defining the ontology on the other, to reduce conflicts caused by reason (2). On the other hand, we do not expect that any effort on our part can resolve conflicts casued by reason (3). Instead, the prevalence of truly ambiguous utterances in our study provides yet more evidence that the common insistence on a single ground truth label for each request may simply be misguided. In other words, such an assumption makes it inherently impossible to correctly represent 100% of natural language utterances.
Acknowledging the existence of ambiguity has two profound implications for nearly all annotation tasks that conforms to the premises of this study:
-
•
Information loss: as [Dumitrache et al., 2018a] have previously suggested, the rejection of ambiguity results in a loss of information that would impact a wide range of supervised learning applications and prevents the development of appropriate model architectures;
-
•
Measurement error: we find it equally concerning that if we are to generate a test set with annotation strategies that ignores ambiguity (such as one-grader, or DG,CR), we would have deliberately introduced noise into the test set by introducing ambiguous examples with an uncertain partial label. This omission would lead to increased measurement error on model performance.999Consider a test set with 5% ambiguous examples for which no (single) ground truth exists. Then even the oracle would be measured to have a 5% error rate on this test set.
6 Conclusion
In this paper, we have introduced a method to estimate the latent inaccuracy of any crowdsourcing annotation process and shown its validity both via formal analysis and simulation experiments. We also investigated the performances and efficiencies of different annotation strategies with an agent-based model. Based on our findings, we further propose a novel annotation strategy that dynamically allocates additional annotation resources to difficult (or ambiguous) requests; it is found to be able to preserve multiple conflicting grades for ambiguous requests and outperform comparable annotation strategies while requiring significantly less annotation resources.
We share insights from a DACR-based annotation project on a classification task and discover patterns among the in-conflict examples that corroborate the existence of ambiguity in the given task. Our goal is to allow the practitioners in the field of machine learning to easily estimate the latent accuracy of any crowdsourcing annotation process. We further hope many will find DACR to be a better alternative to their existing annotation strategies and start thinking about how we should deal with inherent ambiguity.
References
- [Artstein, 2017] Ron Artstein. 2017. Inter-annotator agreement. In Handbook of linguistic annotation, pages 297–313. Springer.
- [Barron and others, 1986] Andrew R Barron et al. 1986. Entropy and the central limit theorem. Ann. Prob., 14(1):336–342.
- [Bayerl and Paul, 2011] Petra Saskia Bayerl and Karsten Ingmar Paul. 2011. What determines inter-coder agreement in manual annotations? a meta-analytic investigation. Computational Linguistics, 37(4):699–725.
- [Bhatia and Davis, 1993] Rajendra Bhatia and Chandler Davis. 1993. More matrix forms of the arithmetic-geometric mean inequality. SIAM Journal on Matrix Analysis and Applications, 14(1):132–136.
- [Cheatham and Hitzler, 2013] Michelle Cheatham and Pascal Hitzler. 2013. String similarity metrics for ontology alignment. In International semantic web conference, pages 294–309. Springer.
- [Checco et al., 2018] Alessandro Checco, Jo Bates, and Gianluca Demartini. 2018. All that glitters is gold—an attack scheme on gold questions in crowdsourcing. In Sixth AAAI Conference on Human Computation and Crowdsourcing.
- [Compton et al., 2012] D’Lane Compton, Tony P Love, and Jane Sell. 2012. Developing and assessing intercoder reliability in studies of group interaction. Sociological Methodology, 42(1):348–364.
- [Craggs and Wood, 2005] Richard Craggs and Mary McGee Wood. 2005. Evaluating discourse and dialogue coding schemes. Computational Linguistics, 31(3):289–296.
- [Dawid and Skene, 1979] Alexander Philip Dawid and Allan M Skene. 1979. Maximum likelihood estimation of observer error-rates using the em algorithm. Journal of the Royal Statistical Society: Series C (Applied Statistics), 28(1):20–28.
- [Dow et al., 2012] Steven Dow, Anand Kulkarni, Scott Klemmer, and Björn Hartmann. 2012. Shepherding the crowd yields better work. In Proceedings of the ACM 2012 conference on computer supported cooperative work, pages 1013–1022. ACM.
- [Dumitrache et al., 2018a] Anca Dumitrache, Lora Aroyo, and Chris Welty. 2018a. Capturing ambiguity in crowdsourcing frame disambiguation. In Sixth AAAI Conference on Human Computation and Crowdsourcing.
- [Dumitrache et al., 2018b] Anca Dumitrache, Oana Inel, Benjamin Timmermans, Carlos Ortiz, Robert-Jan Sips, Lora Aroyo, and Chris Welty. 2018b. Empirical methodology for crowdsourcing ground truth. arXiv preprint arXiv:1809.08888.
- [El Maarry et al., 2014] Kinda El Maarry, Wolf-Tilo Balke, Hyunsouk Cho, Seung-won Hwang, and Yukino Baba. 2014. Skill ontology-based model for quality assurance in crowdsourcing. In International Conference on Database Systems for Advanced Applications, pages 376–387. Springer.
- [Eugenio and Glass, 2004] Barbara Di Eugenio and Michael Glass. 2004. The kappa statistic: A second look. Computational linguistics, 30(1):95–101.
- [Gadiraju et al., 2015] Ujwal Gadiraju, Ricardo Kawase, Stefan Dietze, and Gianluca Demartini. 2015. Understanding malicious behavior in crowdsourcing platforms: The case of online surveys. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems, pages 1631–1640. ACM.
- [Giancola et al., 2018] Michael Giancola, Randy Paffenroth, and Jacob Whitehill. 2018. Permutation-invariant consensus over crowdsourced labels. In Sixth AAAI Conference on Human Computation and Crowdsourcing.
- [Goodman, 1960] Leo A Goodman. 1960. On the exact variance of products. Journal of the American statistical association, 55(292):708–713.
- [Hollenstein et al., 2016] Nora Hollenstein, Nathan Schneider, and Bonnie Webber. 2016. Inconsistency detection in semantic annotation. In Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC 2016), pages 3986–3990.
- [Ipeirotis et al., 2010] Panagiotis G Ipeirotis, Foster Provost, and Jing Wang. 2010. Quality management on amazon mechanical turk. In Proceedings of the ACM SIGKDD workshop on human computation, pages 64–67. ACM.
- [Ipeirotis et al., 2014] Panagiotis G Ipeirotis, Foster Provost, Victor S Sheng, and Jing Wang. 2014. Repeated labeling using multiple noisy labelers. Data Mining and Knowledge Discovery, 28(2):402–441.
- [Kara et al., 2018] Yunus Emre Kara, Gaye Genc, Oya Aran, and Lale Akarun. 2018. Actively estimating crowd annotation consensus. Journal of Artificial Intelligence Research, 61:363–405.
- [Karger et al., 2011] David R Karger, Sewoong Oh, and Devavrat Shah. 2011. Iterative learning for reliable crowdsourcing systems. In Advances in neural information processing systems, pages 1953–1961.
- [Kowtko et al., 1993] Jacqueline C Kowtko, Stephen D Isard, and Gwyneth M Doherty. 1993. Conversational games within dialogue. In University of Edinburgh. Citeseer.
- [Krippendorff, 2004] Klaus Krippendorff. 2004. Reliability in content analysis. Human communication research, 30(3):411–433.
- [Li and Yu, 2014] Hongwei Li and Bin Yu. 2014. Error rate bounds and iterative weighted majority voting for crowdsourcing. arXiv preprint arXiv:1411.4086.
- [Liu et al., 2012] Qiang Liu, Jian Peng, and Alexander T Ihler. 2012. Variational inference for crowdsourcing. In Advances in neural information processing systems, pages 692–700.
- [Oleson et al., 2011] David Oleson, Alexander Sorokin, Greg Laughlin, Vaughn Hester, John Le, and Lukas Biewald. 2011. Programmatic gold: Targeted and scalable quality assurance in crowdsourcing. In Workshops at the Twenty-Fifth AAAI Conference on Artificial Intelligence.
- [Passonneau and Carpenter, 2014] Rebecca J Passonneau and Bob Carpenter. 2014. The benefits of a model of annotation. Transactions of the Association for Computational Linguistics, 2:311–326.
- [Plank et al., 2014] Barbara Plank, Dirk Hovy, and Anders Søgaard. 2014. Learning part-of-speech taggers with inter-annotator agreement loss. In Proceedings of the 14th Conference of the European Chapter of the Association for Computational Linguistics, pages 742–751.
- [Pustejovsky and Stubbs, 2012] James Pustejovsky and Amber Stubbs. 2012. Natural Language Annotation for Machine Learning: A guide to corpus-building for applications. ” O’Reilly Media, Inc.”.
- [Raykar et al., 2010] Vikas C Raykar, Shipeng Yu, Linda H Zhao, Gerardo Hermosillo Valadez, Charles Florin, Luca Bogoni, and Linda Moy. 2010. Learning from crowds. Journal of Machine Learning Research, 11(Apr):1297–1322.
- [Sinha et al., 2018] Vaibhav B Sinha, Sukrut Rao, and Vineeth N Balasubramanian. 2018. Fast dawid-skene: A fast vote aggregation scheme for sentiment classification. arXiv preprint arXiv:1803.02781.
- [Smyth et al., 1995] Padhraic Smyth, U Fayyad, M Burl, P Perona, and P Baldi. 1995. Learning with probabilistic supervision. Computational learning theory and natural learning systems, 3:163–182.
- [Snow et al., 2008] Rion Snow, Brendan O’Connor, Daniel Jurafsky, and Andrew Y Ng. 2008. Cheap and fast—but is it good?: evaluating non-expert annotations for natural language tasks. In Proceedings of the conference on empirical methods in natural language processing, pages 254–263. Association for Computational Linguistics.
- [Sun and Dance, 2012] Yu-An Sun and Christopher Dance. 2012. When majority voting fails: Comparing quality assurance methods for noisy human computation environment. arXiv preprint arXiv:1204.3516.
- [Welinder et al., 2010] Peter Welinder, Steve Branson, Pietro Perona, and Serge J Belongie. 2010. The multidimensional wisdom of crowds. In Advances in neural information processing systems, pages 2424–2432.
- [Zhou et al., 2012] Dengyong Zhou, Sumit Basu, Yi Mao, and John C Platt. 2012. Learning from the wisdom of crowds by minimax entropy. In Advances in neural information processing systems, pages 2195–2203.