This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

11institutetext: Universidad ORT Uruguay, Montevideo, Uruguay
11email: [email protected]
22institutetext: CNRS Univ Lyon INSA Lyon UCBL, LIRIS UMR5205, Lyon, France 22email: [email protected]
33institutetext: Facultad de Ingeniería, Universidad de la República, Montevideo, Uruguay
33email: [email protected]

A Novel Method for Curating Quanti-Qualitative Content through consensus-based versioning Control

Alejandro Adorjan 11 0000-0002-5257-284X    Genoveva Vargas-Solar 22 0000-0001-9545-1821    Regina Motz 33 0000-0002-1426-562X
Abstract

This paper proposes a Researcher-in-the-Loop (RITL) guided content curation approach for quanti-qualitative research methods that uses a version control system based on consensus. The paper introduces a workflow for quanti-qualitative research processes that produces and consumes content versions through collaborative phases validated through consensus protocols performed by research teams. We argue that content versioning is a critical component that supports the research process’s reproducibility, traceability, and rationale. We propose a curation framework that provides methods, protocols and tools for supporting the RITL approach for managing the content produced by quanti-qualitative methods. The paper reports a validation experiment using a use case about the study on disseminating political statements in graffiti.

Keywords:
Data curation content versioning researcher-in-the-loop group consensus protocols.

1 Introduction

Quanti-qualitative methods are applied in social and human sciences data-driven research projects. In this context, the production and validation of generated content and knowledge are determined by consensus protocols performed collaboratively by team research [18]. Research methods with quanti-qualitative perspectives must curate the content flowing along the research tasks, the processes performed (defining research questions, stating, calibrating and validating a theoretical framework, converging to a corpus, defining data collection tools) and the context and conditions in which content is produced (interviews, focus group, research team decision-making criteria and consensus protocols).

The emergence of data science has led to the use of algorithms to perform specific research tasks like automatically processing surveys using statistics and natural language processing. The conditions in which these automatic processes are used and produce results must be assessed, discussed and consensually validated. Related logs gathering information about the assessment, validation and agreement processes must be curated too. Consider the project willing to classify graffiti images harvested in the downtown area of city to identify those with political content. The objective would be to determine whether anti-system political content can be disseminated as hidden messages in artistic graffiti. In the initial phase of a quanti-qualitative project, a team of scientists would synchronise a specification document with “rules” specifying the criteria for choosing people participating in the harvesting process, the area where pictures will be collected and the research question to target. Beyond collecting graffiti pictures, the team would need to maintain the specification documents, information about the people harvesting the data, and the criteria agreed to define the conditions for this harvesting. Moreover, the team would maintain data related to the consensus process that allowed them to decide on the conditions in which graffiti harvesting is done. The consensus process might not be a one-shot task but might need several agreement rounds, during which versions of the specification documents are produced.

Data curation in the context of the previous example maintains the versions of the content and the comments produced by the research team tagging the documents. The curation result is datasets tracing content about the object of study (i.e., graffiti pictures in a city’s downtown) and about the rules agreed upon by a research team to perform graffiti harvesting and information about the consensus process the team followed to make agreements. Curating research artefacts 111We based the notion of artefact according to ACM definition “a digital object that was either created by the authors to be used as part of the study or generated by the experiment itself”, https://www.acm.org/publications/policies/artefact-review-and-badging-current. focusing on the scientific workflow aspects are relevant for reproducibility and improving sciences’ reliability [22]. FAIR (Findable, Accessible, Interoperable and Reusable) principles provide guidelines for scientific data reuse [23]. The FAIR guidelines do not fully provide a framework for systematically exploring, curating and reusing research artefacts. Moreover, scientific publication repositories supporting FAIR principles do not answer questions such as: How do artefacts evolve throughout the research project stages? What software tools and algorithms did researchers rely on? What is the researchers’ narrative concerning applying Artificial Intelligence (AI) algorithms? In the previous example, using AI models for classifying graffiti would require producing explanations on the choice of the algorithms, the principle and the interpretation of their quantitative results, and even the back-and-forth training and calibration phases until the results converge.

This paper introduces a RITL workflow for quanti-qualitative research processes. The workflow produces and consumes content versions through collaborative phases validated through consensus protocols performed by research teams. We argue that content versioning is a critical component that supports the research process’s reproducibility, traceability, and rationale. We propose a curation framework that provides methods, practices, and tools for supporting the RITL approach for managing the content produced by quanti-qualitative methods.

The contribution of the paper concerns mainly the iterative production and validation of content using consensus protocols performed under RITL settings. Therefore we propose:

  1. 1.

    A versioning approach to keep track of the versions of content.

  2. 2.

    Consensus protocols to control the versioning and track the progress of projects across different stages.

The remainder of the paper is organized as follows. Section 2 introduces the background and related work. Section 3 presents a general overview of our proposed content-curation approach for research artefacts produced in quanti-qualitative data-driven research. Section 4 introduces the RITL-based versioning control model based on consensus protocols that define the management of the curated content. Section 5 presents the experimental validation based on two use cases with different content production and validation strategies. Finally, Section 6 concludes the paper and discusses future work.

2 Background and Related Work

Theoretical and methodological frameworks are the basis for sustaining (quali)-qualitative research projects. Quantitative methods study observations of phenomena collected a priori and promote a deductive approach for driving conclusions. For example, analyses numerical data collection methods [18], statistics, machine learning and other artificial intelligence models. On the other hand, qualitative approaches explore complex phenomena. They promote inductive techniques that rely on nonstatistical and nonnumerical data collection methods, analysis, and evidence production [4].

Quanti-quantitative approaches generally incorporate both methods, addressing complex research questions more accurately due to generalizability, contextualisation, and credibility [18]. Quanti-qualitative research methodologies usually apply “ad-hoc” data curation strategies that keep track of the data that describe the tools, techniques, hypothesis, and data harvesting criteria. However, these strategies (if present) usually are not standardised nor systematised. In the next lines we give a summarised vision existing data curation techniques and software used to qualitatively curate content.

Content curation.

According to Garcov et al. [9], research data curation is defined as “the act of preparing research data and artefacts for sharing and long-term preservation”. Research repositories are the standard for publishing data collections to the research communities. Datasets at an early collection stage are generally not ready for analysis or preservation. Thus, extensive preprocessing, cleaning, transformation, and documentation actions are required to support usability, sharing, and preservation over time [11]. We assume that data curation consists of identifying, systematizing, managing, and versioning research data, considering versioning artefacts an essential component of tracking changes along the research project. Curated data collections have the potential to drive scientific progress [25]. Research artefacts curation is relevant for reproducibility and improves the reliability of sciences [22]. However, data curation introduces challenges for supporting data-driven applications [7] adopting quanti-qualitative methods. For example, research challenges curating material across time, space and collaborators [22]. Quantitative and qualitative research methodologies apply ad-hoc data curation strategies that keep track of the data that describe the tools, techniques, hypothesis, and data harvesting criteria defined a priori by a scientific team.

Qualitative Data Processing Software.

Several software tools that apply statistical techniques and machine learning algorithms are available for qualitative researchers. Woods et al. [24] argue that Computer-Assisted Qualitative Data Analysis Software (CAQDAS) is a well-known tool for qualitative research. These tools support qualitative techniques and methods for applying Qualitative Data Analysis (QDA). ATLAS.ti [2], Dedoose [6], MAXQDA [20], NVivo [14] implement the REFI-QDA standard, an interoperability exchange format. CAQDAS [5] researchers and practitioners can perform annotation, labelling, querying, audio and video transcription, pattern discovery, and report generation. Furthermore, CAQDAS tools allow the creation of field notes, thematic coding, search for connections, memos (thoughtful comments), contextual analysis, frequency analysis, word location and data analysis presentation in different reporting formats [8]. The REFI-QDA (Rotterdam Exchange Format Initiative) 222https://www.qdasoftware.org the standard allows the exchange of qualitative data to enable reuse in QDAS [10]. QDA software such as ATLAS.ti [2], Dedoose [6], MAXQDA [20], NVivo [14], QDAMiner [15], Quirkos [16] and Transana [21] adopt REFI-QDA standard.

Discussion.

The researcher’s intervention, defined as RITL by [19], is a crucial aspect of human intervention to assess content concerning (i) the conditions in which it is produced and (ii) to make decisions about the new tasks to perform and the way a research project will move forward. Researcher in the loop (RITL) is a case of Human-in-the-loop (HITL), where the primary output of the process is a selection of the data, not a trained machine learning model. HITL is crucial for handling supervision, exception control, optimization, and maintenance [17, 13]. Under a RITL approach, a human sees all data points in the relevant selection at the end of the process. Using RITL requires a systematic solid way of working 333https://hai.stanford.edu/news/humans-loop-design-interactive-ai-systems. This characteristic is critical for designing content curation for quanti-qualitative research methods.

Scientific content should be extracted and computed, including data, analytics tasks (manual and AI models), and associated metadata. This curated content allows the produced knowledge to be reusable and analytics results to be reproducible [12], thereby adhering to FAIR (Findable, Accessible, Interoperable and Reusable) principles [3].

3 Curating content produced in quanti-qualitative data-driven research

We propose a workflow for modelling (quanti)-qualitative research methods. It is an spiral Γ\Gamma research process for each phase of a quanti-qualitative research approach:

- Γ1\Gamma_{1} Problem statement refers to the theory review stage, formulation of research questions, definition of methodologies, and construction of the theoretical framework.
- Γ2\Gamma_{2} Data acquisition is the phase devoted to data collection, exploration, cleaning, and reliability verification.
- Γ3\Gamma_{3} Data management refers to metadata generation, evaluation, and contextualization.
- Γ4\Gamma_{4} Analysis consists of a round of experiments and measurements, incorporating debates and reflections on the results of previous stages.
- Γ5\Gamma_{5} Reporting is a phase devoted to visualization, evaluation, writing process, and final publication of scientific work.

In these phases, research artefacts with content are produced and versioned, guided by consensus-based decision-making by the research team. Every phase and its stages adopt a spiral process where artefacts are produced through research actions done within fieldwork activities. For certain activities, research teams agree to use automatic models for extracting meta-data and generating quantitative insight into the content using digital solutions such as algorithms, Mathematical models and text processing tools.

A research team RR conducts qualitative activities along these phases and produce content, such as exploratory interviews, weekly meetings reports, manual generation of analytical coding, and analytical coding of the entire process. We propose a model representing the content produced during a quati-qualitative research workflow. This content is modelled as artefacts with an associated life cycle: creation, annotation, versioning and validation. A general view of the concepts of the artefact model is given next.

3.1 Artefact model

Let us define an artefact α\alpha as a concept modelling a document like text documents, records, videos or images.

The artefact α\alpha can contain:
- Surveys, interviews, codebooks, field diaries, that are data harvesting “tools” used in qualitative methods.
- Descriptions of data harvesting protocols with criteria adopted for choosing focal groups, people applying the tools, objectives, and analysis protocols.
- Documents specifying theoretical framework, research questions, bibliography, corpora, and quantitative and categorical analysis results.
- Information describing the context of the (quanti)-qualitative research, including the description of the research team, the conditions in which decision-making is done (agreements, consensus) along the research phases and the provenance of results and agreements.
- Narratives, multimedia content and technical meta-data (e.g., structure, format, document size) associated with the artefacts.

As stated in the data structure (1) below, let us denote α\alpha as the artefact created at time t represented by a timestamp, in a workflow phase Γi\Gamma_{i} of a project Πj\Pi_{j} by a researcher identified with an IdId. Let μ\mu represent metadata and η\eta represent a narrative annotation (both μ\mu and η\eta are artefacts). An artefact α\alpha has associated metadata and narratives.

α:content: Document,producer: ResearcherId,timestamp: Date,projectWfPh:  ΓiΠ  metaData: [μ],listOfTags: [η],listOfActions: [λ\alpha:\begin{tabular}[]{lcl}&&\\ &$\langle$&content: Document,\\ &&producer: ResearcherId,\\ &&timestamp: Date,\\ &&projectWfPh: $\langle$ $\Gamma_{i}$, $\Pi$ $\rangle$ $\rangle$\\ &&metaData: [$\mu$],\\ &&listOfTags: [$\eta$],\\ &&listOfActions: [$\lambda$] $\rangle$\end{tabular} (1)

Metadata. An artefact α\alpha (e.g., a document), by default, brings metadata μ\mu (title, author, subject, keywords, creation and modification date). We assume that metadata μ\mu associated with α\alpha artefacts can be extracted automatically and produced by a research team member at a given time through processing operations λ\lambda applied to artefacts. For example, an AI algorithms or annotation actions produce metadata that must be associated with artefacts.

Narrative. A narrative η\eta of an artefact α\alpha produced by a researcher corresponds to textual content commenting and completing the content of an artefact, or a critical perspective of the content. The narrative is created be a member of the team at a given time τ\tau. Let η\eta (definition (2) below) be an annotation done by a researcher identified with an IdId at timestamp t.

η: content: Document,narrative: Textproducer: Researcher,timestamp: τ \noindent\eta:\begin{tabular}[]{l l}&$\langle$ content: Document,\\ &narrative: Text\\ &producer: Researcher,\\ &timestamp: $\tau$ $\rangle$\\ \end{tabular} (2)

Action. Let λ\lambda represents the meta-data of an operation applied on an artefact (see definition (3) below). It represents input parameters, chosen by a team member identified with an IdId, at time τ\tau, with an associated result and a set of assessment scores. Input parameters and results are modelled as artefacts.

λ: original: α, result: α,operation: Operation,timestamp: τ \lambda:\begin{tabular}[]{l l}&$\langle$ original: $\alpha$, result: $\alpha$,\\ &operation: Operation,\\ &timestamp: $\tau$ $\rangle$\\ \end{tabular} (3)
ID Operation Domain Description
Op1Op_{1} add (α\alpha) α\alpha,αα\sum_{\alpha}\rightarrow\sum_{\alpha} adds α\alpha artefact to the collection of artefacts α\sum_{\alpha}
Op2Op_{2} remove (α\alpha) α\alpha,αα\sum_{\alpha}\rightarrow\sum_{\alpha} removes α\alpha artefact from the collection of artefacts α\sum_{\alpha}
Op3Op_{3} contains(α\alpha) α\alpha,αboolean\sum_{\alpha}\rightarrow boolean returns true if the artefact α\alpha is in the collection of artefacts α\sum_{\alpha}
Op4Op_{4} addMetadata (μ\mu,α\alpha) μ\mu,α\alpha \rightarrow(μ,α)(\mu,\alpha) adds the metadata μ\mu to an existing artefact α\alpha
Op5Op_{5} updateMetadata (μ\mu,α\alpha) μ\mu,α\alpha (μ,α)\rightarrow(\mu,\alpha) updates the metadata μ\mu of an artefact α\alpha
Op6Op_{6} getMetadata (α\alpha) α\alpha \rightarrow(μ,α)(\mu,\alpha) returns the metadata μ\mu of the artefact α\alpha
Op7Op_{7} addRITL (η\eta,λ\lambda,α\alpha) η\eta, λ\lambda,α\alpha ((η,λ),α)\rightarrow((\eta,\lambda),\alpha) addRITL narrative η\eta for the artefact α\alpha once applied the algorithm λ\lambda
Op8Op_{8} getRITL (α\alpha) α\alpha ((η,λ),α)\rightarrow((\eta,\lambda),\alpha) returns the RITL narrative η\eta for the artefact α\alpha with the corresponding algorithm λ\lambda
Op9Op_{9} version (α\sum_{\alpha}) γ\gamma,αα\sum_{\alpha}\rightarrow\sum_{\alpha} commits γ\gamma message to a new version of the collection of artefacts α\sum_{\alpha} denoted as v\sum_{v}
Op10Op_{10} getCurated(urlurl) urlαurl\rightarrow\sum_{\alpha} given a urlurl returns a versioned curated collection of artefacts v\sum_{v}
Table 1: Annotation and versioning operations on artefacts.

Table 1 shows the operations that can be applied to an artefact α\alpha. An operation is executed at a given time τ\tau to curate quanti-qualitative content. In general, an operation produces a new version of an artefact timestamped with τ\tau and signed with the identity of the research team member who executed the operation. Operations can be classified into three groups: (i) manipulating artefacts adding and removing an artefact from the collection of a project; (ii) annotating artefacts by adding and updating metadata automatically or by a researcher; (iii) exploring artefacts versions.

Project. Π\Pi is defined as a process consisting of sequential stages of the curation workflow (problem statement, data acquisition, data management, analysis and report), a collection of artefacts α\sum\nolimits_{\alpha}, a research team RR and a set of released project versions υ\sum\nolimits_{\upsilon}.

Curation process. A version control system implements the artefacts versioning. It traces the life cycle of artefacts and manages artefact versions produced over time by research team members with descriptions of the consensus protocols performed to commit and validate operations. The algorithm 1 shows the curation steps that include creating α\alpha artefacts, annotating them with metadata μ\mu and performing consensus protocols for creating, validating, sharing versions and moving from one workflow step to another.

Input : : Research artefact α\alpha
Output : : Version control branch with data, metadata, and narrative in a given collection v\sum_{v} from the curated artefact collection α\sum_{\alpha}
Step 0: Initialize the curated collection α\sum_{\alpha}\leftarrow\emptyset
Step 1: Select the research artefact α\alpha to be curated and add it to the curated collectionα\sum_{\alpha}
if  α\alpha α\notin\sum_{\alpha} then
       ααα\sum_{\alpha}\leftarrow\alpha\cup\sum_{\alpha}
end if
      Step 2: Create metadata μ\mu for the artefact α\alpha˙α(α,μ)\sum_{\alpha}\leftarrow(\alpha,\mu) adds the corresponding metadata μ\mu to the artefact α\alpha and to the existing curated collection of artefacts α\sum_{\alpha}.
      Step 3: Given an operation λ\lambda, an existing research artefact α\alpha of the collection α\sum_{\alpha}, and the narrative η\eta produced by a researcher, associate both the algorithm λ\lambda and the artefact α\alpha,        α(α,(λ,η))\sum_{\alpha}\leftarrow(\alpha,(\lambda,\eta))
      Step 4: iterate Steps 1, 2 and 3 in a Version Control System.
      Step 5: Generate v\sum_{v}, a curated branch for the artefact collection α\sum_{\alpha}.
Algorithm 1 Curating Research Artefacts

The algorithm consists in initialising the containers where the curated artefacts’ versions produced during the workflow are gathered and versioned (Step 0). Then in Steps 1 to 4, artefacts’ versions are created (Step 1), and meta-data can be associated automatically or manually (Step 2). A subset of artefacts can be then analysed and processed (Step 3), leading to new curated artefacts containing meta-data about the operations (input, output, parameters, results, assessment scores and interpretation narratives). The curation process is performed at every phase of the curation workflow and is iterative until the workflow progresses to the next phase (Step 4). Finally, when a curation iteration converges, a new validated artefact set is liberated and accessible to the whole team (Step 5).

4 Consensus-based versioning control

A versioning pipeline and a consensus protocol guide the manipulation of artefacts within a project in our approach illustrated in Figure 1.

Refer to caption
Figure 1: Versioning and consensus-based artefacts manipulation.

The research team chooses the evolution of the status of every stage with its artefacts through a consensus protocol. Through this protocol, the research team members decide whether the status of a project stage will change depending on the validation/disapproval of the artefacts produced in that stage. The research team decides the evolution of the status of every stage of the project and the validation of artefacts through consensus protocols. Through these protocols, the research team members decide whether the status of a project stage will change depending on the validation/disapproval of the artefacts α\sum_{\alpha} produced in that stage. This version control system implements the artefacts versioning. It traces the life cycle of artefacts and manages artefact versions produced over time by research team members with descriptions of the consensus protocols performed to commit and validate operations.

4.1 Versioning curated content

We propose versioning operators to guide a project’s workflow for managing curated artefacts α\sum_{\alpha} representing content (documents) and metadata. Versioning operators represents decisions made at specific project stages to create, validate (globally or by specific team members) and share artefacts’ versions v\sum_{v}. The operators are executed within consensus-based protocols performed by the research team members. This execution model is an original aspect of the versioning control proposed for the quanti-qualitative content curation.

Versioning We have defined four groups of operators for dealing with the versioning of the stages of a research project. Versioning a stage means versioning the artefacts produced within its different versions as illustrated in Figure 1.

- Initialisation θinit,clone,drop\theta_{init,clone,drop}. The operator θinit\theta_{init} initialises a version workflow with an initial stage with a first branch (main) with a set of artefacts and a history with a first element representing an initial state ( θinit\theta_{init}, θclone\theta_{clone}). There are two options for initialising versioning locally θinit\theta_{init} or cloning θclone\theta_{clone} from a remote location.

The operator θdropS\theta_{dropS} removes all the tracking versioning of a workflow leading to the set of artefacts at the stage of the operation θdropS\theta_{dropS}. The operator θdropB\theta_{dropB} removes the branch from a stage with its artefacts.

- Adding/removing artefacts from a version pool:ϕadd,remove\phi_{add,remove}, ϕadd\phi_{add} Add and remove artefacts from the version history of a given project stage branch.

- Commit γ{\gamma}. The operator commits γ{\gamma} a stage adding an element to that stage’s versioning curated collection.

Tagging artefacts ζtag\zeta_{tag} ζtag\zeta_{tag} represents the operation of associating content with an artefact α\alpha by creating tags. A tag contains observations and interpretations (understanding) generated automatically and by a human. Tags are α\alpha artefacts containing text, interviews, photos, videos, and data in general.

- Branch β{\beta} Forks the main branch of a stage to create a new branch within the versioning workflow. The new branch is associated with a copy of all or part of the artefacts of the stage. It also adds an element to the versioning history of the stage from which the branch is created.

4.2 Consensus protocols

We proposed in our approach consist of functions gpref(G,i) and dis(G, i) that aggregate individual team members’ preferences to reflect the overall team’s decision-making (according to [1] definition on consensus functions for group recommendation)444Notice that several algorithms could be applied, for example Plurality Voting, Average, Least Misery, Quadratic Voting or Expert Judgement.. In our proposal, for each set of artefacts α\sum_{\alpha}, of a given research project Π\Pi, the agreement gpref(G,i) or disagreement function dis(G, i) of consensus must be applied for before versioning the curating state of artefacts.

There are two main aspects of a consensus function that refer to group preference and group disagreement:
(1) Group preference: The preference of a group for an item needs to reflect the degree to which all group members prefer the item. The more group members prefer an item, the higher its group preference. We adopt the average preference that computes the group preference for an item as the average of individual group members’ preferences for that item [1]. The preference of an item i by a group G, is given by:

gpref(G,i)=1|G|pref(u,i)\textsc{gpref}(G,i)=\frac{1}{|G|}\sum\textsc{pref}(u,i) (4)

(2) Group disagreement: The group preference must capture the level at which members disagree or agree. Pair-wise disagreement [1] computes the group preference for an item as the combination of its average and its pair-wise disagreement between individual group members’ preferences. The disagreement of a group GG over an item ii is given by:

dis(G,i)=2|G|(|G|1)|pref(u,i)pref(v,i)|\textsc{dis}(G,i)=\frac{2}{|G|(|G|-1)}\sum|\textsc{pref}(u,i)-\textsc{pref}(v,i)| (5)

where uv,u,vGu\neq v,u,v\in G, reflects the degree of consensus in the user-item preferences for ii among group members.

5 Experimental validation

The first version of the RITL curation approach and framework has been validated in the context of two use cases that apply quanti-qualitative research to perform political and ethnographic data-driven studies. 555The project is funded by **to be completed if accepted**.

  1. 1.

    UC1UC_{1} Political, artistic project for detecting political statements within grafitti in the downtown of a city in the global south.

  2. 2.

    UC2UC_{2} An anthropological study of stud reading practices.

To validate our approach, we first implemented a quanti-qualitative curation framework with tools used by pilot groups of social and human sciences research teams. This paper reports our return of experience on use case UC1UC_{1}.

5.1 Quanti-qualitative curation framework

We propose a RITL-driven curation system that allows researchers to curate the content produced during the phases of a project (see Figure 2). It provides tools for uploading digital content, creating associated meta-data (comments, feedback, tags), and interacting asynchronously with other researchers commenting and completing their meta-data. It also provides tools for maintaining the versioning of the content along the different phases of a project and its validation, and also the consensus process that makes the phases evolve from one to another (e.g., branching, validating, committing). The system allows the exploration and analysis of the projects’ curated content. Exploration and analysis of curated content will enable them to make informed decisions and have insight into their research practice.

Refer to caption
Figure 2: RITL driven curation framework for (quanti)-qualitative research methodologies.

5.2 Experimental results

The experimental validation of the framework that implements our content curation approach is intended to observe: (1) How versioning protocols based on consensus models are integrated into the quanti-qualitative workflows implemented in each use case by research teams. (2) How many cycles performed in each workflow phase took for the research team to converge? (3) How many artefacts were produced in each use case, and the average number of versions and branches created per phase and cycle of each use case workflow? According to our notation, Γ\Gamma represents the phases of the research project, proposed by a research team RR, with a set of curated artefacts α\sum\nolimits_{\alpha}, and a set of released versions υ\sum\nolimits_{\upsilon}. Let us define η\sum_{\eta} as the collection of narratives, regarding a collection of artefacts. Let us represent the research team RR implementing the research workflow consisting of team members of different types Ri,pR_{i,p} with different levels in a hierarchy pp. Let us also label the phases of the quanti-qualitative research workflow as the following: problem statement Γ1\Gamma_{1}, data acquisition Γ2\Gamma_{2}, data management Γ3\Gamma_{3}, data analysis Γ4\Gamma_{4} and report Γ5\Gamma_{5}.

UC1UC_{1} Political content in graffiti: The project was held by a research team RR with two members, a junior R0R_{0} and a senior scientist R1R_{1} working in a hierarchical organisation. The project adopted the spiral qualitative research phases workflow with the narrative and analytics phases Γ3,4\Gamma_{3,4} merged into one single phase.

The problem statement phase Γ1\Gamma_{1} had two cycles: four versions of an artefact α\alpha defining one research question and a narrative containing the definition of political graffiti, inclusion and exclusion rules to select graffiti were enumerated, and the theory framework supporting the pertinence of the research question. In this use case, the final research question emerged from sentence variations: Is it possible to trace and classify political messages disseminated in graffiti in a city? The first version of the artefact α1\alpha_{1} had fifteen comments (artefacts η115\eta_{1...15} proposed by the team members). No branches were created in this phase, and a consensus protocol was implemented until the second cycle using a pairwise disagreement protocol, given the hierarchical organisation and the pair number of members.

The data collection phase Γ2\Gamma_{2} had three cycles to define the area(s) in the city to collect the photographs of the graffiti by the junior team member. An artefact produced in the first cycle included a set of potential areas in the city to be explored for collecting graffiti samples to acquire evidence to select a data harvesting eventual area. The areas were old town, docks, commercial area, administrative area in downtown. The sample graffiti sets were artefacts’ first versions, with each graffiti tagged with its geographic location. In a second cycle, the team generated artefacts with comments about the elements of the samples. A new consensus pairwise consensus protocol was launched to validate a new artefact with the criteria to select the final collection area. As a result of the consensus, the collection areas were agreed upon in the second cycle. During the third cycle, the junior researcher collected and tagged 1050 graffiti photographs in different downtown areas of the old city. The team validated 546 photos defining a dataset artefact during a consensus pairwise protocol 666The selected final data set is published on Instagram (address to be specified if the paper is accepted).. An overview of the final dataset organised according to the downtown areas is shown in Figure 3. On the right, we can see the graffiti collected downtown, and on the left side, zoom in on those concerning the old city.

Refer to caption
Figure 3: Graffiti dataset organised according to their geographical area validated in phase Γ1\Gamma_{1} graffiti-areas.

The research team developed a phase integrating data management and analysis tasks Γ3,4\Gamma_{3,4}. The phase was performed in three cycles. In the first cycle, the team chose an analytics protocol consisting of a classification graffiti task performed in two branches, one with a manual task where an artefact containing classification categories specified how to classify the graffiti validated through a consensus protocol. The second branch consisting in a classification task applying machine learning classification models λ\lambda provided by Orange 777https://orangedatamining.com. Therefore the team agreed on k-means and hierarchical classification models (λ\lambda algorithms). Both branches produced several versions of artefacts classifying the graffiti photographs with associated metadata μ\mu, gathering information about decision-making and assessment scores used to converge into a final classification in each branch. Metadata μ\mu containing the comments on the partial results and consensus protocols within the branches were also associated artefacts. In the second cycle, both branches were merged, and the team discussed the classification results generating narratives to interpret them (see Figure 5). Four versions of the narratives were produced, including comments and modification requests (i.e., new artefacts) that led to these versions. The narrative with classification was validated through a consensus protocol. The final validated result of the phase Γ3,4\Gamma_{3,4} led to fifteen artefacts.

Refer to caption
Figure 4: Phases Γ1..5\Gamma_{1..5}.

Phases Γ1..5\Gamma_{1..5} evolves also in a versioning cycle begin with θinit\theta_{init} (initialization of the repository), addition and removing of the corresponding artifacts with ϕadd,remove\phi_{add,remove} operation, and corresponding versioning commit . Subsequently, once the artifacts were tag by ζtag\zeta_{tag}, the corresponding branches β{\beta} were unified to generate the stable version v\sum_{v}.

Finally, phase Γ5\Gamma_{5} consisted of two cycles during which the team produced two versions of the final report with analytics and conclusions about the study. In cycle two, the team validated within a consensus protocol three artefacts, the quantitative results from the classification algorithms, the qualitative results gathered in a report and the graffiti dataset published on Instagram.

Refer to caption
Figure 5: Analytics artefacts example of the branch 2 of the second cycle of the phase Γ3\Gamma_{3} applying λ\lambda algorithms provided by Data Mining software Orange.

5.3 Lessons learned

In quanti-qualitative data-driven projects, research teams design, application and processing of exploratory interviews for collecting data, define criteria to include and exclude content, annotate content, and recurrently discuss with the team members to converge on the progress of the project and document discussions and agreements.

The consensus within the research team allows rigorous analysis and content assessment throughout the process. Content is contained in artefacts produced as a result of these actions. Besides, these actions can be supported by digital tools such as algorithms, mathematical models, and text processing tools. The criteria used to choose the digital tools and the parameters to calibrate and validate results are also gathered as metadata.

Researchers’ intervention is essential throughout the research process and determines how the activities should be done. Establishing consensus mechanisms as the number of research teams increases is a challenge.

As the use cases were implemented in our framework, we observed that we must define tools for specifying research workflows according to the characteristics of the projects. The team must have ways to set inputs, outputs of tasks, restrictions, dependencies and other research objectives to guide the curation process. They should also be able to adapt consensus mechanisms for validating artefacts and assessment metrics.

6 Conclusion and Future Work

This paper proposes a content curation for quanti-qualitative research methods with a RITL approach. The versioning control guides the evolution of a quanti-qualitative research workflow that produces and consumes content versions through collaborative phases, validated through consensus protocols performed by the research team. The research curation system we propose allows researchers to work on collaborative data-driven projects by processing and managing digital artefacts, their changes, metadata, and narrative throughout the research process.

We argue that versioning research artefacts with a RITL approach can promote the creation, acquisition, and sharing. This approach also encourages reproducibility, traceability, and rationale of the research process. Providing a querying model to explore and exploit curated content and workflow phases to derive, reproduce, and trace workflows is part of our future work.

References

  • [1] Amer-Yahia, S., Elbassuoni, S., Omidvar-Tehrani, B., Borromeo, R.M., Farokhnejad, M.: Grouptravel: Customizing travel packages for groups. In: 22nd International Conference on Extending Database Technology (EDBT) (2019)
  • [2] ATLAS.ti: ATLAS.ti. https://atlasti.com (last accessed April 2023)
  • [3] Barcelos, P.P.F., Sales, T.P., Fumagalli, M., Fonseca, C.M., Sousa, I.V., Romanenko, E., Kritz, J., Guizzardi, G.: A fair model catalog for ontology-driven conceptual modeling research. Conceptual Modeling. ER 73(3) (2022)
  • [4] Bhangu, S., Provost, F., Caduff, C.: Introduction to qualitative research methods–part i. Perspectives in Clinical Research 14(1),  39 (2023)
  • [5] Chen, N., Drouhard, M., Kocielnik, R., Suh, J., Aragon, C.: Using machine learning to support qualitative coding in social science: Shifting the focus to ambiguity. ACM Transactions on Interactive Intelligent Systems 8(2), 1–20 (2018)
  • [6] Dedoose: Dedoose. https://www.dedoose.com/ (last accessed April 2023)
  • [7] Esteva, M., Xu, W., Simone, N., Nagpal, K., Gupta, A., Jah, M.: Synchronic curation for assessing reuse and integration fitness of multiple data collections (2022)
  • [8] Evers, J.C.: Current issues in qualitative data analysis software (qdas): A user and developer perspective. The Qualitative Report 23(13), 61–73 (2018)
  • [9] Garkov, D., Müller, C., Braun, M., Weiskopf, D., Schreiber, F.: ” research data curation in visualization: Position paper”(data) (2023)
  • [10] Karcher, S., Kirilova, D.D., Pagé, C., Weber, N.: How data curation enables epistemically responsible reuse of qualitative data. The Qualitative Report 26(6), 1996–2010 (2021)
  • [11] Lafia, S., Thomer, A., Bleckley, D., Akmon, D., Hemphill, L.: Leveraging machine learning to detect data curation activities. In: 2021 IEEE 17th International Conference on eScience (eScience). pp. 149–158. IEEE (2021)
  • [12] Leipzig, J., Nüst, D., Hoyt, C.T., Ram, K., Greenberg, J.: The role of metadata in reproducible computational research. Patterns 2(9), 100322 (2021)
  • [13] Mosqueira-Rey, E., Hernández-Pereira, E., Alonso-Ríos, D., Bobes-Bascarán, J., Fernández-Leal, Á.: Human-in-the-loop machine learning: A state of the art. Artificial Intelligence Review 56(4), 3005–3054 (2023)
  • [14] NVivo: Nvivo. https://www.qsrinternational.com/ (last accessed April 2023)
  • [15] QDAMiner: Qdaminer. https://provalisresearch.com/products/qualitative-data-analysis-software/ (last accessed April 2023)
  • [16] Quirkos: Quirkos. https://www.quirkos.com (last accessed April 2023)
  • [17] Rahwan, I.: Society-in-the-loop: programming the algorithmic social contract. Ethics and information technology 20(1), 5–14 (2018)
  • [18] Sardana, N., Shekoohi, S., Cornett, E.M., Kaye, A.D.: Qualitative and quantitative research methods. In: Substance Use and Addiction Research, pp. 65–69. Elsevier (2023)
  • [19] Van de Schoot, R., de Bruin, J.: Researcher-in-the-loop for systematic reviewing of text databases. Zenodo: SciNLP: Natural Language Processing and Data Mining for Scientific Text (2020)
  • [20] Software, V.: Maxqda. http://maxqda.com (last accessed April 2023)
  • [21] Transana: Transana. https://www.transana.com (last accessed April 2023)
  • [22] Vuorre, M., Curley, J.P.: Curating research assets: A tutorial on the git version control system. Advances in Methods and Practices in Psychological Science 1(2), 219–236 (2018)
  • [23] Wilkinson, M.D., Dumontier, M., Aalbersberg, I.J., Appleton, G., Axton, M., Baak, A., Blomberg, N., Boiten, J.W., da Silva Santos, L.B., Bourne, P.E., et al.: The fair guiding principles for scientific data management and stewardship. Scientific data 3(1),  1–9 (2016)
  • [24] Woods, M., Macklin, R., Lewis, G.K.: Researcher reflexivity: exploring the impacts of caqdas use. International Journal of Social Research Methodology 19(4), 385–403 (2016)
  • [25] Zuiderwijk, A., Shinde, R., Jeng, W.: What drives and inhibits researchers to share and use open research data? A systematic literature review to analyze factors influencing open research data adoption. PloS One 15(9) (2020)