This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Can We Enhance the Quality of Mobile Crowdsensing Data Without Ground Truth?

Jiajie Li, Bo Gu,   Shimin Gong,   Zhou Su,  
and Mohsen Guizani
This work was supported in part by the National Science Foundation of China (NSFC) under Grant U20A20175. (Corresponding author: Bo Gu.) Jiajie Li, Bo Gu, and Shimin Gong are with the School of Intelligent Systems Engineering, Sun Yat-sen University, Shenzhen 518107, China, and with the Guangdong Provincial Key Laboratory of Fire Science and Intelligent Emergency Technology, Guangzhou 510006, China. E-mail: [email protected], [email protected], and [email protected]. Zhou Su is with the School of Cyber Science and Engineering, Xi’an Jiaotong University, Xi’an 710049, China. E-mail: [email protected]. Mohsen Guizani is with the Computer Science and Engineering Department, Qatar University, Doha 2713, Qatar. E-mail: [email protected].
Abstract

Mobile crowdsensing (MCS) has emerged as a prominent trend across various domains. However, ensuring the quality of the sensing data submitted by mobile users (MUs) remains a complex and challenging problem. To address this challenge, an advanced method is needed to detect low-quality sensing data and identify malicious MUs that may disrupt the normal operations of an MCS system. Therefore, this article proposes a prediction- and reputation-based truth discovery (PRBTD) framework, which can separate low-quality data from high-quality data in sensing tasks. First, we apply a correlation-focused spatio-temporal Transformer network that learns from the historical sensing data and predicts the ground truth of the data submitted by MUs. However, due to the noise in historical data for training and the bursty values within sensing data, the prediction results can be inaccurate. To address this issue, we use the implications among the sensing data, which are learned from the prediction results but are stable and less affected by inaccurate predictions, to evaluate the quality of the data. Finally, we design a reputation-based truth discovery (TD) module for identifying low-quality data with their implications. Given the sensing data submitted by MUs, PRBTD can eliminate the data with heavy noise and identify malicious MUs with high accuracy. Extensive experimental results demonstrate that the PRBTD method outperforms existing methods in terms of identification accuracy and data quality enhancement.

Index Terms:
Data quality, mobile crowdsensing, reputation, data prediction, truth discovery.

1 Introduction

Mobile crowdsensing (MCS) has drawn considerable attention in recent years. As a novel data acquisition paradigm, MCS leverages embedded sensors in mobile devices (e.g., smartphones and cameras) to gather large amounts of data for various applications [1]. Typically, an MCS system comprises three key types of components: a sensing platform, task initiators (TIs), and mobile users (MUs). Specifically, the TIs delegate sensing tasks to MUs through the sensing platform. The sensing platform gathers and sends the data submitted by the MUs to the TIs, subsequently assigning payments to the MUs. Compared with traditional sensing methods, MCS has advantages such as deployment flexibility, cost-effectiveness, extensive coverage, and scalability. Consequently, numerous applications based on MCS (e.g., ParkNet[2], Nextdoor[3], and Google Crowdsource[4]) have emerged.

However, the sensing data submitted by MUs may be unreliable because of sensor damage and the presence of malicious MUs [9, 5, 6, 7]. These low-quality sensing data (e.g., erroneous or outdated data) can mislead and negatively impact applications and systems based on sensing data, leading to reduced efficiency and accuracy. In contrast, applications and systems can improve the accuracy of their decisions and actions based on high-quality sensing data, helping them perform more efficiently and effectively. The impact of low-quality data highlights the importance of evaluating and enhancing the quality of the sensing data. Since the sensing data are measured and submitted by MUs, the sensing platform lacks the ground truth of these data, making it difficult to directly evaluate their quality. Consequently, considerable effort has been devoted to estimating the ground truth of sensing data, leading to the application of three types of methods in MCS systems: data aggregation methods [8, 9, 12, 11, 10], prediction-based methods [15, 13, 16, 17, 14, 18, 19, 20, 21], and truth discovery (TD) methods [22, 23, 24, 25, 26, 27, 29, 30, 31, 28]. Data aggregation methods estimate the ground truth of sensing data through clustering algorithms. However, their estimates may be inaccurate when insufficient sensing data are available [11, 10]. To address this problem, several schemes have designed MCS frameworks with pre-exploration stages to obtain sufficient baseline sensing data [12, 11]. Unfortunately, these methods increase the overhead associated with MCS. As many sensing tasks in MCS applications are repeated periodically or continue for a very long period, historical data can be obtained in many cases and are useful for ground truth estimation. Prediction-based methods initially use these historical data to learn and predict the ground truth of the sensing data submitted by MUs, subsequently calculating the errors between these data and the corresponding predicted ground truth to evaluate the data quality. On this basis, Xie et al. [13] and Zhang et al. [14] attempted to capture the spatial and temporal correlations among data and used them to predict the ground truth. However, these methods may not perform well in realistic MCS systems with sensing data, including bursty values.

Numerous schemes have applied TD in quality-based MCS systems to evaluate the quality of conflicting data [22, 23, 24, 25]. TD methods initially aggregate the sensing data submitted by MUs based on their weights (i.e., trustworthiness levels) to estimate the ground truth (e.g., calculate the weighted mean of all submitted data as the ground truth). The weights of the MUs are subsequently updated based on the errors between their submitted data and the estimated ground truth. Specifically, the weights of MUs are increased if they provide sensing data that are close to the estimated ground truth, resulting in greater impacts on the estimation process. The ground-truth estimates subsequently need to be calculated again as the weights of the MUs are adjusted. Therefore, the weights of the MUs and the ground-truth estimates are alternately updated until convergence is reached. To further improve the framework of TD methods, Cheng et al. [26] adopted a reputation mechanism [32, 33, 34] for TD in an MCS system to replace the weights of MUs with reputations. This approach effectively reduces the misclassification of MUs. Moreover, Yin et al. [27] utilized the implications among conflicting data and designed a framework based on TD to identify incorrect and fake data. However, the effectiveness of TD methods is limited by the quantity of the given sensing data. Like the data aggregation methods, TD methods estimate ground truth with low accuracy in scenarios where the sensing data are insufficient and sparse.

Note that in MCS systems, the sensing data submitted by MUs exhibit spatial and temporal correlations [5, 24, 35, 14] in most scenarios (e.g., the traffic flows of a city and the temperature and humidity of different regions), which can be utilized to improve the performance of the data quality evaluation. On this basis, we propose a prediction- and reputation-based TD (PRBTD) framework to evaluate the quality of sensing data with spatio-temporal correlations.

The proposed PRBTD method leverages the advantages of prediction-based methods, TD methods, and reputation mechanisms. First, we adopt a correlation-focused spatio-temporal Transformer network (CFSTTN), which is an improved version of the global-local spatio-temporal Transformer network (GLSTTN) [17], to learn from historical sensing data and predict the ground truth of the submitted sensing data. The CFSTTN incorporates a long-range residual connection to enhance its ability to extract the spatio-temporal correlations among data. However, in practice, the prediction results may be inaccurate because of the noise in historical sensing data and the bursty data values. Therefore, instead of directly using the prediction results for data quality evaluation, we extract the implications among data at different locations and in different periods based on the prediction results, which can remain stable and thus reduce the impact of inaccurate predictions. Finally, our PRBTD framework identifies malicious MUs and low-quality data through a TD strategy based on the reputations of the MUs and the implications among the data. Thus, our proposed method enhances the quality of the sensing data by filtering out incorrect and fake data. The main contributions of this work can be summarized as follows.

  • The proposed PRBTD framework consists of a CFSTTN prediction module, a data feature and implication calculation module, and a reputation-based TD module. This approach combines prediction, TD, and reputation mechanisms for the first time to evaluate the quality of the sensing data and is applicable in most scenarios, including cases with bursty data values and data sparsity.

  • The data feature and implication calculation module in PRBTD uses the prediction results to analyze the implications among the data acquired from various locations and in different periods. Since these implications remain stable, PRBTD reduces the impact of prediction inaccuracies by focusing on implications instead of directly using the prediction results. Therefore, PRBTD overcomes the shortcomings of prediction-based methods, especially in scenarios with bursty sensing data values.

  • The reputation-based TD module in PRBTD leverages the implications among the data and the reputations of MUs to achieve accurate evaluation of data quality. By utilizing these implications, the reputation-based TD module indirectly increases the amount of data available for evaluating the quality of the submitted data. Consequently, PRBTD can more accurately evaluate data quality in scenarios with sparse sensing data.

  • We conduct extensive simulation experiments on real-world datasets. The results indicate that our PRBTD framework is applicable to complex scenarios in MCS systems and outperforms existing methods.

2 Related Work

In this section, we briefly review the related work on quality-based MCS systems and methods for evaluating data quality.

2.1 Quality-Based MCS Systems

Numerous studies have been dedicated to the establishment of data quality-based MCS systems [37, 39, 38, 40, 36]. An effective approach is to design incentive and task assignment mechanisms that induce MUs to submit high-quality data. For example, Jin et al. [37] designed an incentive mechanism based on reverse combinatorial auctions, aiming to encourage the participation of MUs and maximize social welfare. Xiao et al. [38] used a multi-armed bandit algorithm to select MUs with good performance to ensure data quality. Gu et al. [39] adopted a Stackelberg leader-follower game framework in MCS systems to design a task assignment mechanism that is able to choose MUs with high potential through multiagent reinforcement learning. However, these methods assume prior data quality knowledge, which does not always hold in practice. Therefore, some methods have added pre-exploration stages to MCS systems to estimate the capabilities of MUs [40]. For example, Zhang et al. [11] and Gao et al. [12] used the data sensed in the pre-exploration stages as baselines for evaluating the sensing errors of MUs, which could represent their abilities. Although these methods can more accurately evaluate the quality of the sensing data, the overhead of the resulting MCS platform is greatly increased.

2.2 Data Quality Evaluation

Many studies have focused on evaluating the quality of the data collected within MCS systems [41, 35]. Zhang et al. [11] and Jin et al. [37] compared sensing data with the corresponding ground truth to calculate the induced sensing errors and to evaluate the quality of the data. On the basis of these results, Gao et al. [12] decomposed the sensing errors into deviations and variances to represent the sensing abilities and stabilities of MUs. However, these methods assume that the ground truth is known to the sensing platform, which is impractical in MCS systems. Therefore, many studies have been conducted on ground truth estimation, resulting in three types of methods: data aggregation methods [8, 9, 10, 12, 11], prediction-based methods [15, 13, 16, 17, 14, 18, 19, 20, 21], and TD methods [22, 23, 24, 25, 26, 27, 29, 30, 31].

Data aggregation methods. Zhang et al. [11] applied a weighted averaging method to estimate the ground truth of the sensing data. However, this method requires sufficient sensing data to obtain accurate estimates, which is difficult to achieve in scenarios with sparse data. To address data sparsity, Meng et al. [8] introduced a method for estimating the quality of the sensing data via a sparse data matrix. Specifically, they constructed a data matrix from submitted data and then filled in the missing data via Bayesian probabilistic matrix factorization. Building on this contribution, Yang et al. [9] deployed a small number of unmanned aerial vehicles to collect trustworthy data for constructing the sparse baseline data matrix. However, these methods focus only on the spatial correlations among data and overlook the temporal aspect, which limits their effectiveness.

Prediction-based methods. Min et al. [15] utilized the spatio-temporal correlations among traffic data and achieved improved prediction performance with their proposed dynamic space-time autoregressive integrated moving average (STARIMA) model. Furthermore, Yao et al. [16] applied a Transformer framework for traffic prediction and proposed a novel multi-view spatio-temporal graph network. This model combines attention and convolution mechanisms in its traffic pattern analysis strategy to comprehensively determine the spatio-temporal characteristics. On the basis of this approach, Gu et al. [17] proposed a Transformer model with global and local spatio-temporal modules to analyze the strong dynamics and spatio-temporal dependencies that are inherent in cellular traffic data. However, the predictions yielded by these methods may be inaccurate when bursty values are present in the sensing data, which limits the performance of prediction-based methods in MCS systems.

TD methods. TD methods are considered efficient for estimating the ground truth from unreliable data [22, 23, 24, 25] and are thus suitable for MCS systems. For example, Cheng et al. [26] presented a framework that combines TD with zero-knowledge proof in an MCS system to improve the accuracy of the ground truth produced for sensing data while preserving user privacy. Yin et al. [27] leveraged TD to evaluate the trustworthiness of web-based information sources and identify incorrect and fake data. However, these traditional TD methods typically produce unsatisfactory results in practical applications because of the possibility of sparse sensing data being present in MCS systems.

In summary, most existing methods can handle only specific scenarios. Inspired by these works, we have found that prediction-based methods and TD methods can effectively complement each other. Therefore, we propose leveraging the principles of prediction-based methods and reputation-based TD methods while fully utilizing the spatio-temporal correlations among data to implement our PRBTD method, which is able to enhance the effectiveness of data quality evaluation across various complex scenarios.

3 MCS System Model

Refer to caption
Figure 1: Architecture of an MCS system.

The architecture of an MCS system is illustrated in Fig. 1. In MCS, a sensing task typically spans a duration of TT time slots and covers an area with NN regions [35, 30]. In this work, we study scenarios with historical sensing data, and these data collected in a task exhibit temporal and spatial correlations (e.g., collecting traffic flow data in a city for several days). The time slot of the task is denoted by t{1,2,,T}t\in\left\{1,2,...,T\right\}. The N(=H×W)N(=H\times W) sensing regions are identified based on their latitudes and longitudes. To simplify the expression, we use n{1,2,,N}n\in\left\{1,2,...,N\right\} to identify a region (h,wh,w), where n=(h1)×W+wn=(h-1)\times W+w. An MU in the MCS system is identified as i{1,2,,I}i\in\left\{1,2,...,I\right\}, where II is the total number of MUs. In particular, we assume that each MU submits sensing data at most once per time slot. The data submitted to the platform by an MU should include the ID ii of the MU, the sensing time slot tt, the sensing location nn, and the sensing value vi,ntv_{i,n}^{t}. In our systems, we assume that the sensing data have continuous data values. However, our proposed PRBTD method can be generalized to other data types (e.g., categorical sensing data). Thus, we represent the sensing data point submitted by MU ii in region nn at time slot tt as di,nt=(i,t,n,vi,nt)d_{i,n}^{t}=\left(i,t,n,v_{i,n}^{t}\right), and use 𝑫nt\bm{D}_{n}^{t} to denote the set of all the data sensed in region nn at time slot tt. Moreover, there are malicious MUs in our system. Malicious MUs are users who continuously submit low-quality data that are harmful to the system. Similar to [27, 21], we assume that the proportion of malicious MUs in our system should not exceed 50%. This assumption is based on the expectation that the majority of MUs are honest [27]. If the proportion of malicious MUs exceeds 50%, the system can be considered under attack, making the system untrustworthy [33]. In such a case, the majority of the sensing data may be malicious, rendering most existing methods for data quality evaluation ineffective. The interactions among the TI, MUs, and sensing platform in the MCS system are as follows [30, 39].

  1. 1.

    At the beginning of a sensing task, the TI broadcasts the task requirements to the available MUs through the sensing platform.

  2. 2.

    During each time slot of the task, the MUs may choose to sense and submit their data to the sensing platform. At the end of each time slot, the sensing platform gathers and identifies the incorrect or fake data among all submitted data. These low-quality data are then eliminated by the platform to enhance the quality of the remaining data.

  3. 3.

    At the end of the task, payments are assigned to the MUs based on the quality of their sensing data, and high-quality data are delivered to the TI.

4 Our Method

In this section, we introduce the proposed PRBTD framework. As shown in Fig. 2, our PRBTD framework consists of three modules: a CFSTTN prediction module, a data feature and implication calculation module, and a reputation-based TD module. Table I lists the notations used in the PRBTD framework.

Refer to caption
Figure 2: Architecture of the PRBTD framework.
TABLE I: Notations Used in the PRBTD Framework
Notation Description
I,iI,i The total number of MUs, the ID of an MU
T,tT,t The total number of time slots for a sensing task, a time slot in the task duration
N,nN,n The total number of regions in a sensing task, a region in the task area
di,ntd_{i,n}^{t} The sensing data submitted by MU ii for region nn at time slot tt
vi,ntv_{i,n}^{t} The value of sensing data point di,ntd_{i,n}^{t}
𝑰(vi,nt)\bm{I}(v_{i,n}^{t}) The set of MUs who sense the same value vi,ntv_{i,n}^{t} in region nn at time slot tt
𝑫nt\bm{D}_{n}^{t} The set of data sensed in region nn at time slot tt
𝑫,𝑫^t\bm{D}^{\prime},\widehat{\bm{D}}^{t} The data buffer, the set of high-quality data at time slot tt
Mnt,𝑴nM_{n}^{t},\bm{M}_{n} The mean of the sensing data values in region nn at time slot tt, the set of mean data values at time slot tt in all regions
G^nt,𝑮^t\widehat{G}_{n}^{t},\widehat{\bm{G}}^{t} The ground-truth predictions of data in region nn at time slot tt, the set of ground-truth predictions at time slot tt
δi,nt\delta_{i,n}^{t} The sensing error of data point di,ntd_{i,n}^{t}
δ˙nt,δ¨it\dot{\delta}_{n}^{t},\ddot{\delta}_{i}^{t} The circumstance-related error of region nn at time slot tt, the user-related error of MU ii at time slot tt
rit,Ritr_{i}^{t},R_{i}^{t} The reputation of MU ii at time slot tt, the corresponding reputation score
q(di,nt),Q(di,nt)q(d_{i,n}^{t}),Q(d_{i,n}^{t}) The data quality of sensing data point di,ntd_{i,n}^{t}, the corresponding quality score
imp(di,nt,dj,nt)imp\!\left(d_{i,n}^{t},d_{j,n^{\prime}}^{t^{\prime}}\right) The implication between sensing data points di,ntd_{i,n}^{t} and dj,ntd_{j,n^{\prime}}^{t^{\prime}}

4.1 CFSTTN Prediction Module

In an MCS system, the sensing platform cannot directly evaluate the quality of the submitted data due to the absence of the ground truth. However, in scenarios with historical sensing data, prediction methods can be employed to estimate the ground truth. Thus, we propose a CFSTTN to learn from historical data and predict the ground truth of the submitted sensing data.

Refer to caption
Figure 3: Structure of the CFSTTN.

As shown in Fig. 3, the CFSTTN comprises three components: an embedding module, a global spatio-temporal module, and a local spatio-temporal module. First, the embedding module adjusts the dimensions of the input data. Then, the global spatio-temporal module deploys a spatio-temporal Transformer framework based on a multi-head attention mechanism to extract multidimensional global spatial and temporal characteristics from the input data. Finally, the local spatio-temporal module uses densely connected convolutional neural networks to further extract local spatio-temporal information based on the global characteristics of the input data.

Our proposed CFSTTN improves the original model by integrating a correlation-focused part with a long-range residual connection with the input data. As presented in Fig. 3, our CFSTTN predicts a spatio-temporal relationship weight matrix, which is subsequently multiplied with the input data to derive the final output. The integration of the long-range residual connection enhances the spatio-temporal correlations between the output and input data, thus improving the prediction accuracy.

The CFSTTN uses historical data in the sensing task for prediction purposes. We represent the mean values of the data at time slot tt as a column vector 𝑴t=[M1t,M2t,,MNt]𝖳\bm{M}^{t}=\left[M_{1}^{t},M_{2}^{t},...,M_{N}^{t}\right]^{\mathsf{T}} where

Mnt=di,nt𝑫ntvi,nt|𝑫nt|M_{n}^{t}=\frac{\sum_{d_{i,n}^{t}\in\bm{D}_{n}^{t}}v_{i,n}^{t}}{|\bm{D}_{n}^{t}|} (1)

is the mean value of the data sensed at time slot tt in region nn. We construct a matrix [𝑴tp+1,,𝑴t]N×p\left[\bm{M}^{t-p+1},...,\bm{M}^{t}\right]\in\mathbb{R}^{N\times p} as the input of the CFSTTN, which consists of the mean values of data derived from the preceding pp time slots up to tt, where 0<p<T0<p<T is a hyperparameter. The predicted ground truth, which includes the mean values of the data at time slot t+1t+1 from every region, is denoted as 𝑮^t+1=[G^1t+1,G^2t+1,,G^Nt+1]𝖳\widehat{\bm{G}}^{t+1}=\left[\widehat{G}_{1}^{t+1},\widehat{G}_{2}^{t+1},...,\widehat{G}_{N}^{t+1}\right]^{\mathsf{T}}. Additionally, we use f()f(\cdot) to represent a series of operations conducted within the three components of the CFSTTN. The spatio-temporal correlation weight matrix can be represented as f(𝑴tp+1,,𝑴t)f(\bm{M}^{t-p+1},...,\bm{M}^{t}). Finally, we obtain the prediction of the ground truth at time slot t+1t+1 by performing matrix multiplication between the spatio-temporal correlation weight matrix and the input data:

𝑮^t+1=[𝑴tp+1,,𝑴t]f(𝑴tp+1,,𝑴t).\widehat{\bm{G}}^{t+1}=\left[\bm{M}^{t-p+1},...,\bm{M}^{t}\right]f(\bm{M}^{t-p+1},...,\bm{M}^{t}). (2)

In this paper, the mean absolute error (MAE) is utilized as the training metric. Specifically, we use the backpropagation method to minimize the MAE between the predicted ground truth 𝑮^t+1\widehat{\bm{G}}^{t+1} and the mean values of the actual submitted data 𝑴t+1\bm{M}^{t+1}, which can be expressed as follows:

(θ)=1Nn=1N|G^nt+1Mnt+1|\mathcal{L}\left(\theta\right)=\frac{1}{N}\sum_{n=1}^{N}\left|\widehat{G}_{n}^{t+1}-M_{n}^{t+1}\right| (3)

where θ\theta represents the parameters in the CFSTTN, which are learned and updated during the training process. However, the prediction results in an MCS task may be inaccurate because of the noise in the training data and the bursty values in the sensing data. Therefore, the output of the CFSTTN prediction module cannot be directly used to evaluate the quality of the data.

4.2 Data Feature and Implication Calculation Module

In the MCS system, implicit correlations are present among the sensing data with spatial and temporal correlations. We denote the implicit correlation among the sensing data points as implicationimplication, which reflects the extent to which one data point can predict or support another. For example, consider the scenario shown in Fig. 4 where MUs collect traffic flow data on a one-way road. In this case, the traffic flow on segment C is the sum of the traffic flows on segment A and segment B at any period tt. Therefore, when the traffic data points submitted by MUs satisfy v1,At+v2,Bt=v3,Ctv_{1,A}^{t}+v_{2,B}^{t}=v_{3,C}^{t}, the implications among these data points are strong, as they mutually support each other’s accuracy. In contrast, the implications are weak when the data points follow v1,At+v2,Btv3,Ctv_{1,A}^{t}+v_{2,B}^{t}\neq v_{3,C}^{t}, which means that some of these MUs have submitted incorrect data that thus contradict the other data. Similarly, in most MCS scenarios, we can infer the existence of implications among the sensing data. However, owing to the complex spatial and temporal correlations among the data, quantitatively analyzing these implications is challenging in most cases. Therefore, we attempt to understand these implications from historical data. Specifically, we perform a series of computations in this module to quantify these implications from the output of the CFSTTN prediction module. Before providing the detailed calculations, we define the degree of implication as follows:

Definition 1.

(Degree of implication) The degree of implication between data points di,ntd_{i,n}^{t} and dj,ntd_{j,n^{\prime}}^{t^{\prime}} is represented as imp(di,nt,dj,nt)imp(d_{i,n}^{t},d_{j,n^{\prime}}^{t^{\prime}}), where n{1,2,,N}n^{\prime}\in\left\{1,2,...,N\right\} and 0<ttT0<t^{\prime}\leq t\leq T. Moreover, the degree satisfies 1imp(di,nt,dj,nt)1-1\leq imp(d_{i,n}^{t},d_{j,n^{\prime}}^{t^{\prime}})\leq 1.

Refer to caption
Figure 4: Simple example of MUs sensing traffic flow data.

Specifically, imp(di,nt,dj,nt)=0imp(d_{i,n}^{t},d_{j,n^{\prime}}^{t^{\prime}})\!=\!0 indicates a lack of implication between data points di,ntd_{i,n}^{t} and dj,ntd_{j,n^{\prime}}^{t^{\prime}}, whereas imp(di,nt,dj,nt)>0imp(d_{i,n}^{t},d_{j,n^{\prime}}^{t^{\prime}})>0 indicates mutual support between the data, and imp(di,nt,dj,nt)<0imp(d_{i,n}^{t},d_{j,n^{\prime}}^{t^{\prime}})<0 means that the data contradict each other. Implications close to 1 indicate that the data points strongly support each other, whereas implications close to -1 suggest that the data points are likely to be entirely contradictory.

Refer to caption

(a) MUs sense data in the same region at the same time slot.

Refer to caption

(b) MUs sense data in different regions at different time slots.

Figure 5: MCS scenarios with multiple MUs.

To calculate the degree of implication between sensing data points, we first reduce the dimensionality [42] of the sensing data di,nt=(i,t,n,vi,nt)d_{i,n}^{t}=\left(i,t,n,v_{i,n}^{t}\right) to extract the features. Specifically, since sensing errors can reflect the characteristics of data [11, 12], we calculate the sensing errors as data features. Since the actual ground truth of the sensing data are unknown to us during a sensing task, we calculate the sensing errors of the data points based on the ground truth prediction results (i.e., the output of the CFSTTN). When the corresponding predicted ground truth G^nt0\widehat{G}_{n}^{t}\neq 0, the error of data point di,ntd_{i,n}^{t} is calculated as

δi,nt=vi,ntG^ntG^nt.\delta_{i,n}^{t}=\frac{v_{i,n}^{t}-\widehat{G}_{n}^{t}}{\widehat{G}_{n}^{t}}. (4)

In most MCS scenarios, the sensing errors in the data are caused primarily by two factors [12]. The first factor is related to external sensing circumstances, such as weather conditions, which are influenced by the time and location of the data collection. The second factor is related to the MUs, such as issues with the devices or improper operations when sensing data. Therefore, we decompose the sensing errors in the data into two components, which are defined below.

Definition 2.

(Circumstance-related error) The error caused by external factors (e.g., the illumination intensity or weather conditions) during data collection is denoted as δ˙nt\dot{\delta}_{n}^{t}. This error is influenced by the sensing region and period.

Definition 3.

(User-related error) The error caused by internal factors (e.g., devices or operations) of MU ii during data collection is denoted as δ¨it\ddot{\delta}_{i}^{t}. Note that δ¨it\ddot{\delta}_{i}^{t} is also related to the sensing period tt, as the internal factors of an MU may vary in different periods.

Hence, the total sensing error of a data point di,ntd_{i,n}^{t} can be divided into the circumstance-related error and the user-related error as follows:

δi,nt=δ˙nt+δ¨it\delta_{i,n}^{t}=\dot{\delta}_{n}^{t}+\ddot{\delta}_{i}^{t} (5)

On this basis, we learn that the error vector [δ˙nt,δ¨it][\dot{\delta}_{n}^{t},\ddot{\delta}_{i}^{t}] corresponds to, and is uniquely related to, the data point di,ntd_{i,n}^{t}. Therefore, we use the error vector of a data point as its data feature, which can reflect most of the characteristics of that data point. Specifically, we can analyze the relationship between the features of the data points and the implications among them through the following conjecture, which allows us to infer the implications among data points based on their features.

Conjecture 1.

The degree of implication between two data points is positively correlated with the similarity between the corresponding data features [27, 28].

This conjecture is applicable in most MCS scenarios. The reason is that data features reflect their patterns; thus, data with similar features usually follow a similar pattern and can infer each other. In our work, the features we extract from the sensing data include both circumstance-related errors and user-related errors. Therefore, if two data points have similar features, their circumstance-related errors and user-related errors are similar. According to Definitions 2 and 3, these two data points may be collected by users with similar characteristics (e.g., using similar devices or similar operations) in similar circumstances (e.g., under similar weather conditions). In such cases, the two data points can strongly support each other’s authenticity [31]. For example, when traffic flow data are sensed on different road segments with similar traffic characteristics (e.g., parallel roads with comparable traffic flows), if the data from two segments are very close, they can mutually support each other. Hence, the implication between two data points with similar features is strong. In contrast, data points with dissimilar features may contradict each other; therefore, the implications among these data are weak. Consequently, we can infer that in most MCS scenarios, Conjecture 1 is reasonable.

Based on Conjecture 1, we calculate the cosine similarity of the data features to determine the degree of implication between the data points as shown in Fig. 5(b). Even though the cosine similarity focuses only on direction, it is an appropriate choice for calculating the similarity of data features because our data features already contain magnitude information. The detailed process of calculating the data features and degrees of implication is as follows.

Calculate the circumstance-related error δ˙nt\bm{\dot{\delta}_{n}^{t}}. In our MCS system, the proportion of malicious MUs is low [27, 33, 21], and the MCS platform can identify and exclude these malicious MUs during sensing tasks. Therefore, we assume that the impact of malicious MUs on the calculations is limited.

According to (5), we calculate the average sensing error of all the data collected in region nn as follows:

τ=1di,nτ𝑫nτδi,nττ=1|𝑫nτ|=τ=1di,nτ𝑫nτ(δ˙nτ+δ¨iτ)τ=1|𝑫nτ|\frac{\sum_{\tau=1}^{\infty}\sum_{d_{i,n}^{\tau}\in\bm{D}_{n}^{\tau}}\delta_{i,n}^{\tau}}{\sum_{\tau=1}^{\infty}|\bm{D}_{n}^{\tau}|}=\frac{\sum_{\tau=1}^{\infty}\sum_{d_{i,n}^{\tau}\in\bm{D}_{n}^{\tau}}\left(\dot{\delta}_{n}^{\tau}+\ddot{\delta}_{i}^{\tau}\right)}{\sum_{\tau=1}^{\infty}|\bm{D}_{n}^{\tau}|} (6)

where 𝑫nτ\bm{D}_{n}^{\tau} represents the set of data sensed in region nn at time slot τ\tau. We use cntc_{n}^{t} to represent the amount of sensing data collected in region nn up until time slot tt. Specifically, cntc_{n}^{t} is calculated as follows:

cnt={0,t=0cnt1+|𝑫nt|,0<tT.c_{n}^{t}=\begin{cases}0,&t=0\\ c_{n}^{t-1}+|\bm{D}_{n}^{t}|,&0<t\leq T.\end{cases} (7)

Based on (7), we can rewrite (6) as follows:

τ=1(cnτcnτ1)δ˙nτ=τ=1di,nτ𝑫nτδi,nτ.\sum_{\tau=1}^{\infty}\left(c_{n}^{\tau}-c_{n}^{\tau-1}\right)\dot{\delta}_{n}^{\tau}=\sum_{\tau=1}^{\infty}\sum_{d_{i,n}^{\tau}\in\bm{D}_{n}^{\tau}}\delta_{i,n}^{\tau}. (8)

Note that when simplifying (6) to (8), we utilize the property that the user-related error of each normal MU ii satisfies

τ=1δ¨iτ=0.\sum_{\tau=1}^{\infty}\ddot{\delta}_{i}^{\tau}=0. (9)

The reason is that the long-term user-related error expectation for each normal MU in an MCS system should be zero.

However, the relationship between δ˙nt\dot{\delta}_{n}^{t} and δi,nt\delta_{i,n}^{t} shown in (8) holds only as tt\to\infty, which is not feasible in practice because the duration of a sensing task is limited. Therefore, we rewrite (8) in the following slot-based update form to calculate the estimated circumstance-related error δ˙n(t)\dot{\delta}_{n}^{(t)} in region nn at time slot tt:

δ˙n(t)={0,cnt=0δ˙n(t1)cnt1+di,nt𝑫ntδi,ntcnt,cnt0\dot{\delta}_{n}^{(t)}=\begin{cases}0,&c_{n}^{t}=0\\ \frac{\dot{\delta}_{n}^{(t-1)}c_{n}^{t-1}+\sum_{d_{i,n}^{t}\in\bm{D}_{n}^{t}}\delta_{i,n}^{t}}{c_{n}^{t}},&c_{n}^{t}\neq 0\end{cases} (10)

Note that limtδ˙n(t)=δ˙nt\lim_{t\to\infty}\dot{\delta}_{n}^{(t)}=\dot{\delta}_{n}^{t}. Thus, a longer sensing task duration and more recent time slots can result in more accurate estimates. Therefore, we use δ˙n(t)\dot{\delta}_{n}^{(t)} calculated from (10) instead of δ˙nt\dot{\delta}_{n}^{t} in the following calculations.

Calculate the user-related error δ¨it\bm{\ddot{\delta}_{i}^{t}}. Suppose we have a case where G^nt0\widehat{G}_{n}^{t}\neq 0, then δ¨it\ddot{\delta}_{i}^{t} can be calculated according to (4) and (5) as follows:

δ¨it=vi,nt(1+δ˙n(t))G^ntG^nt.\ddot{\delta}_{i}^{t}=\frac{v_{i,n}^{t}-\left(1+\dot{\delta}_{n}^{(t)}\right)\widehat{G}_{n}^{t}}{\widehat{G}_{n}^{t}}. (11)

However, the calculation in (11) neglects the situation in which G^nt=0\widehat{G}_{n}^{t}=0. Considering that the cosine similarity between vectors remains unchanged when the vectors are scaled, we scale the error vector [δ˙n(t),δ¨it][\dot{\delta}_{n}^{(t)},\ddot{\delta}_{i}^{t}] of data point di,ntd_{i,n}^{t} by G^nt\widehat{G}_{n}^{t} to form a new data feature Fi,nt=[G^ntδ˙n(t),vi,nt(1+δ˙n(t))G^nt]F_{i,n}^{t}=[\widehat{G}_{n}^{t}\dot{\delta}_{n}^{(t)},v_{i,n}^{t}-(1+\dot{\delta}_{n}^{(t)})\widehat{G}_{n}^{t}]. The scaled feature Fi,ntF_{i,n}^{t} can be utilized to calculate the degree of implication, and the calculation process remains valid even when G^nt=0\widehat{G}_{n}^{t}=0.

Calculate the degree of implication between sensing data points di,nt\bm{d_{i,n}^{t}} and dj,nt\bm{d_{j,n^{\prime}}^{t^{\prime}}}. We calculate the cosine similarity between the data feature vectors to determine the degree of implication as follows:

imp(di,nt,dj,nt)\displaystyle imp(d_{i,n}^{t},d_{j,n^{\prime}}^{t^{\prime}}) =cosinesimilarity(Fi,nt,Fj,nt)\displaystyle=cosine\ similarity(F_{i,n}^{t},F_{j,n^{\prime}}^{t^{\prime}}) (12)
=Fi,ntFj,ntFi,ntFj,nt\displaystyle=\frac{F_{i,n}^{t}\cdot F_{j,n^{\prime}}^{t^{\prime}}}{\|F_{i,n}^{t}\|\|F_{j,n^{\prime}}^{t^{\prime}}\|}

where Fi,ntF_{i,n}^{t} is the scaled feature vector of the sensing data point di,ntd_{i,n}^{t} and Fj,nt=[G^ntδ˙n(t),vj,nt(1+δ˙n(t))G^nt]F_{j,n^{\prime}}^{t^{\prime}}=[\widehat{G}_{n^{\prime}}^{t^{\prime}}\dot{\delta}_{n^{\prime}}^{(t)},v_{j,n^{\prime}}^{t^{\prime}}-(1+\dot{\delta}_{n^{\prime}}^{(t)})\widehat{G}_{n^{\prime}}^{t^{\prime}}] corresponds to the data point dj,ntd_{j,n^{\prime}}^{t^{\prime}}. Note that when we calculate imp(di,nt,dj,nt)imp(d_{i,n}^{t},d_{j,n^{\prime}}^{t^{\prime}}) according to (12), the circumstance-related error of region nn^{\prime} is δ˙n(t)\dot{\delta}_{n^{\prime}}^{(t)} instead of δ˙n(t)\dot{\delta}_{n^{\prime}}^{(t^{\prime})}. Since the time slots satisfy ttt^{\prime}\leq t according to Definition 1, indicating that time slot tt is more recent than the other time slots are, the estimated circumstance-related error of region nn^{\prime} should be updated to δ˙n(t)\dot{\delta}_{n^{\prime}}^{(t)}, which is more accurate.

Finally, the implication between any two data points can be extracted and utilized for the calculations in the subsequent module. Even in scenarios with bursty sensing data values, the implicit spatial and temporal correlations between data points remain stable [24] (e.g., when the traffic flow on a segment of a road undergoes a sudden change, the traffic flows on adjacent segments also exhibit certain changes). Therefore, although the prediction results based on historical sensing data are inaccurate in some cases, they retain spatio-temporal information, making it possible to extract accurate implications among data points. Thus, our method has greater advantages than prediction-based methods, which directly use predictions.

4.3 Reputation-Based TD Module

TD methods can estimate the ground truth based on submitted data and the weights of their providers. Traditional TD methods require a large amount of data for the same target (e.g., multiple MUs that sense the traffic flow data of the same segment of a road at the same time) to accurately estimate the corresponding ground truth. However, in an MCS system, the distributions of MU locations and periods may be sparse, and the amount of sensing data available for each target may vary, resulting in inaccurate estimates when traditional TD methods are used. Therefore, we design a reputation-based TD module that considers data implications, which allows for more accurate identification of low-quality sensing data and malicious MUs in scenarios characterized by data sparsity. Before providing a detailed description of this module, we first present the definitions that are used in the module.

Definition 4.

(Data quality level) The quality level of the sensing data point di,ntd_{i,n}^{t} is denoted as q(di,nt)q(d_{i,n}^{t}) and satisfies 0q(di,nt)10\leq q(d_{i,n}^{t})\leq 1. This quality level indicates the trustworthiness of the data point di,ntd_{i,n}^{t}. Additionally, the expected data quality level, denoted as q^(di,nt)[0,1]\hat{q}(d_{i,n}^{t})\in\left[0,1\right], is an estimation of q(di,nt)q(d_{i,n}^{t}) when the actual q(di,nt)q(d_{i,n}^{t}) is unknown. The estimate is calculated based on the credibility probabilities (i.e., the reputations) of the MUs who provided similar or the same data points as di,ntd_{i,n}^{t}.

Specifically, the closer the value of data point di,ntd_{i,n}^{t} is to the ground truth, the higher the trustworthiness of this data point is, thereby making q(di,nt)q(d_{i,n}^{t}) closer to 1.

Definition 5.

(Reputation) The reputation of MU ii at time slot tt is denoted as rit[0,1]r_{i}^{t}\in\left[0,1\right]. Reputation ritr_{i}^{t} represents the probability that MU ii is a normal and credible user. Furthermore, the reputation of an MU equals the mean data quality level of all the sensing data submitted by this MU [27].

Refer to caption
Figure 6: Reputation of an MU influenced by data quality levels.

Integrating a reputation mechanism into MCS can effectively reduce the misclassification of MUs. Even if normal MUs occasionally submit low-quality data under special circumstances (e.g., sudden device damage or abrupt weather changes), they can maintain a high reputation by consistently providing high-quality data over time [33]. In contrast, malicious MUs, which consistently submit low-quality or fake data, will maintain a persistently low reputation. Fig. 6 presents an example of how the reputation of a normal MU changes over time in relation to the quality levels of the submitted data. This example indicates that the reputations of MUs decrease when they submit low-quality sensing data but can remain high if they consistently submit high-quality data.

Like traditional TD methods [25, 26, 27], our reputation-based TD module consists of three steps, as follows.

Calculate the quality levels of the sensing data. We first consider the simple scenario shown in Fig. 5(a). This case includes two independent mobile users, MU ii and MU jj, who collect data in region nn at time slot tt. The values of their sensing data satisfy vi,nt=vj,ntv_{i,n}^{t}=v_{j,n}^{t}. In this case, vi,ntv_{i,n}^{t} and vj,ntv_{j,n}^{t} are incorrect data values only if both MU ii and MU jj provide low-quality data simultaneously. Owing to the independence of the MUs, the probability of them providing simultaneous low-quality data should be (1rit)(1rjt)(1-r_{i}^{t})(1-r_{j}^{t}). Therefore, the expected quality levels of data points di,ntd_{i,n}^{t} and dj,ntd_{j,n}^{t} are given by 1(1rit)(1rjt)1-(1-r_{i}^{t})(1-r_{j}^{t}), which represents the probability that data points di,ntd_{i,n}^{t} and dj,ntd_{j,n}^{t} are correct. By generalizing the results to common scenarios, the expected quality level of data point di,ntd_{i,n}^{t} can be calculated as follows:

q^(di,nt)=1i𝑰(vi,nt)(1rit)\hat{q}(d_{i,n}^{t})=1-\!\prod_{i\in\bm{I}(v_{i,n}^{t})}(1-r_{i}^{t}) (13)

where 𝑰(vi,nt)\bm{I}(v_{i,n}^{t}) represents the set of all MUs that submit data with values equal to vi,ntv_{i,n}^{t} in region nn at time slot tt.

However, 1rit1-r_{i}^{t} may approach 0 when MU ii has a high reputation, leading to the disappearance of the cumulative product term in (13). To address this problem, we define the reputation score of MU ii at time slot tt as

Rit=ln(1rit)R_{i}^{t}=-\ln(1-r_{i}^{t}) (14)

where Rit[0,+)R_{i}^{t}\in\left[0,+\infty\right) as rit[0,1]r_{i}^{t}\in[0,1], and limrit1Rit=+\lim_{r_{i}^{t}\to 1}R_{i}^{t}=+\infty. Similarly, the expected quality score of the sensing data point di,ntd_{i,n}^{t} is defined as

Q^(di,nt)=ln(1q^(di,nt))\widehat{Q}(d_{i,n}^{t})=-\ln\left(1-\hat{q}(d_{i,n}^{t})\right) (15)

where Q^(di,nt)[0,+)\widehat{Q}(d_{i,n}^{t})\in\left[0,+\infty\right) and limq^(di,nt)1Q^(di,nt)=+\lim_{\hat{q}(d_{i,n}^{t})\to 1}\widehat{Q}(d_{i,n}^{t})=+\infty. Subsequently, we can rewrite (13) with the reputation score and the expected quality score as follows:

Q^(di,nt)=i𝑰(vi,nt)Rit.\widehat{Q}(d_{i,n}^{t})=\sum_{i\in\bm{I}(v_{i,n}^{t})}R_{i}^{t}. (16)

Moreover, the implications between sensing data points can influence the quality levels of these data. Taking the scenario shown in Fig. 5(a) as an example, the values of the data points submitted by MUs ii and kk satisfy vi,ntvk,ntv_{i,n}^{t}\neq v_{k,n}^{t}. When imp(di,nt,dk,nt)>0imp(d_{i,n}^{t},d_{k,n}^{t})>0, data points di,ntd_{i,n}^{t} and dk,ntd_{k,n}^{t} exhibit mutual support, so the trustworthiness of data point dk,ntd_{k,n}^{t} can increase the trustworthiness of di,ntd_{i,n}^{t}. Therefore, the quality level of data point di,ntd_{i,n}^{t} increases the quality level of dk,ntd_{k,n}^{t}. In contrast, the quality level of data point di,ntd_{i,n}^{t} decreases the quality level of dk,ntd_{k,n}^{t} if imp(di,nt,dk,nt)<0imp(d_{i,n}^{t},d_{k,n}^{t})<0. Hence, the overall data quality score needs to consider both the expected quality and the influence of the implication as follows:

Q(di,nt)=Q^(di,nt)+ρdj,nt𝑫Q^(dj,nt)imp(di,nt,dj,nt)Q(d_{i,n}^{t})=\widehat{Q}(d_{i,n}^{t})+\rho\!\sum_{d_{j,n^{\prime}}^{t^{\prime}}\in\bm{D}^{\prime}}\!\widehat{Q}(d_{j,n^{\prime}}^{t^{\prime}})imp(d_{i,n}^{t},d_{j,n^{\prime}}^{t^{\prime}}) (17)

where 0<ρ<10<\rho<1 is a hyperparameter that determines the proportion of the overall quality score influenced by the implication. 𝑫\bm{D}^{\prime} is a data cache that stores the sensing data derived from the previous ll time slots and is updated at the end of each time slot. When the hyperparameter ll is set to a larger value, more historical sensing data are considered when calculating the implications that influence the quality of the data. Thus, the data quality evaluation becomes more accurate, but the calculation time also increases. Finally, the quality level of the sensing data point di,ntd_{i,n}^{t} can be calculated as follows:

q(di,nt)=1eQ(di,nt).q(d_{i,n}^{t})=1-e^{-Q(d_{i,n}^{t})}. (18)

Calculate the reputations of the MUs. The reputations of all the MUs need to be updated at each time slot t{1,2,,T}t\in\left\{1,2,...,T\right\}. First, we initialize the reputation for every MU at time slot tt as

rit=rit1,i{1,2,,N}.r_{i}^{t}=r_{i}^{t-1},\forall i\in\left\{1,2,...,N\right\}. (19)

Note that when t=1t=1, ritr_{i}^{t} is initialized as ri0r_{i}^{0}, which is the initial reputation of MU ii set by the sensing platform. Since each MU submits data at most once per time slot, we need only update the reputations of the MUs who submit data at the current time slot tt. Specifically, the reputations of these MUs are updated based on the quality of the data they submit at time slot tt. The reputation update function is designed according to [32]:

rit={12(1+α(q(di,nt)γ1γ)+(1α)(2rit1)),γq(di,nt)112(1+β(q(di,nt)γγ)+(1β)(2rit1)),0q(di,nt)<γ.r_{i}^{t}=\begin{cases}\begin{aligned} \frac{1}{2}\bigg{(}&1+\alpha\Big{(}\frac{q(d_{i,n}^{t})-\gamma}{1-\gamma}\Big{)}\\ &+\left(1-\alpha\right)\left(2r_{i}^{t}-1\right)\bigg{)},\end{aligned}&\gamma\leq q(d_{i,n}^{t})\leq 1\\ \begin{aligned} \frac{1}{2}\bigg{(}&1+\beta\Big{(}\frac{q(d_{i,n}^{t})-\gamma}{\gamma}\Big{)}\\ &+\left(1-\beta\right)\left(2r_{i}^{t}-1\right)\bigg{)},\end{aligned}&0\leq q(d_{i,n}^{t})<\gamma.\\ \end{cases} (20)

where α\alpha and β\beta are hyperparameters that determine the increasing and decreasing rates of the reputation update function, which satisfies 0<α<β<10<\alpha<\beta<1, and γ\gamma is the quality threshold set by the platform, which satisfies 0<γ<10<\gamma<1. Since the data quality level represents the trustworthiness of the data, sensing data with quality levels that are greater than or equal to γ\gamma are identified as high-quality data, whereas data with quality levels lower than γ\gamma are considered low-quality data. According to (20), the reputations of MUs who have submitted high-quality data increase, whereas those of MUs who have submitted low-quality data decrease.

The reputation update function shown in (20) satisfies the demands mentioned in [32] and [33] for monitoring the reputations of users in long-term multi-user scenarios and thus can be applied in MCS systems.

Refer to caption
Figure 7: Workflow of the reputation-based TD module.

Update the data quality levels and the reputations of the MUs alternately until convergence is reached. When the reputations of the MUs are updated according to (20), the expected quality levels of the data submitted by these MUs may change accordingly. Consequently, the data quality levels should be updated again according to (16) to (18). Moreover, since the data quality level q(di,nt)q(d_{i,n}^{t}) changes, the reputation of MU ii needs to be recalculated by (20). Therefore, as shown in Fig. 7, the data quality level q(di,nt)q(d_{i,n}^{t}) and the MU reputation ritr_{i}^{t} are alternately updated until convergence is achieved, resulting in the final quality level of data point di,ntd_{i,n}^{t} and the final reputation of MU ii at time slot tt.

Input: Sensing data 𝑫t={di,nt}i=1,n=1I,N\bm{D}^{t}=\left\{d_{i,n}^{t}\right\}_{i=1,n=1}^{I,N} and MU reputations {rit1}i=1I\left\{r_{i}^{t-1}\right\}_{i=1}^{I}
Output: Data quality levels {q(di,nt)}i=1,n=1I,N\left\{q(d_{i,n}^{t})\right\}_{i=1,n=1}^{I,N}, MU reputations {rit}i=1I\left\{r_{i}^{t}\right\}_{i=1}^{I}, and high-quality data cache 𝑫^t\bm{\widehat{D}}^{t}
Initialize 𝑫^t\bm{\widehat{D}}^{t}\leftarrow\varnothing, ritrit1r_{i}^{t}\leftarrow r_{i}^{t-1} for i=1,2,,Ii=1,2,...,I
Predict 𝑮^t\widehat{\bm{G}}^{t} with the CFSTTN
if t>lt>l then
       Remove old data from the cache: 𝑫𝑫𝑫tl\bm{D}^{\prime}\leftarrow\bm{D}^{\prime}\setminus\bm{D}^{t-l} 
end if
Add new data to the cache: 𝑫𝑫𝑫t\bm{D}^{\prime}\leftarrow\bm{D}^{\prime}\cup\bm{D}^{t}
repeat
       for i=1,2,,Ii=1,2,...,I do
             rioldritr_{i}^{old}\leftarrow r_{i}^{t}
             Δi0\varDelta_{i}\leftarrow 0
       end for
      for n=1,2,,Nn=1,2,...,N do
             Calculate cntc_{n}^{t} according to (7)
             Calculate δ˙nt\dot{\delta}_{n}^{t} with cntc_{n}^{t} according to (10)
            
       end for
      for di,nt𝐃td_{i,n}^{t}\in\bm{D}^{t} do
             for dj,nt𝐃d_{j,n^{\prime}}^{t^{\prime}}\in\bm{D}^{\prime} do
                   Calculate imp(di,nt,dj,nt)imp(d_{i,n}^{t},d_{j,n^{\prime}}^{t^{\prime}}) according to (12)
                  
             end for
            Calculate q(di,nt)q(d_{i,n}^{t}) according to (13) to (18)
             Update ritr_{i}^{t} with q(di,nt)q(d_{i,n}^{t}) according to (20)
             Calculate Δi=|ritriold|\varDelta_{i}=\left|r_{i}^{t}-r_{i}^{old}\right|
            
       end for
      Calculate Δmax=max{Δi}i=1I\varDelta_{max}=\max\left\{\varDelta_{i}\right\}_{i=1}^{I}
      
until Δmax<ϵ\varDelta_{max}<\epsilon
for di,nt𝐃td_{i,n}^{t}\in\bm{D}^{t} do
       if q(di,nt)γq(d_{i,n}^{t})\geq\gamma then
             𝑫^t𝑫^tdi,nt\bm{\widehat{D}}^{t}\leftarrow\bm{\widehat{D}}^{t}\cup d_{i,n}^{t} 
       end if
      
end for
Algorithm 1 Enhancing Data Quality with PRBTD

4.4 Data Enhancement with PRBTD

Finally, we present Algorithm 1, which outlines the process of enhancing the quality of the sensing data with the proposed PRBTD method at time slot tt. First, as shown in Fig. 2, the PRBTD method predicts the ground truth 𝑮^t\widehat{\bm{G}}^{t} via the CFSTTN prediction module. The data feature and implication calculation module subsequently uses the prediction results to calculate the sensing errors of the submitted data as features and uses these features to determine the degrees of implication between the data. The PRBTD method ultimately evaluates the quality levels of the data and the reputations of the MUs with the reputation-based TD module while considering the implications between the data. Moreover, sensing data with quality levels lower than γ\gamma are eliminated, whereas the remaining data are stored in a high-quality data cache 𝑫^t\bm{\widehat{D}}^{t}, thus achieving data cleansing and quality enhancement. Similarly, PRBTD can identify malicious MUs in MCS systems. After the task begins, the sensing platform can periodically check the reputations of the MUs and identify those whose reputations consistently remain below the predefined threshold as malicious MUs.

5 Experiments and Analysis

5.1 Experimental Setup

5.1.1 Dataset

The TaxiBJ dataset [14] is utilized for conducting simulation experiments in our MCS system to validate the ability of PRBTD to enhance the quality of the sensing data. This dataset comprises real-world taxi flow data with spatio-temporal correlations derived from Beijing city, making it suitable for simulating MCS systems. Specifically, the TaxiBJ dataset contains 7220 entries of data spanning four time intervals: 1st Jul. 2013 – 30th Oct. 2013, 1st Mar. 2014 – 30th Jun. 2014, 1st Mar. 2015 – 30th Jun. 2015, and 1st Nov. 2015 – 10th Apr. 2016. The data were sampled at 30-minute intervals within each period. In addition, the dataset partitions Beijing into a grid of 32×3232\times 32 spatial regions based on longitude and latitude information.

5.1.2 Setup

In our simulation, the sensing task spans T=120T=120 continuous time slots, each lasting 30 minutes, and covers N=32N=32 regions. Specifically, we use the data from the last 120 time slots within 32 randomly selected regions of the TaxiBJ dataset for the simulations. These data are regarded as the ground truth of the sensing data in the task. Our MCS system includes 90 normal MUs and 10 malicious MUs that participate in the sensing task. We set the sensing errors of normal MUs to 0, ensuring that the sensing data they submit are equal to the corresponding ground truth. However, the sensing errors of malicious MUs are sampled from a Gaussian distribution N(μ,σ2)N(\mu,\sigma^{2}). Specifically, we set μ=0.3\mu=0.3 and σ=0.15\sigma=0.15 as the basic setup to simulate the sensing errors carried by the data provided by malicious MUs. Additionally, each MU is set to submit sensing data a total of kk times during the sensing task, where 0<k<T0<k<T. In our experiments, we set k=30k=30 to simulate a scenario where MUs do not need to collect and submit data frequently. When calculating the reputations of MUs, we initialize the reputation of each MU as ri0=0.5r_{i}^{0}=0.5 and set the data quality threshold as γ=0.5\gamma=0.5. The reason is that the reputations of MUs and the quality levels of data represent their respective probabilities of trustworthiness, making 0.5 a suitable initial value and threshold. Moreover, the increasing and decreasing rates in the reputation update function (20) are set as α=0.018\alpha=0.018 and β=0.06\beta=0.06, respectively, whereas the proportion of the data quality levels influenced by implications is set as ρ=0.02\rho=0.02. These parameters are hyperparameters in our method, and their specific settings were determined through a series of preliminary experiments. Finally, the detailed parameter settings for the simulations are listed in Table II.

TABLE II: Simulation Parameters
Parameters Values
Number of MUs (II) 100
Number of times a MU performs the task (kk) 30
Initial reputation of each MU (ri0r_{i}^{0}) 0.5
Mean of the sensing errors induced by malicious MUs (μ\mu) 0.3
Standard deviation of the sensing errors induced by malicious MUs (σ\sigma) 0.1
Increase rate of reputation (α\alpha) 0.018
Decrease rate of reputation (β\beta) 0.06
The data quality threshold (γ\gamma) 0.5
Proportion of the implication-influenced part of the quality score (ρ\rho) 0.02
Size of the data cache 𝑫\bm{D}^{\prime} (ll) 5
Number of time slots used as the prediction input (pp) 5
Ending factor of PRBTD (ϵ\epsilon) 0.001

In the CFSTTN, we set the network parameters in accordance with the setup used by the GLSTTN [17]. The first 7100 time slots of data derived from the TaxiBJ dataset are used for training, whereas the last 120 time slots are used for validation. We adopt the adaptive moment estimation (Adam) optimizer with an initial learning rate of 0.001 for training. This learning rate is reduced to 10410^{-4} at 50% of the total training epochs and further reduced to 10510^{-5} at 75% of the total. The total number of training epochs is set to 300, with a batch size of 32 per epoch. Additionally, we set the prediction input of the CFSTTN module to consist of data from the preceding 3 time slots along with data from 1 day and 1 week prior. We implement the CFSTTN via PyTorch 1.1.0. All the experiments are conducted on a system equipped with an NVIDIA 1080Ti GPU. Our source code is publicly available at https://github.com/glab2019/PRBTD.

5.2 Evaluation of the CFSTTN

We first evaluate the convergence performance of the CFSTTN prediction module in our PRBTD framework. The training and validation loss curves produced by the CFSTTN on the TaxiBJ dataset are shown in Fig. 8. The CFSTTN converges after approximately 220 epochs, with an average training time of 82.56 seconds per epoch, leading to an average convergence duration of 5.04 hours. These results were obtained after repeating the training 6 times and indicate that the convergence time of the CFSTTN is acceptable.

Refer to caption
Figure 8: Convergence of the CFSTTN.

We analyze the prediction performance of the CFSTTN via the root mean square error (RMSE) and the coefficient of determination (R2). The RMSE measures the difference between the prediction results and the ground truth. Therefore, a lower RMSE reflects better performance. R2 represents the explanatory ability of the prediction method. An R2 closer to 1 indicates a higher prediction accuracy. In our experiment, which used the TaxiBJ dataset, the CFSTTN achieves an RMSE of 16.88 and an R2 of 0.9583. Moreover, we compare our CFSTTN with the following 6 baselines:

  • ARIMA. The auto-regressive integrated moving average (ARIMA) is a popular time series prediction model that combines auto-regression, differencing, and moving average to predict future values.

  • SARIMA. The seasonal auto-regressive integrated moving average (SARIMA) extends ARIMA by including seasonal components, allowing it to model and predict data with seasonal patterns.

  • VAR. The vector auto-regressive (VAR) model is a multivariate time series model that captures linear interdependencies among multiple time series.

  • ST-ANN. The spatio-temporal artificial neural network (ST-ANN) extracts spatial and temporal features and uses an artificial neural network to model complex dependencies for prediction.

  • DeepST. This deep neural network-based model employs convolutional neural networks (CNNs) to model spatial dependencies and recurrent neural networks (RNNs) to capture temporal dependencies.

  • ST-ResNet. The spatio-temporal residual networks (ST-ResNet) utilize residual learning to capture spatial and temporal dependencies in data.

The model comparisons are presented in Table III, demonstrating that our CFSTTN outperforms the baselines.

TABLE III: Comparison Among the RMSEs Induced by Different Methods on TaxiBJ
Method RMSE
ARIMA 22.78
SARIMA 26.88
VAR 22.88
ST-ANN 19.57
DeepST 18.18
ST-ResNet 17.30
CFSTTN (ours) 16.88
Refer to caption

(a) Ground-truth and prediction curves.

Refer to caption

(b) Prediction errors.

Refer to caption

(c) CDF of the absolute prediction error.

Figure 9: Comparison of prediction results for region (24,9).

Finally, we analyze the ability of the CFSTTN by visualizing its prediction results obtained on the TaxiBJ dataset and comparing them with the predictions of ST-ResNet, which is the best-performing method among the baselines. We take region (24,9) as an example. Fig. 9(a) shows that the prediction results of the CFSTTN for this region closely align with the ground truth and are better than those of ST-ResNet. Moreover, Figs. 9(b) and (c) show that the errors of the CFSTTN in the prediction results are smaller than those of ST-ResNet, thus highlighting that the CFSTTN has a good prediction ability.

5.3 Experimental Results

We analyze the performance of the PRBTD method based on the following metrics.

  • F1-score. F1-scores of the methods when classifying normal and malicious MUs. The F1-score is a metric that combines precision and recall and can be used to evaluate the ability of a method to correctly classify normal and malicious MUs in the MCS system.

  • Reputation distance. Interclass reputation distance between normal MUs and malicious MUs. A larger reputation distance indicates better performance in terms of classifying normal MUs and malicious MUs in MCS systems.

  • Noise reduction ratio. Proportion of noise removed relative to the total noise after applying the enhancement method. Specifically, we calculate the noise reduction ratio as 1 minus the ratio of the average noise in the enhanced data to the average noise in the original data. A higher noise reduction ratio indicates better performance in identifying low-quality data and enhancing data quality.

Each set of experiments in our simulations is repeated 6 times, and the results reflect the mean results obtained across the repetitions.

5.3.1 Baseline Comparisons

We compare PRBTD with the following 4 baseline methods.

  • Weight truth inference (WEI). A data aggregation method. This method calculates the average of all the data in region nn prior to time slot tt as the ground truth.

  • CFSTTN-based (CNB). A prediction-based method. This method directly uses the prediction results obtained from the CFSTTN as the ground truth.

  • TD [28]. A traditional user-weight-based TD method. This method iteratively estimates the ground truth based on the data submitted by MUs and the weights of these MUs.

  • Decentralized trust inference (DTI) [9]. A data filling method. This method uses trusted baseline data (e.g., data collected by dispatching UAVs during the task) to construct a sparse data matrix, which is then completed via Bayesian probabilistic matrix factorization (BPMF). Specifically, we utilize historical data as trusted baseline data when deploying DTI in our experiments. The completed matrix serves as the estimated ground truth.

TABLE IV: Performances of the PRBTD and Baseline Methods in the Scenario with the Basic Setup
Method F1-score Reputation distance Noise reduction ratio
WEI 0.9667 0.1004 0.2687
CNB 0.9741 0.1420 0.2863
TD 0.9778 0.2951 0.4252
DTI 0.9870 0.1562 0.3926
PRBTD (ours) 0.9907 0.3871 0.8100

We choose these methods as baselines due to their reproducibility, strong representativeness and high performance among various types of methods, as well as their good adaptability to our MCS scenario. To use these methods in our experiments, we adjusted them to evaluate the quality of a sensing data point based on the distance between the data value and the estimated ground truth and update the reputations of the MUs according to (20).

We first analyze the runtime efficiency of our PRBTD method. Specifically, in our experiments, the average runtime of the PRBTD method per slot was 0.5097 seconds, which is significantly less than the 30-minute duration we set for each time slot. Therefore, in our experiments, the PRBTD method is able to perform real-time data quality analysis.

The performances of the PRBTD method and baselines are presented in Table IV, which shows that PRBTD outperforms the other methods in terms of the three metrics. Specifically, the proposed PRBTD framework achieves the highest F1-score and a 31% improvement in reputation distance over TD, the most commonly used method in MCS. Additionally, the noise reduction ratio of PRBTD is 0.8100, meaning that after executing the data quality enhancement via PRBTD, the average noise in the sensing data decreases by 81%. This reduction demonstrates the strong denoising efficiency of PRBTD over the baselines.

Refer to caption

(a) F1-score.

Refer to caption

(b) Reputation distance.

Refer to caption

(c) Noise reduction ratio.

Figure 10: Comparisons among the PRBTD and baseline methods with different proportions of ground truth data used as historical data.
Refer to caption

(a) F1-score.

Refer to caption

(b) Reputation distance.

Refer to caption

(c) Noise reduction ratio.

Figure 11: Comparisons among the PRBTD and baseline methods in scenarios with and without bursty data values.

5.3.2 Experiments in Scenarios with Noisy Historical Data

In practical MCS scenarios, historical data usually contain a certain amount of noise, which affects the performance of learning methods that rely on such data. To investigate the impact of noisy historical data on our PRBTD framework, we introduce noise to a certain proportion of the data while keeping the remaining data unchanged to create historical data with varying proportions of ground truth data. Specifically, we multiply the values of the selected data by 1+ξ1+\xi, where ξN(0,0.12)\xi\sim N(0,0.1^{2}). We construct sets of historical data with 50%, 60%, 70%, 80%, and 90% ground truth data and evaluate the performance of the PRBTD, DTI, and CNB methods on these sets. We did not include the other two baselines, TD and WEI, because they do not utilize historical data. As shown in Fig. 10, although some biased results exist due to experimental randomness, the overall trend indicates that the effectiveness of all methods tends to decline as the proportion of ground truth in the historical data decreases. This decline is reasonable, as an increase in noise within the training data is likely to affect the prediction results of the methods. However, PRBTD consistently outperforms CNB and DTI. In practice, historical data typically undergo some level of data cleaning and enhancement, meaning that noise is usually limited. Even with 50% of the historical data containing noise, PRBTD still performs effectively and better than traditional prediction-based methods do, demonstrating its independence from the ground truth.

TABLE V: Performances of the PRBTD and Baseline Methods in the Scenario with Bursty Data Values
Method F1-score Reputation distance Noise reduction ratio
WEI 0.9667 0.0927 0.2657
CNB 0.9704 0.1309 0.2679
TD 0.9778 0.2953 0.4240
DTI 0.9852 0.1554 0.3726
PRBTD (ours) 0.9889 0.3686 0.8089

5.3.3 Experiments in Scenarios with Bursty Data Values

In some cases, the sensing data of tasks may encompass data with bursty values, which can lead to inaccurate ground-truth predictions by prediction modules. To analyze the performance of PRBTD in scenarios with bursty data values, we randomly select 3 sets of data and reduce their values by 50% for comparison experiments. Specifically, each set consists of data derived from 7 continuous time slots in 4 consecutive task regions. The simulation results are shown in Table V. Furthermore, we compare the results obtained in scenarios with and without bursty data values. The comparisons presented in Fig. 11 show that in the scenario with bursty data values, the performances of the PRBTD, DTI, and CNB methods decreases, whereas those of the TD and WEI methods remain stable. These results are reasonable because the PRBTD and CNB methods rely on predictions of sensing data, whose accuracy is impacted by the occurrence of bursty data values, and DTI has a limited ability to handle sudden changes in data values. However, the performance decline of PRBTD is smaller than that of CNB, confirming that our method effectively mitigates the negative impact of prediction inaccuracies. Overall, our PRBTD framework still outperforms the other methods in terms of all three metrics, thus demonstrating strong robustness.

Refer to caption

(a) F1-score.

Refer to caption

(b) Reputation distance.

Refer to caption

(c) Noise reduction ratio.

Figure 12: Comparisons among the PRBTD and baseline methods under different data sparsity levels.
Refer to caption

(a) F1-score.

Refer to caption

(b) Reputation distance.

Refer to caption

(c) Noise reduction ratio.

Figure 13: Comparisons among the PRBTD and baseline methods under different levels of sensing noise.

5.3.4 Experiments in Scenarios with Sparse Data

In MCS, the sensing data submitted by MUs may be sparse due to the uncertainty of the MU locations and period distributions, resulting in inaccurate quality evaluations in cases with insufficient data. Thus, we conduct experiments under different degrees of data sparsity by adjusting the amount of sensing data submitted by the MUs. Specifically, we analyze the performances of different methods in 5 scenarios, which use different percentages (i.e., 60%, 70%, 80%, 90%, and 100%) of the total amount of submitted sensing data for quality evaluation purposes. The simulation results are shown in Fig. 12. The results show that as the sparsity of the data increases, the performance of the PRBTD, DTI, TD, and WEI methods gradually decreases, whereas that of the CNB method remains stable. The reason for these declines is the reduction in the amount of available data, which leads to a decrease in the accuracy of the ground-truth estimates yielded by the PRBTD, DTI, TD, and WEI methods. Nevertheless, in most case, PRBTD still outperforms these baseline methods, as it utilizes the implications between the sensing data derived from different locations and periods, which indirectly increases the amount of data available for quality evaluation.

TABLE VI: Performances of the PRBTD and Baseline Methods in the Scenario with Lower Levels of Noise
Method F1-score Reputation distance Noise reduction ratio
WEI 0.9333 0.0480 0.2022
CNB 0.9611 0.0950 0.2145
TD 0.9556 0.2099 0.3937
DTI 0.9422 0.0858 0.3229
PRBTD (ours) 0.9667 0.2954 0.6787

5.3.5 Experiments in Scenarios with Lower Levels of Sensing Noise

Finally, to evaluate the effectiveness of the PRBTD method in extreme MCS scenarios, we analyze the performances of different methods in cases where the malicious MUs are more sophisticated (i.e., the sensing data submitted by the malicious MUs have lower levels of sensing noise, and some malicious MUs even submit data without noise in some time slots). Specifically, we adjust μ\mu to 0.15 and keep the other parameters unchanged to reduce the sensing noise of the data submitted by malicious MUs. Table VI and Fig. 13 present and compare the performances of the different methods. The results show that with the reduction in the amount of sensing noise contained in the data derived from malicious MUs, the performances of all methods decrease because of the difficulty in identifying incorrect or fake data with less noise. However, PRBTD remains the best approach, outperforming the other methods in terms of the three metrics, thus indicating its robustness in cases involving sophisticated malicious MUs.

6 Conclusion

In this paper, we combine prediction, TD, and reputation mechanisms for the first time to propose a novel framework, PRBTD, for evaluating and enhancing the quality of spatio-temporally correlated sensing data in MCS systems. We propose using the implications between data points to assist in data quality evaluation and evaluate the two significant benefits of integrating implications: reducing the impact of inaccurate prediction results on data quality evaluation and increasing the amount of data available for inferring and verifying the authenticity of data. This approach makes PRBTD superior to existing methods regardless of the presence of bursty values or data sparsity.

However, our PRBTD algorithm has limitations. As its calculations rely on historical data, it cannot achieve optimal performance initially in the absence of historical data or prior knowledge. Additionally, PRBTD must learn and extract the implications between data, so it can handle data with implicit correlations only. To address these limitations, future work will focus on developing this method further to extend its applicability.

Overall, the proposed PRBTD framework makes significant advancements in the quality evaluation of sensing data in MCS systems, thereby promising more accurate and reliable data classification even in challenging scenarios.

References

  • [1] A. Capponi, C. Fiandrino, B. Kantarci, L. Foschini, D. Kliazovich, and P. Bouvry, “A survey on mobile crowdsensing systems: Challenges, solutions, and opportunities,” IEEE Commun. Surveys Tuts., vol. 21, no. 3, pp. 2419–2465, 3rd Quart. 2019.
  • [2] S. Mathur, S. Kaul, M. Gruteser, and W. Trappe, “ParkNet: a mobile sensor network for harvesting real time vehicular parking information,” in Proc. ACM MobiHoc S3 Workshop, 2009, pp. 25–28.
  • [3] Nextdoor, [Online]. Available: https://nextdoor.com/
  • [4] Google crowdsource, [Online]. Available: https://crowdsource.google.com/about/
  • [5] M. Li, J. Wu, W. Wang, and J. Zhang, “Toward privacy-preserving task assignment for fully distributed spatial crowdsourcing,” IEEE Internet Things J., vol. 8, no. 18, pp. 13991–14002, Sep. 2021.
  • [6] Z. Wang, J. Hu, R. Lv, J. Wei, Q. Wang, D. Yang, and H. Qi, “Personalized privacy-preserving task allocation for mobile crowdsensing,” IEEE Trans. Mobile Comput., vol. 18, no. 6, pp. 1330–1341, Jun. 2019.
  • [7] L. Xiao, Y. Li, G. Han, H. Dai, and H. V. Poor, “A secure mobile crowdsensing game with deep reinforcement learning,” IEEE Trans. Inf. Forensics Secur., vol. 13, no. 1, pp. 35–47, Jan. 2018.
  • [8] C. Meng, H. Xiao, L. Su, and Y. Cheng, “Tackling the redundancy and sparsity in crowd sensing applications,” in Proc. 14th ACM Conf. Embedded Netw. Sens. Syst., 2016, pp. 150–163.
  • [9] X. Yang, Z. Zeng, A. Liu, N. N. Xiong, T. Wang, and S. Zhang, “A decentralized trust inference approach with intelligence to improve data collection quality for mobile crowd sensing,” Inf. Sci., vol. 644, pp. 119286, Oct. 2023.
  • [10] S. B. Azmy, N. Zorba, and H. S. Hassanein, “Quality estimation for scarce scenarios within mobile crowdsensing systems,” IEEE Internet Things J., vol. 7, no. 11, pp. 10955–10968, Nov. 2020.
  • [11] X. Zhang and X. Gong, “Online data quality learning for quality-aware crowdsensing,” in Proc. 16th Annu. IEEE Int. Conf. Sens. Commun. Netw., 2019, pp. 1–9.
  • [12] X. Gao, S. Chen, and G. Chen, “MAB-based reinforced worker selection framework for budgeted spatial crowdsensing,” IEEE Trans. Knowl. Data Eng., vol. 34, no. 3, pp. 1303–1316, Mar. 2022.
  • [13] Y. Xie, J. Niu, Y. Zhang, and F. Ren, “Multisize patched spatial-temporal transformer network for short- and long-term crowd flow prediction,” IEEE Trans. Intell. Transp. Syst., vol. 23, no. 11, pp. 21548–21568, Nov. 2022.
  • [14] J. Zhang, Y. Zheng, and D. Qi, “Deep spatio-temporal residual networks for citywide crowd flows prediction,” in Proc. 31st AAAI Conf. Artif. Intell., 2017, pp. 1655–1661.
  • [15] X. Min, J. Hu, Q. Chen, T. Zhang, and Y. Zhang, “Short-term traffic flow forecasting of urban network based on dynamic STARIMA model,” in Proc. 12th IEEE Int. Conf. Intell. Transp. Syst., 2009, pp. 1–6.
  • [16] Y. Yao, B. Gu, Z. Su, and M. Guizani, “MVSTGN: A multi-view spatial-temporal graph network for cellular traffic prediction,” IEEE Trans. Mobile Comput., vol. 22, no. 5, pp. 2837–2849, May 2023.
  • [17] B. Gu, J. Zhan, S. Gong, W. Liu, Z. Su, and M. Guizani, “A spatial-temporal transformer network for city-level cellular traffic analysis and prediction,” IEEE Trans. Wireless Commun., vol. 22, no. 12, pp. 9412–9423, Dec. 2023.
  • [18] F. Xu, Y. Lin, J. Huang, D. Wu, H. Shi, J. Song, and Y. Li, “Big data driven mobile traffic understanding and forecasting: A time series approach,” IEEE Trans. Services Comput., vol. 9, no. 5, pp. 796–805, Sep./Oct. 2016.
  • [19] Y. Wang, M. Papageorgiou, and A. Messmer, “Real-time freeway traffic state estimation based on extended Kalman filter: A case study,” Transp. Sci., vol. 41, no. 2, pp. 167–181, May 2007.
  • [20] Z. Zhou, H. Liao, B. Gu, K. M. S. Huq, S. Mumtaz, and J. Rodriguez, “Robust mobile crowd sensing: When deep learning meets edge computing,” IEEE Netw., vol. 32, no. 4, pp. 54–60, Jul./Aug. 2018.
  • [21] N. Owoh, J. Riley, M. Ashawa, S. Hosseinzadeh, A. Philip, and J. Osamor, “An adaptive temporal convolutional network autoencoder for malicious data detection in mobile crowd sensing,” Sensors, vol. 24, no. 7, Apr. 2024, Art. no. 2353.
  • [22] Y. Zheng, H. Duan, and C. Wang, “Learning the truth privately and confidently: Encrypted confidence-aware truth discovery in mobile crowdsensing,” IEEE Trans. Inf. Forensics Security., vol. 13, no. 10, pp. 2475–2489, Oct. 2018.
  • [23] C. Miao, L. Su, W. Jiang, Y. Li, and M. Tian, “A lightweight privacy-preserving truth discovery framework for mobile crowd sensing systems,” in Proc. IEEE Int. Conf. Comput. Commun., 2017, pp. 1–9.
  • [24] W. Mo, Z. Li, Z. Zeng, N. N. Xiong, S. Zhang, and A. Liu, “SCTD: A spatiotemporal correlation truth discovery scheme for security management of data platform,” Future Gener. Comput. Syst., vol. 139, pp. 109–125, Feb. 2023.
  • [25] R. W. Ouyang, M. Srivastava, A. Toniolo, and T. J. Norman, “Truth discovery in crowdsourced detection of spatial events,” in Proc. 23rd ACM Int. Conf. Inform. Knowl. Manage., 2014, pp. 461–470.
  • [26] Y. Cheng, J. Ma, Z. Liu, Z. Li, Y. Wu, C. Dong, and R. Li, “A privacy-preserving and reputation-based truth discovery framework in mobile crowdsensing,” IEEE Trans. Dependable Secure Comput., vol. 20, no. 6, pp. 5293–5311, Nov./Dec. 2023.
  • [27] X. Yin, J. Han, and P. S. Yu, “Truth discovery with multiple conflicting information providers on the web,” in Proc. 13th ACM SIGKDD Int. Conf. Knowl. Discovery Data Mining, 2007, pp. 1048–1052.
  • [28] Y. Du, H. Xu, Y.-E. Sun, and L. Huang, “A general fine-grained truth discovery approach for crowdsourced data aggregation,” in Proc. Int. Conf. Database Syst. Adv. Appl., 2017, pp. 3–18.
  • [29] G. Xu, H. Li, S. Liu, M. Wen, and R. Lu, “Efficient and privacy-preserving truth discovery in mobile crowd sensing systems,” IEEE Trans. Veh. Technol., vol. 68, no. 4, pp. 3854–3865, April. 2019.
  • [30] Y. Kang, A. Liu, N. N. Xiong, S. Zhang, T. Wang, and M. Dong, “DTD: An intelligent data and bid dual truth discovery scheme for MCS in IIoT,” IEEE Internet Things J., vol. 11, no. 2, pp. 2507–2519, Jan. 2024.
  • [31] H. Tian, W. Sheng, H. Shen, and C. Wang, “Truth finding by reliability estimation on inconsistent entities for heterogeneous data sets,” Knowl. Based Syst., vol. 187, Jan. 2020, Art. no. 104828.
  • [32] K. Zhao, S. Tang, B. Zhao, and Y. Wu, “Dynamic and privacy-preserving reputation management for blockchain-based mobile crowdsensing,” IEEE Access, vol. 7, pp. 74694–74710, 2019.
  • [33] T. Salman, R. Jain, and L. Gupta, “A reputation management framework for knowledge-based and probabilistic blockchains,” Proc. IEEE Int. Conf. Blockchain, 2019, pp. 520–527.
  • [34] D. C. Nguyen, M. Ding, Q.-V. Pham, P. N. Pathirana, L. B. Le, A. Seneviratne, J. Li, D. Niyato, and H. V. Poor, “Federated learning meets blockchain in edge computing: Opportunities and challenges,” IEEE Internet Things J., vol. 8, no. 16, pp. 12806–12825, Aug. 2021.
  • [35] L. Wang, Z. Yu, D. Zhang, B. Guo, and C. H. Liu, “Heterogeneous multi-task assignment in mobile crowdsensing using spatiotemporal correlation,” IEEE Trans. Mobile Comput., vol. 18, no. 1, pp. 84–97, Jan. 2019.
  • [36] Y. Wang, Z. Cai, Z.-H. Zhan, Y.-J. Gong, and X. Tong, “An optimization and auction-based incentive mechanism to maximize social welfare for mobile crowdsourcing,” IEEE Trans. Computat. Social Syst., vol. 6, no. 3, pp. 414–429, Jun. 2019.
  • [37] H. Jin, L. Su, D. Chen, K. Nahrstedt, and J. Xu, “Quality of information aware incentive mechanisms for mobile crowd sensing systems,” in Proc. 16th ACM Int. Symp. Mobile Ad Hoc Netw. Comput., 2015, pp. 167–176.
  • [38] M. Xiao, B. An, J. Wang, G. Gao, S. Zhang, and J. Wu, “CMAB-based reverse auction for unknown worker recruitment in mobile crowdsensing,” IEEE Trans. Mobile Comput., vol. 21, no. 10, pp. 3502–3518, Oct. 2022.
  • [39] B. Gu, X. Yang, Z. Lin, W. Hu, M. Alazab, and R. Kharel, “Multiagent actor-critic network-based incentive mechanism for mobile crowdsensing in industrial systems,” IEEE Trans Ind. Informat., vol. 17, no. 9, pp. 6182–6191, Sep. 2021.
  • [40] H. Gao, C. H. Liu, J. Tang, D. Yang, P. Hui, and W. Wang, “Online quality-aware incentive mechanism for mobile crowd sensing with extra bonus,” IEEE Trans. Mobile Comput., vol. 18, no. 11, pp. 2589–2603, Nov. 2019.
  • [41] B. Guo, H. Chen, Z. Yu, W. Nan, X. Xie, D. Zhang, and X. Zhou, “TaskMe: Toward a dynamic and quality-enhanced incentive mechanism for mobile crowd sensing,” Int. J. Hum.-Comput. Stud., vol. 102, pp. 14–26, Jun. 2017.
  • [42] D. Zhang, J. Yang, J. Yang, and A. F. Frangi, “Two-dimensional PCA: A new approach to appearance-based face representation and recognition,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 26, no. 1, pp. 131–137, Jan. 2004.