[BACK]
Computers, Materials & Continua
DOI:10.32604/cmc.2020.011995
images
Article

A Crowdsourcing Recommendation that Considers the Influence of Workers

Zhifang Liao1, Xin Xu1,*, Peng Lan1, Liu Yang1,*, Yan Zhang2 and Xiaoping Fan3,*

1School of Computer Science and Engineering, Central South University, Changsha, 410075, China
2Glasgow Caledonian University, Glasgow, G4 0BA, UK
3Hunan University of Finance and Economics, Changsha, 410205, China
*Corresponding Author: Xiaoping Fan. Email: Xpfan@csu.edu.cn
Received: 09 June 2020; Accepted: 04 July 2020

Abstract: In the context of the continuous development of the Internet, crowdsourcing has received continuous attention as a new cooperation model based on the relationship between enterprises, the public and society. Among them, a reasonably designed recommendation algorithm can recommend a batch of suitable workers for crowdsourcing tasks to improve the final task completion quality. Therefore, this paper proposes a crowdsourcing recommendation framework based on workers’ influence (CRBI). This crowdsourcing framework completes the entire process design from task distribution, worker recommendation, and result return through processes such as worker behavior analysis, task characteristics construction, and cost optimization. In this paper, a calculation model of workers’ influence characteristics based on the ablation method is designed to evaluate the comprehensive performance of workers. At the same time, the CRBI framework combines the traditional open-call task selection mode, builds a new task characteristics model by sensing the influence of the requesting worker and its task performance. In the end, accurate worker recommendation and task cost optimization are carried out by calculating model familiarity. In addition, for recommending workers to submit task answers, this paper also proposes an aggregation algorithm based on weighted influence to ensure the accuracy of task results. This paper conducts simulation experiments on some public datasets of AMT, and the experimental results show that the CRBI framework proposed in this paper has a high comprehensive performance. Moreover, CRBI has better usability, more in line with commercial needs, and can well reflect the wisdom of group intelligence.

Keywords: Crowdsourcing; recommendation framework; workers’; influence; worker recommendation; weighted voting

1  Introduction

Crowd-based cooperative computing is a current research hotspot in cloud computing and big data. Crowdsourcing, an important branch of collaborative intelligent computing, integrates computers of unknown mass on the Internet to accomplish tasks that individual computers alone cannot accomplish [1]. The concept of crowdsourcing was proposed by Howe in June 2006 [2], and its main participants include task requesters (i.e., requesters) and task completers (i.e., workers), who are linked together through human intelligence tasks. Crowdsourcing is a problem solving strategy that can solve complex problems in various scenarios, such as audio annotation, image classification, and software development. It allows the crowd to complete tasks by accepting tasks and participate in solving complex problems. The task requester completes the work of distributing the crowdsourcing task, and the crowdsourcing worker selects, accepts, and submits the answer to the task. However, if too many tasks are distributed to workers, the workers will become bored with the tasks, and the quality of tasks will be reduced. Therefore, how to choose the right workers from a large crowd has be-come a key issue for crowdsourcing recommendation.

At present, crowdsourcing recommendations generally target workers’ behavior analysis and task characteristic construction. For example, Guo et al. [3] proposed a recommendation mechanism based on interest collaboration that used the professional level and interests of workers as the bases for recommendation. Difallah et al. [4] generated recommendations based on the degree of matching between workers and task descriptions. However, these studies did not consider the recommendation process as a whole and ignored the control of recommendation costs and the treatment of recommendation results during the recommendation process. Furthermore, the ability weight measurement of crowdsourcing workers is a key factor in the rational distribution of crowdsourcing tasks. Therefore, identifying an effective indicator to evaluate the characteristics of workers and using this indicator to recommend workers is a key task in crowdsourcing. Zhong et al. [5] have studied this issue, and most related studies use workers’ task accuracy and preferences as indicators. However, if the task accuracy rate of the workers is used as an index for capability evaluation, it is impossible to distinguish between new manual workers and skilled workers with the same accuracy. If the completion accuracy of workers on different tasks is considered, the complexity of tasks is ignored. Therefore, we propose a new indicator of workers ‘influence, which takes into account workers’ correct rate, error rate, and task complexity. It can not only evaluate the comprehensive characteristics of workers, but also distinguish the behavior differences between workers in more detail. At the same time, for the problem of recommending cold start, this paper proposes a worker recommendation algorithm based on open-call mode to perform worker recommendation and cost optimization. In addition, for the task answers submitted by workers, we propose a weighted voting algorithm based on influence to aggregate workers’ answers to ensure the accuracy of task results. Therefore, this paper proposes a crowdsourcing recommendation framework based on the workers’ influence (CRBI).

The main contributions of this paper are as follows:

1.    Proposing a new evaluation index of worker influence and designing a calculation model of worker influence characteristics based on the ablation method to effectively evaluate the comprehensive performance of workers.

2.    Proposing a crowdsourcing task recommendation algorithm based on open-call mode to achieve effective recommendation efficiency and optimize task costs.

3.    Proposing a weighted voting algorithm based on worker influence that cooper-ates with the recommendation algorithm to achieve good performance.

The rest of this paper is organized as follows. Section 2 describes the existing recommendation methods. Section 3 introduces the system model of the CRBI framework. Section 4 introduces the specific process design of the CRBI frame-work. Section 5 analyzes the experimental results. Finally, Section 6 draws conclusions and discusses directions for future research.

2  Related Work

Crowdsourcing can effectively reduce the cost of innovation, break the boundaries of the source of innovation, fully develop the wisdom of the public, and increase public participation. The research topics in crowdsourcing mainly focuses on the decomposition of crowdsourcing tasks, spatial crowdsourcing, incentive mechanism, workers’ behavior analysis, aggregation of crowdsourcing results, and task selection. Moreover, crowdsourcing, as a problem-solving strategy, is also continuously combined with blockchain, Internet of Things [6], cloud computing [7] and social network, which can better solve complex problems in various scenarios. Crowdsourcing is also widely used in teaching. Wang et al. [8] proposed a crowdsourcing-based framework for evaluating teaching quality in classrooms. Moreover, benefiting from the development of public crowdsourcing platforms, such as Amazon Machine Turk (AMT), Crowdflower (CF), TopCoder, etc., the crowd can accept and solve tasks through the crowdsourcing platform. The research on crowdsourcing has focused on task result aggregation, incentive mechanisms, spatial crowdsourcing, task decomposition, and quality control. Although crowdsourcing has become a promising problem-solving model, there is a lack of overall design research on crowdsourcing processes. Therefore, designing an efficient and available crowdsourcing recommendation framework is important for crowdsourcing systems.

The research on crowdsourcing recommendation is based mainly on traditional recommendation [9], including probability matrix decomposition [10], content-based recommendation [11], collaborative filtering [12], and hybrid recommendation algorithms [13]. Collaborative filtering, which generates recommendations based on the similarity of workers or tasks, is one of the most commonly used algorithms in crowdsourcing recommendation. Safran et al. [14] proposed two top-N recommendation algorithms for crowdsourcing systems and combined them with algorithms such as minimum mean square error and Bayesian personalized ranking for optimization. Generally, workers participating in crowdsourcing have different abilities and interests. Kurup et al. [15] proposed a new task recommendation model for reward-based crowdsourcing applications, derived task-based and reward-based probabilities, and analyzed their effectiveness. In general, crowdsourcing recommendation systems are designed for a single stakeholder; Aldhahri et al. [13] proposed a multiobjective recommendation model for the interests of crowdsourcing workers, requesters, and platforms, but the model suffers from the cold start problem. Wang et al. [16] used the negative exponential learning curve model to fit workers’ skill upgrades, and eliminated the impact of task difficulty on worker scores through the task difficulty weighting algorithm, and proposed a recommendation framework based on the model. Considering the independent contributions of crowdsourcing developers, Ye et al. [17] proposed a teammate recommendation method for developers based on their expertise and preferences. In the application of software development, Yu et al. [18] considered the dynamic changes in the development capabilities of software crowdsourcing developers, and combined with task types and complexity, proposed a dynamic task assignment algorithm DUTA. Habib et al. [19] considered a developer’s preference mode, technical popularity and winning probability in software crowdsourcing development in a competitive mode, and proposed a dynamic task recommendation system that combines task exploration and development.

3  System Model

Based on traditional crowdsourcing research, this paper proposes a CRBI frame-work that analyzes workers’ influence characteristics and combines workers’ task performance models and potential task models of new tasks.

As shown in Fig. 1, the CRBI framework is divided into three main steps: Calculation of worker influence, recommendation of workers for new tasks, and aggregation of results generated by recommended workers. The detailed process is as follows.

images

Figure 1: Design flow of the CRBI framework

Step 1: Calculation of workers’ influence. First, the historical task information on the crowdsourcing platform is used to construct the initial influence of the workers and the accuracy of the workers in completing tasks. Simultaneously, a weight correction model is used to perform weight training on the initial influence to obtain the corrected worker influence, which is used as the worker weight to evaluate the overall performance of the worker.

Step 2: Worker recommendation for new tasks. Based on the traditional open-call task selection mode, worker influence and task complexity are combined to construct worker performance models for different tasks and to perceive and build characteristic models for new tasks. The recommended set of workers is obtained by calculating the similarity between workers’ performance characteristics and new task characteristics. In addition, the task distribution quantity is controlled by fitting the change curve of task entropy to reduce the task cost of the requester, which is an optimization strategy of the CRBI framework.

Step 3: Aggregation of crowdsourcing task results. A weighted majority vote based on workers’ influence (WIVT) is applied to the task answers generated by recommended workers to complete the result aggregation. Finally, the task results are returned to the crowdsourcing platform to complete the crowdsourcing process.

4  CRBI Framework

4.1 Definition of Symbols

This paper considers a crowdsourcing problem of n tasks and m workers, where workers choose a label for the task from z categories. The specific symbols used in this paper are defined in Tab. 1.

Table 1: Symbol definitions

images

4.2 Calculation of Workers’ Influence

In contrast to the traditional workers’ ability evaluation index based on the accuracy of task completion, this paper proposes a model of worker influence characteristics based on the ablation method. This model calculates the influence of aggregated workers’ task results to obtain a measure of worker influence that serves as a new assessment indicator for the comprehensive ability of workers. The main process is shown in Fig. 2.

images

Figure 2: Workers’ influence calculation model (The left is the process of the initial aggregation accuracy of the worker group; the right is the process of aggregation after excluding worker images. The difference in accuracy is the degree of worker disturbance to the result aggregation)

Step 1: Calculate the aggregation accuracy of workers in historical tasks. The initial aggregation accuracy of the worker group is obtained as shown in Fig. 2 (left).

images

where images is an indicator function; when images, the function is 1.

Step 2: Eliminate the task data of worker images and recalculating the result aggregation accuracy again, as shown in Fig. 2 (right).

images

where images is the number of remaining tasks.

Step 3: Calculate the difference between images and images to obtain the degree of disturbance of the aggregation result by worker images.

images

where images is the disturbance degree of the worker. If images, the answer of images has a negative disturbance on the result; thus, the worker is classified as a low-quality worker, where images indicates its low-quality degree. If images, the answer of images has a positive disturbance on the result; thus, the worker is classified as a key worker, where images indicates the importance. Furthermore, we map images to [0, 1] according to the sigmod function, and the amplified influence is defined as follows:

images

where images is a scaling factor. By means of images, the degree of disturbance of workers in historical tasks can be evaluated and used as the initial influence of workers.

Step 4: Construct a weight correction model to adjust the initial influence of workers based on the accuracy of workers’ task completion.

images

where images, images are correction parameters, images is a reward factor for the weight of key workers, and images is a penalty factor for the weight of low-quality workers to finely distinguish the differences between workers.

Step 5: Calculating the worker’s influence images as the weight of the worker and using the result to evaluate the comprehensive ability of the worker.

images

when images, the influence of worker images on the final aggregation result tends to 0; thus, this worker is regarded as a general worker, and the task accuracy rate is used as the weight. When images, worker images has a positive or negative influence on the final aggregation result, so the corrected influence images is taken as the weight.

4.3 Worker Recommendation Based on Open-Call Mode

In the real world, the crowdsourcing recommendation will have a cold start problem. To address this problem, this paper applies the traditional task selection mode and proposes a worker recommendation algorithm based on the open-call mode. The algorithm process is shown in Fig. 3. First, a worker’s weighted task performance model is constructed based on historical tasks; then, the characteristic model of the task applicant is analyzed to obtain the characteristic model of the new task. Finally, the similarity between the task model and the worker model is calculated to generate recommendations.

images

Figure 3: Flow of worker recommendation based on open-call mode

4.3.1 Construction of the Task Performance Model of Workers

Different types of tasks have different difficulties, even tasks of the same category have slight differences. This characteristic is often ignored when building worker models. Therefore, this paper builds a weighted task performance model based on task complexity to distinguish workers’ performance on different types of tasks.

Step 1: Calculate the entropy images of the worker answer for task images as the complexity of task images. That is, the greater the uncertainty in the worker’s answer to the task is, the higher the complexity of the task.

images

where images is the probability that the worker completed task images correctly.

Step 2: Use the task complexity to weight the accuracy images of worker images on each type of task to obtain the weighted accuracy images of the worker.

images

4.3.2 Generation of Recommended Workers

Due to the lack of prior knowledge and descriptive information in crowdsourcing tasks, it is difficult for crowdsourcing platforms to obtain some relevant characteristics of new tasks. However, we can obtain the relevant characteristics of the task by mining the subtle links between the new task and the applicants. In this paper, we combine the traditional open-call task selection model and perceive the potential characteristics of a new task based on the worker’s task performance model and influence.

Step 1: Integrate the workers applying for a new task and build the connection between the new task and the worker images. Then, use the workers’ influence images to perform a weighted average of their task performance models to obtain a new task characteristic model images.

images

where images represents the confidence level of the new task as images and images is the number of recommended workers.

Step 2: Calculate the similarity of the characteristic model images of the new task and the task performance characteristic model images of each worker to measure how well workers match the new task.

images

Map the similarity vector of the worker group images to the worker list to obtain the final recommendation list according to the number of workers set by the requester.

Step 3: Use an exponential curve to fit the trend of task entropy images to select the appropriate task recommendation number.

images

Let images to obtain the minimum number of tasks for which the accuracy of the task converges, and use this value to control the number of task distributions. We attempt to obtain the best accuracy with a small number of tasks to improve the overall performance of the CRBI framework.

4.3.3 Aggregation of Task Results

After recommending suitable workers for new tasks, the crowdsourcing platform must also collect workers’ task result tags. Aggregation rules must be used to infer the true results of the task from the results of the task with noise. The quality of the task results depends not only on the capabilities of the workers but also on the aggregation rules. Therefore, good aggregation rules are important.

In general, aggregation rules include algorithms based on majority voting and statistical methods. However, statistics-based aggregation entails considerable computational cost, while majority-based aggregation assigns different quality workers the same weight and produces inefficient results. Therefore, we propose a weighted voting aggregation algorithm based on workers’ influence (WIVT) that uses worker influence as a weight for weighted voting to infer the trusted label of the task. The formula is as follows:

images

where images is the label category of task images inferred by the algorithm and images represents the probability that task images belongs to category images. The formula is as follows:

images

In contrast to the traditional weighted majority voting algorithm, the WIVT algorithm takes into account workers’ task completion accuracy in addition to workers’ task amount, task complexity and other characteristics. The characteristics of workers covered by WIVT are more comprehensive. Moreover, the weight calculation process, weight optimization is performed to maximize the accuracy of the aggregation. The WIVT algorithm can effectively alleviate problems such as voting ties and malicious voting while ensuring the accuracy of the aggregation.

5  Experiment

5.1 Dataset

The dataset involved in the experiment is shown in Tab. 2. Each dataset is collected from the AMT platform, and the main attributes are worker ID, task ID, worker answer, and golden label (that is, the correct label provided by an expert).

Table 2: Dataset

images

The MG dataset requires workers to listen to a 30–second song sample and classify it into one of 10 music types. The SP–2013 and SP–2015 datasets require workers to classify the polarity of sentences in movie reviews as positive or negative. The ZC–India and ZC–USA datasets contain links between entity names extracted from news articles and URIs describing the entities, requiring each worker to distinguish whether the URI is related or unrelated to the entity.

The experimental program is written mainly in Python. The execution environment is a Windows 10 operating system, and the CPU is an Intel i7 quad-core processor.

In the experiment, accuracy and RMSE are used as indicators to evaluate the recommendation results. Accuracy is used to measure the accuracy of recommended workers in completing a task, and RMSE is used to evaluate the accuracy with respect to the average cumulative residual. Avg-Precision, Avg-Recall, and F1-Measure are used as indicators to evaluate the task aggregation effectiveness of the recommended workers: Avg-Precision indicates the average accuracy of task result aggregation; Avg-Recall indicates the probability of a worker completing a certain type of task correctly; and F1-Measure is a metric that comprehensively considers accuracy and recall.

5.2 Experiment and Results Analysis

To illustrate the scientificity and effectiveness of the CRBI framework, the experiment aims to solve the following four problems:

Q1: How are workers’ influence indicators structured? What is the effect?

Q2: How effective is the CRBI recommendation?

Q3: How is cost optimization implemented in the CRBI?

Q4: How to integrate the task answers submitted by recommended workers?

5.2.1 Construction and Evaluation of Workers’ Influence

In this experiment, we first identified the existing problems by assessing the accuracy of the workers, then calculated the disturbance and selected the parameters to obtain workers’ influence, and finally evaluated and analyzed the workers’ influence indicators. We use 50% of the dataset as known historical data to construct and evaluate the worker influence characteristics. We obtain the distribution map of the completion accuracy of the workers on historical tasks via statistical analysis of the historical tasks of the workers (Fig. 4).

images

Figure 4: Characteristic distribution of workers in the MG dataset. (the line chart is the change in the accuracy of each worker; the histogram is the total number of workers’ historical tasks, where the number of incorrect tasks is shown in red and the number of correct is shown in blue)

Fig. 4 shows that using the accuracy of workers as their weight leads to an excessively broad index, making it difficult for the platform to distinguish the behavioral differences between workers with the same accuracy. For example, in the Fig. 4, the worker with ID 36 completed 12 tasks correctly, while the worker with ID 4 completed 1 task correctly: the accuracy of both workers is 100%. Furthermore, the worker with ID 16 is correct in 61 of 76 tasks completed, while the worker with ID 32 is correct in 27 of 34 tasks, and both have an accuracy of approximately 80%. Clearly, the different orkloads of the workers lead to differences in the accuracy assessment; therefore, accuracy is not an effective indicator of workers’ ability. Moreover, workers’ accuracy can be higher or lower depending on the difficulty of the task type (Fig. 5).

images

Figure 5: Average accuracy of each dataset

Therefore, in response to the above problems, we propose an indicator of worker influence to evaluate the comprehensive ability of workers. First, the disturbance of workers to the task aggregation result is calculated, as shown in Fig. 6, where the vertical axis is the degree of worker disturbance. When the value is positive, the worker has a promoting effect on the task; when the value is negative, the worker’s answer has a negative effect on the task. When the value is 0, the worker does not play a key role in the task aggregation, that is, an ordinary worker. However, ordinary workers are the most common, which satisfies with the normal distribution of the population.

images

Figure 6: Distribution of worker disturbances

After the worker’s disturbance value is calculated, the reward factor α and the penalty factor β are selected to perform weight correction. According to previous experiments and mathematical statistical analysis, we set the selection interval of the worker’s reward factor α to [1, 3] and the selection interval of the penalty factor β to [0, 2.5]. Then, the values in the interval are combined in pairs, and the aggregation accuracy of each combination is calculated to select the optimal parameter.

Fig. 7 shows the aggregation accuracy under different reward factors α and penalty factors β. The horizontal axis is the worker’s reward factor α, the vertical axis is the worker’s penalty factor β, and the value in the matrix is the aggregation accuracy of the task results. Then, these values are normalized. In Fig. 7 (left), the coordinates corresponding to a value of 1 are the optimal values of parameters α and β. As shown in Fig. 7 (right), if there are multiple coordinates with a value of 1, the historical data are randomly divided and aggregated multiple times, and parameters are selected from the intersection set of the matrices.

images

Figure 7: Aggregation accuracy under different parameters (the accuracy on the MS dataset on the left and the accuracy on the SP-2015 dataset on the right)

Fig. 8 shows that the indicator of worker influence objectively reflects the difference in ability among workers. For example, even though the worker with ID 36 has completed only 12 tasks, he receives a high weight because of the 100% accuracy. The worker with ID 16 has the same accuracy as the worker with ID 32, but the worker with ID 16 has a higher weight because of the greater number of completed tasks. Therefore, this indicator can distinguish between “new” and “experienced” workers, and there are few cases where the weights are the same, which aids in distinguishing among workers.

images

Figure 8: Distribution of workers’ influence

In addition, some workers have the same accuracy and are considered to have equal chances of being recommended. Moreover, dishonest workers may improve their accuracy through cheating. As shown in Fig. 9, when the accuracy rates of the workers are all 1.0, the accuracy rate index cannot be used to effectively distinguish the excellence degree of the workers. However, the influence index can be used to distinguish between homogeneous workers and reduce the weight of some low-quality workers.

images

Figure 9: Changes in worker accuracy and influence on the ZC-USA dataset

5.2.2 Performance Evaluation of Recommendation Algorithms

To evaluate the recommendation performance of the CRBI framework, this experiment compares our recommendation algorithm with the following four recommendation methods: RBAcc, recommendation based on workers’ task accuracy; RBCf, recommendation based on collaborative filtering; RBInf, recommendation based on workers’ influence; Random, no recommendation is used (workers are randomly selected).

As shown in Fig. 10, under different numbers of recommended workers, CRBI performs better than the Random, RBAcc, RBCF, and RBFf algorithms by approximately 29%, 15%, 10%, and 11%, respectively, and the curve of the algorithm is more stable. Moreover, CRBI achieves better accuracy with a small number of workers.

images

Figure 10: The accuracy of the recommended workers in completing the task with different numbers of recommended workers

However, since we use existing data for the recommendation simulation, as the number of recommenders increases, the number of available workers gradually decreases, and the set of recommended workers will tend to be the same. Therefore, the recommendation accuracy will gradually decrease. However, even in this case, our algorithm maintains an accuracy improvement of 10%~29% compared to the other algorithms.

In addition, to effectively evaluate the performance advantage of the recommendation algorithm in CRBI, we choose N = 3 (that is, the number of recommended workers for each task is 3). The task completion accuracy and the RMSE of the recommended worker are shown in Fig. 11.

images

Figure 11: Accuracy and RMSE of the recommendation algorithm

As shown in Fig. 11 (left), compared to the Random algorithm without recommendation, our algorithm achieves an accuracy improvement of approximately 29%~39% by considering the workers’ influence and task performance characteristics. Compared to the other recommendation algorithms, CRBI achieves different degrees of accuracy improvement. Moreover, Fig. 11 (right) shows that as the number of tasks increases, the accuracy and RMSE of the compared methods tend to converge. By contrast, the accuracy of our recommendation algorithm steadily improves throughout the entire process, and the RMSE is always low, indicating good performance.

Tab. 3 presents the recommendation results when the number of tasks is 20, 40, and 60. The accuracy of the recommendation algorithm based on the open-call mode in the CRBI framework is the highest and the MSE and RMSE are the lowest, indicating that our recommendation algorithm has good performance. In addition, the RBInf algorithm performs better than the Random algorithm, indicating that the index of worker influence we constructed is beneficial in evaluation worker performance.

Table 3: Evaluation of various recommendation algorithms under different task numbers

images

In summary, the recommendation algorithm in the CRBI framework can greatly improve the accuracy of task completion, and it performs better than recommendation based on traditional ideas and characteristics.

5.2.3 Cost Optimization in Recommendation

In the general task distribution or recommendation process, the same number of workers will be recommended to complete each type of task. However, in the actual crowdsourcing process, if the number of tasks is insufficient, the accuracy of task completion will be low, whereas if too many tasks are distributed, the cost will increase cost and the accuracy of the tasks will be difficult to improve. Therefore, choosing an appropriate number of distributions for different tasks is key to optimizing the cost optimization.

In this experiment, we use three methods to fit the relationship between the number of tasks and task entropy to select the optimal task distribution volume for the recommendation algorithm to optimize the cost.

Fig. 12 shows that the exponential approximation matches the trend of task entropy, so we select the exponential curve. Due to the difficulty of different types of tasks and the number of tasks required to reach convergence, we use exponential curves to fit other types of tasks (Fig. 13).

images

Figure 12: Fit between task number and task entropy

images

Figure 13: Entropy fitting curves

As shown in Fig. 13, as the number of tasks increases, the task results tend to stabilize. When the curve approaches 0, the tasks tends to converge, and the value of the abscissa at this time is the minimum number of tasks required for this type of task. When the task entropy converges, the accuracy of the task results cannot be improved even if the task continues to be distributed. These experiments illustrated that if the task distribution is stopped when the task entropy converges, the cost can be reduced by 20~30%. Therefore, the task number selection method based on task entropy in the CRBI framework can effectively optimize the cost of crowdsourcing platforms.

5.2.4 Aggregation of the Task Results of Recommended Workers

Crowdsourcing can produce efficient output but also substantial noise, which leads to unreliable quality of the results. Therefore, the true value must be inferred from unreliable results in a short time. For the task answers submitted by workers, we use the WIVT in the CRBI framework to aggregate the results and compare them with the results of four other aggregation algorithms: Majority Vote (MV), Vote Distribution (VD) [20], Dawid & Skene (DS), and Bayesian (BCC) [21].

As shown in Fig. 14, the variation curve of the WIVT algorithm is always at a relatively high position, and the curve growth is stable, indicating that WIVT has the best stability. Furthermore, Tab. 4 indicates that at 50% of the data volume, the WIVT algorithm achieves a 9% performance improvement over the majority voting algorithm and a 6%, 4%, and 16% performance improvement over other aggregation algorithms, respectively. At 100% data volume, the WIVT algorithm achieves a performance improvement of approximately 10%, 8%, 1%, and 3% compared to the voting algorithm and the other aggregation algorithms.

images

Figure 14: Aggregated results of various algorithms (accuracy line chart on the left and histogram of F1-measures on the right)

Table 4: Aggregation results of various algorithms

images

The MV and VD algorithms assume that the task finishers are high-quality workers and do not distinguish between the voting weights of workers, which leads to inefficient aggregation results. The other two statistics-based methods, despite having good performance, have a substantial time cost. Because the workers are filtered by the recommendation algorithm, the recommended workers not only have better comprehensive ability but also a good matching degree with the new task, producing a good aggregation effect. In addition, the WIVT algorithm has a considerable advantage with respect to time: a 50 to 100 times improvement in time compared to the statistical method. In addition to ensuring time efficiency, the WIVT algorithm produces high-quality task results, which can maximize the application performance of the CRBI framework.

6  Conclusion

Crowdsourcing is a problem-solving mechanism with broad application prospects that can apply crowd intelligence to large-scale and complex tasks. Crowdsourcing recommendation is an important means to improve the quality of task completion. However, some traditional recommendation methods have certain shortcomings and ignore the control of task costs and the processing of task results generated by recommended workers. Therefore, we propose a crowdsourcing recommendation framework based on worker influence to systematically implement the entire crowdsourcing process. Among them, the worker’s influence takes into account the worker’s correct rate, error rate, and task complexity, which can more accurately assess the worker’s characteristic performance. Moreover, the weighted voting based on workers’ influence can also ensure the accuracy and safety of task aggregation. The CRBI framework proposed in this paper is used for the entire crowdsourcing recommendation process. From a user perspective, CRBI implements the entire process from task distribution to task answer acquisition; from a platform perspective, CRBI implements various processes, including worker behavior analysis, characteristic mining, worker recommendation, and task result aggregation. However, because this paper uses the classification tasks of the AMT platform for experiments, further experiments are needed for the suitability of other types of tasks. Moreover, the protection of workers’ privacy and incentive mechanisms in the recommendation also need to be further improved.

In the future, we will consider using technologies such as big data to perform parallel calculations on massive crowdsourced data to improve the efficiency of the model. And use blockchain technology to optimize the crowdsourcing voting algorithm to ensure the security of the crowdsourcing voting process. In addition, we plan to conduct research on other types of crowdsourcing tasks, such as crowdsourcing software development, to design a more applicable recommendation framework.

Funding Statement: The works that are described in this paper are supported by Ministry of Science and Technology: Key Research and Development Project (2018YFB003800), Hunan Provincial Key Laboratory of Finance & Economics Big Data Science and Technology (Hunan University of Finance and Economics) 2017TP1025 and HNNSF 2018JJ2535.

Conflicts of Interest: The authors declare that they have no conflicts of interest to report regarding the present study.

References

1.   P. Kucherbaev, F. Daniel, S. Tranquillini and M. Marchese. (2015). “Crowdsourcing processes: A survey of approaches and Opportunities,” IEEE Internet Computing, vol. 20, no. 2, pp. 50–56. [Google Scholar]

2.   J. Howe. (2006). “The rise of crowdsourcing,” Wired Magazine, vol. 14, no. 6, pp. 1–4. [Google Scholar]

3.   Z. Guo, C. Tang, W. Niu, Y. Fu, T. Wu et al.. (2017). , “Fine-grained recommendation mechanism to curb astroturfing in crowdsourcing systems,” IEEE Access, vol. 5, pp. 15529–15541. [Google Scholar]

4.   D. E. Difallah, G. Demartini and P. Cudré–Mauroux. (2013). “Pick-a-crowd: Tell me what you like, and I will tell you what to do,” in Proc. of the 22nd Int. Conf. on World Wide Web, New York, NY, USA, pp. 367–374. [Google Scholar]

5.   Q. Zhong, Y. Zhang, C. Li and Y. Li. (2017). “Task recommendation method based on workers’ interest and competency for crowdsourcing,” System Engineering Theory and Practice, vol. 37, pp. 3270–3280. [Google Scholar]

6.   S. Lazarova-Molnar and N. Mohamed. (2019). “Collaborative data analytics for smart buildings: Opportunities and models,” Cluster Computing, vol. 22, no. 1, pp. 1065–1077. [Google Scholar]

7.   M. Hosseini, C. M. Angelopoulos, W. K. Chai and S. Kunding. (2019). “Crowdcloud: A crowdsourced system for cloud infrastructure,” Cluster Computing, vol. 22, no. 2, pp. 455–470. [Google Scholar]

8.   T. Wang, T. Wu, A. H. Ashrafzadeh and J. He. (2018). “Crowdsourcing-based framework for teaching quality evaluation and feedback using linguistic 2-tuple,” Computers, Materials & Continua, vol. 57, no. 1, pp. 81–96. [Google Scholar]

9.   H. Hu, Y. Zheng, Z. Bao, G. Li, J. Feng et al.. (2016). , “Crowdsourced poi labelling: Location-aware result inference and task assignment,” in IEEE 32nd Int. Conf. on Data Engineering, Helsinki, Finland, pp. 61–72. [Google Scholar]

10. M. C. Yuen, I. King and K. S. Leung. (2015). “TaskRec: A task recommendation framework in crowdsourcing systems,” Neural Processing Letters, vol. 41, no. 2, pp. 223–238. [Google Scholar]

11. K. Mao, Y. Yang, Q. Wang, Y. Jia and M. Harman. (2015). “Developer recommendation for crowdsourced software development tasks,” in IEEE Sym. on Service-Oriented System Engineering, San Francisco Bay, CA, USA, pp. 347–356. [Google Scholar]

12. M. Nilashi, O. Ibrahim and K. Bagherifard. (2018). “A recommender system based on collaborative filtering using ontology and dimensionality reduction techniques,” Expert Systems with Applications, vol. 92, pp. 507–520. [Google Scholar]

13. E. Aldahari, V. Shandilya and S. Shiva. (2018). “Crowdsourcing multi-objective recommendation system,” in Companion Proc. of the Web Conf. Republic and Canton of Geneva Switzerland, pp. 1371–1379. [Google Scholar]

14. M. Safran and D. Che. (2018). “Efficient learning-based recommendation algorithms for top-N tasks and top-N workers in large-scale crowdsourcing systems,” ACM Transactions on Information Systems, vol. 37, no. 1, pp. 1–46. [Google Scholar]

15. A. R. Kurup and G. P. Sajeev. (2017). “Task recommendation in reward-based crowdsourcing systems,” in Int. Conf. on Advances in Computing, Communications and Informatics, Udupi, India, pp. 1511–1518. [Google Scholar]

16. Z. Wang, H. Sun, Y. Fu and L. Ye. (2017). “Recommending crowdsourced software developers in consideration of skill improvement,” in 32nd IEEE/ACM Int. Conf. on Automated Software Engineering, Urbana, IL, USA, pp. 717–722. [Google Scholar]

17. L. Ye, H. Sun, X. Wang and J. Wang. (2018). “Personalized teammate recommendation for crowdsourced software developers,” in 33rd IEEE/ACM Int. Conf. on Automated Software Engineering, Montpellier, France, pp. 808–813. [Google Scholar]

18. D. Yu, Y. Wang and Z. Zhou. (2019). “Software crowdsourcing task allocation algorithm based on dynamic utility,” IEEE Access, vol. 7, pp. 33094–33106. [Google Scholar]

19. A. Habib, S. Hussain, A. A. Khan, M. K. Sohail, M. Ilahi et al.. (2019). , “Knowledge based quality analysis of crowdsourced software development platforms,” Computational and Mathematical Organization Theory, vol. 25, no. 3, pp. 122–131. [Google Scholar]

20. W. Wang and Z. H. Zhou. (2015). “Crowdsourcing label quality: A theoretical analysis,” Science China Information Sciences, vol. 58, no. 11, pp. 1–12. [Google Scholar]

21. H. C. Kim and Z. Ghahramani. (2012). “Bayesian classifier combination,” in Artificial Intelligence and Statistics, La Palma, Canary Islands, pp. 619–627. [Google Scholar]

images This work is licensed under a Creative Commons Attribution 4.0 International License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.