Beruflich Dokumente
Kultur Dokumente
The Task
We asked the crowd to validate company websites. We provided them with a Company Name and a Company Description, which was actually the description of the SIC industry classification that had been assigned to the business, and we asked the crowd workers to confirm whether the link we gave them was the correct website for the company. We asked the contributors to select one of the following answers: 1 Yes, the Name and SIC match the site 2 No, neither the Name nor the SIC match the site 3 Maybe, the Name matches but the SIC does not 4 Bad link As you can see from our answer choices, we were actually trying to get more than one answer from the workers. In addition to providing the SIC code as a means to confirm that the website belonged to the business (matching not only the Company Name but also the Company Description to the website), we also hoped to validate the industry classification previously assigned to the business (answer choice Maybe).
TIP #1
Avoid overloading questions instead, break complex questions into multiple steps.
Our instructions explained how to find the Company Name on the website and where workers might easily locate a description of services that could be used to compare the Company Description. We distributed the work to U.S.-based crowdworkers and paid $0.04 per unit. The entire file included 391 records with 13 test questions (or gold records) that were used to gauge the trust level of the workers on this task. Contributors were paid for every judgment (answer) provided and, apart from the test questions which were included in each task and used to determine whether the worker could continue working on the job, the accuracy of the work was not checked until the entire file had been completed.
Findings TIP #2 If there is a predominant correct answer to your question, select test questions where the answer is not the easiest guess.
Overall, the quality of the work we received was not great however, this did not surprise us, as we knew that there were flaws in the design of our task. We will discuss the mistakes we made and our takeaways in the next section. First, we share our analysis of the results file: ! ! The same universe of workers completed both the golds and full file records, and had an average trust level of 89.9% The quality-checked sample from the full file was not randomly selected. We included more units where the correct answer was not likely to be Yes in order to gauge the quality of judgments where the correct answer was No or Maybe, as these judgments had a higher error rate in the gold record results. Based on the golds and the sample checked from the full file, the correct answer was Yes more than half of the time.
www.k2reimagine.com
! !
! ! !
TIP #3 While worker trust scores can be useful, the design of your job is what will determine the level of accuracy.
www.k2reimagine.com
TIP #4 Design your tasks bearing in mind that the crowd does not know the context of the task (or your business).
Designing crowd-supported workflows and microtasks successfully requires not only an appreciation for the nuance of the processes being restructured but also a familiarity with the behavior of the crowd. Just as a manager building a new team in a new location would consider the economics and the culture distinguishing their newcomers, the same attention must be given to this on-demand, virtual workforce. If there is one golden rule for working with the crowd, it is to avoid the common trap of simply retooling and redeploying existing processes. Rather, crowdsourcing for data management thrives when the work is reimagined, keeping in mind the capacity as well as the constraints, and equally important, the motivations of the crowd. We have discussed the capabilities that can be unleashed from the crowd; now we will highlight some of the constraints and motivations that were exemplified in this exercise.
Most data managers commit significant time and resource to ensuring that their production analysts understand the business demands driving their work. The best analysts and data specialists connect the dots and are able to see the big picture. A good analyst would have looked at our task in this case and would have understood that we could optimize the task of confirming the company website to include confirming the industry classification that had been previously assigned (by noticing while on the website the business in which the company is engaged). Workers in the crowd, however, do not have the same context. The task they have been assigned has been necessarily sliced down to just a fractional view of the overall scope of work. It is for this reason that microtask designers must frame questions in a way that does not assume any background knowledge or perspective on the part of the crowdworkers that would allow them to add value above and beyond what the task explicitly requests. In-house analysts, and even outsourced teams, are typically motivated by quality and reputation, key elements for continued employment and career growth. The crowd is motivated by shorter-term incentives, typically money. Whereas quality and reputation apply to the Crowd as well, maximizing compensation is
2013 K2 Case Studies www.k2reimagine.com 3
TIP #5 As with any workforce, understand the behaviors and motivations of the crowd.
www.k2reimagine.com