crowdsourcing and human computation
play

Crowdsourcing and Human Computation Tuesdays and Thursdays - PowerPoint PPT Presentation

Crowdsourcing and Human Computation Tuesdays and Thursdays 3pm-4:30pm 3401 Walnut St room 401B Instructor: Chris Callison-Burch Website: crowdsourcing-class.org Inter-related concepts Groups of individuals doing things collectively


  1. Crowdsourcing and Human Computation Tuesdays and Thursdays 3pm-4:30pm 3401 Walnut St room 401B Instructor: Chris Callison-Burch Website: crowdsourcing-class.org

  2. Inter-related concepts “Groups of individuals doing things collectively Collective that seem intelligent” “A paradigm for utilizing Intelligence human processing power Human to solve problems that “A labor market The Gig com puters cannot yet characterized by Computation Economy solve.” the prevalence of short-term contracts or Crowd- freelance work as sourcing “Outsourcing a job traditionally opposed to performed by an employee to an permanent jobs.” undefined, generally large group of people via open call.” Data Mining “Applying algorithms to extract patterns from data.”

  3. Francis Galton

  4. Collective Intelligence? Group Think

  5. Collective Intelligence? Mob Mentality

  6. Collective Intelligence? Fake News

  7. Collective Intelligence? Misinformation campaigns

  8. Popular Delusions and the Madness of Crowds • Economic bubbles • Alchemy & Psuedoscience • Witch hunts • Prophecies

  9. Tulip Mania “A tulip, known as "the Viceroy" displayed in a 1637 Dutch catalog. Its bulb was offered for sale between 3,000 and 4,200 guilders depending on size. A skilled craftsworker at the time earned about 300 guilders a year.”

  10. "Looking back, it’s clear that the Beanie Baby craze was an economic bubble, fueled by frenzied speculation and blatantly baseless optimism. Bubbles are quite common, but bubbles over toys are not."

  11. Wisdom of Crowds Requirements for a crowds to be wise • Diversity of Opinion • Independence • De-centralization • Aggregation

  12. Groups / Crowds • Employees of a business • Participants in a poll • Sports fans betting on games • Independent stock market investors • Internet users linking to sites • Citizens in a democracy

  13. Ways of aggregating collective intelligence • Point spreads / parimutuel odds • Stock prices • Futures contracts • Voting • Computer algorithms, interfaces

  14. 2010 Haitian Earthquake

  15. Disaster Response The maps are bad Jan 12 Robert Munro

  16. Disaster Response Better maps from Crowdsourcing Jan 12 Jan 23 Robert Munro

  17. Disaster Response The responders don’t speak Kreyol • Fanmi mwen nan Kafou, 24 • My family in Carrefour, 24 Cote Cote Plage, 41A bezwen manje Plage,41A needs food and ak dlo water • Moun kwense nan Sakre Kè • People trapped in Sacred Heart nan Pòtoprens Church, PauP • Ti ekipman Lopital General • General Hospital has less than 24 hrs. supplies genyen yo paka minm fè 24 è • Fanm gen tranche pou fè yon • Undergoing children delivery piIt nan Delmas 31 Delmas 31 Robert Munro

  18. Disaster Response Maps + Translation + Local Knowledge Workers collaborated to find Apo locations: Dalila: I need Thomassin Apo please Dalila Apo: Kenscoff Route: Lat: 18.495746829274168, Long:-72.31849193572998 Haiti responders Apo: This Area after Petion-Ville and Pelerin 5 is not on Google Map. We have no streets name Apo: I know this place like my pocket Dalila: thank God u was here Feedback from responders: "just got emergency SMS, child delivery, USCG are acting, and, the GPS (18.4957, -72.3185) coordinates of the location we got from someone of your team were 100% accurate!" Robert Munro

  19. Disaster Response • Clark Craig of the Marine Corps: – “I cannot overemphasize to you what the work of the Ushahidi/HaiI has provided. It is saving lives every day.” • Secretary of State Hillary Clinton: – “The technology community has set up interacIve maps to help us idenIfy needs and target resources. And on Monday, a seven-year-old girl and two women were pulled from the rubble of a collapsed supermarket by an American search-and-rescue team aVer they sent a text message calling for help.” • Cr aig Fulgate, FEMA Task Force: – “[The] Crisis Map of HaiI represents the most comprehensive and up-to-date map available to the humanitarian community.” • Ushahidi@TuVs : – “The World Food Program delivered food to an informal camp of 2500 people, having yet to receive food or water, in Diquini to a locaIon that 4636 had idenIfied for them.” Robert Munro

  20. How can computer science and economics help facilitate collective intelligence?

  21. NASA Clickworkers (2000) NASA showed that public volunteers could do routine science analysis that would normally be done by a graduate student working for months on end. From November 2000 to January 2002, they had 101,000 clickworkers volunteering 14,000 work hours , 612,832 sessions , and 2,378,820 entries ! We try to have several people cover each region on Mars so that we can compute a consensus, throwing out any mistaken or frivolous entries and averaging out the inaccuracies. Here are all Here is the consensus the clicks we received for this region

  22. NASA Clickworkers (2000) Mars age map produced directly from clickworker inputs. Mars age map produced from scientists Color guide: red=heavily cratered (old), green=medium, violet=lightly cratered (young).

  23. Postmark City: Barre Postmark State: MA Postmark Date: Oct-11 Postmark Year: 1886 Stamp: 1c $ 0. 01

  24. Help African Refugees samasource.org

  25. Choose the right word What would you call these colors? Dolores Labs

  26. Catch some zzzzs thesheepmarket.com

  27. Dark Side of Crowdsourcing Real Time with Bill Maher: The "Sharing" Economy – August 21, 2015 (HBO)

  28. Are Workers Treated Fairly? �40

  29. https://crowd-workers.com/

  30. 3k workers 20k requesters 3.8m task records

  31. A Data-Driven Analysis of Workers’ Earnings on Amazon Mechanical Turk CHI-2018 Kotaro Hara Abigail Adams Kristy Milland Saiph Savage Chris Callison-Burch Jeffrey P. Bigham ABSTRACT A growing number of people are working as part of on-line crowd work. Crowd work is often thought to be low wage work. However, we know little about the wage distribution in practice and what causes low/high earnings in this setting. We recorded 2,676 workers performing 3.8 million tasks on Amazon Mechanical Turk. Our task-level analysis revealed that workers earned a median hourly wage of only ~$2/h, and only 4% earned more than $7.25/h. While the average requester pays more than $11/h, lower-paying requesters post much more work. Our wage calculations are influenced by how unpaid work is accounted for, e.g. , time spent searching for tasks, working on tasks that are rejected, and working on tasks that are ultimately not submitted. We further explore the characteristics of tasks and working patterns that yield higher hourly wages. Our analysis informs platform design and worker tools to create a more positive future for crowd work.

  32. Takeaways Crowd workers are underpaid < $2/h and they often earn below $2/h $ Unpaid work, particularly returning tasks has a large impact on the hourly wage Majority of the requesters reward workers below $5/h

  33. How to put crowdsourcing towards good uses �46

  34. Use of MTurk-like systems in research • Participant pool for user studies, polling, cognitive science experiments • Annotation for machine learning tasks like computer vision or NLP • Human Computer Interaction: worker pools are hardwired into the UI • New Programming Languages Concepts • Study markets themselves for economics research, cost-optimization

  35. Annotation for machine learning / artificial intelligence tasks Task: Dog? Answer: Yes Broker Pay: $0.01 Is this a dog? www.mturk.com o Yes o No $0.01

  36. Human Computer Interaction

  37. Crowdsourcing - Rad Lab Talk - UC Berkeley Fall 2010

  38. New Programming Languages Concepts

  39. New Programming Languages Concepts

  40. New Programming Languages Concepts • Latency • Cost • Parallelization • Non-determinism • Iterative improvement

  41. Study Markets Themselves • What predictions of economics hold true on MTurk? • What incentives can we give to increase throughput, quality, worker retention? • What is the cost-optimal solution to a problem?

  42. What will we cover in this class (and should you take it)?

  43. Topics • Taxonomy of crowdsourcing and human computation • Microtasking platforms like Mechanical Turk and Figure-eight • Programming concepts for human computation • The economics of crowdsourcing • Crowdsourcing and machine learning • Applications to human computer interaction • Crowdsourcing and social science

  44. Who should take this class • Anyone who wants to be on the cutting edge of this new field • Entrepreneurial students who want to start their own companies • Students from the business school who want to experiment with markets • Students from the social sciences who want to conduct large-scale students with people

  45. Course Requirements Weekly assignments Writing and Coding Company profile, Presentations project pitch Self-designed, Final project groups of 4-5 Final presentation Show off your work

Recommend


More recommend