skip to main content
research-article

AutoMan: a platform for integrating human-based and digital computation

Published:19 October 2012Publication History
Skip Abstract Section

Abstract

Humans can perform many tasks with ease that remain difficult or impossible for computers. Crowdsourcing platforms like Amazon's Mechanical Turk make it possible to harness human-based computational power at an unprecedented scale. However, their utility as a general-purpose computational platform remains limited. The lack of complete automation makes it difficult to orchestrate complex or interrelated tasks. Scheduling more human workers to reduce latency costs real money, and jobs must be monitored and rescheduled when workers fail to complete their tasks. Furthermore, it is often difficult to predict the length of time and payment that should be budgeted for a given task. Finally, the results of human-based computations are not necessarily reliable, both because human skills and accuracy vary widely, and because workers have a financial incentive to minimize their effort.

This paper introduces AutoMan, the first fully automatic crowdprogramming system. AutoMan integrates human-based computations into a standard programming language as ordinary function calls, which can be intermixed freely with traditional functions. This abstraction lets AutoMan programmers focus on their programming logic. An AutoMan program specifies a confidence level for the overall computation and a budget. The AutoMan runtime system then transparently manages all details necessary for scheduling, pricing, and quality control. AutoMan automatically schedules human tasks for each computation until it achieves the desired confidence level; monitors, reprices, and restarts human tasks as necessary; and maximizes parallelism across human workers while staying under budget.

References

  1. S. Ahmad, A. Battle, Z. Malkani, and S. Kamvar. The Jabberwocky Programming Environment for Structured Social Computing. In UIST, pp. 53--64, 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Amazon. Mechanical Turk. http://www.mturk.com.Google ScholarGoogle Scholar
  3. C. N. Anagnostopoulos. MediaLab LPR Database. http://www.medialab.ntua.gr/research/LPRdatabase.html.Google ScholarGoogle Scholar
  4. M. S. Bernstein, G. Little, R. C. Miller, B. Hartmann, M. S. Ackerman, D. R. Karger, D. Crowell, and K. Panovich. Soylent: A Word Processor with a Crowd Inside. In UIST, pp. 313--322, 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. H. Chafi, Z. DeVito, A. Moors, T. Rompf, A. K. Sujeeth, P. Han-rahan, M. Odersky, and K. Olukotun. Language Virtualization for Heterogeneous Parallel Computing. In Onward!, 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. S. Cooper, F. Khatib, A. Treuille, J. Barbero, J. Lee, M. Beenen, A. Leaver-Fay, D. Baker, Z. Popovi, and F. Players. Predicting Protein Structures With a Multiplayer Online Game. Nature, 466(7307):756--760, 2010.Google ScholarGoogle ScholarCross RefCross Ref
  7. A. DasGupta. Probability for Statistics and Machine Learning: Fundamentals and Advanced Topics. Springer, 1st edition, 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. J. Dean and S. Ghemawat. MapReduce: Simplified Data Processing on Large Clusters. In OSDI, pp. 137--150, 2004. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. J. R. Douceur. The Sybil Attack. In IPTPS, pp. 251--260, 2001. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. S. Dow, A. Kulkarni, B. Bunge, T. Nguyen, S. Klemmer, and B. Hartmann. Shepherding the Crowd: Managing and Providing Feedback to Crowd Workers. In CHI, pp. 1669--1674, 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. S. Due, M. Ibrahim, M. Shehata, and W. Badawy. Automatic License Plate Recognition (ALPR): A State of the Art Re-view. IEEE Transactions on Circuits and Systems for Video Technology, 2012.Google ScholarGoogle Scholar
  12. W. Feller. An Introduction to Probability Theory and Applica-tions, volume 1. John Wiley & Sons Publishers, 3rd edition, 1968.Google ScholarGoogle Scholar
  13. M. Fogus. BAYSICK---a DSL for Scala implementing a subset of BASIC. https://github.com/fogus/baysick, March 2009.Google ScholarGoogle Scholar
  14. M. J. Franklin, D. Kossmann, T. Kraska, S. Ramesh, and R. Xin. CrowdDB: Answering Queries with Crowdsourcing. In SIGMOD, pp. 61--72, 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. M. Gentile. UI widget images come from the "Vector UI Kit" package. http://www.icondeposit.com/design:9, Mar 2012. Creative Commons c 2012 Matt Gentile.Google ScholarGoogle Scholar
  16. J. Howe. The Rise of Crowdsourcing. Wired Magazine, 14(6):176--178, 2006.Google ScholarGoogle Scholar
  17. P. G. Ipeirotis. Demographics of Mechanical Turk. Tech. Rep. Working Paper CeDER-10-01, NYU Center for Digital Econ-omy Research, 2010.Google ScholarGoogle Scholar
  18. P. G. Ipeirotis, F. Provost, and J. Wang. Quality Management on Amazon Mechanical Turk. In HCOMP, pp. 64--67, 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. A. Kittur, B. Smus, and R. E. Kraut. CrowdForge: Crowdsourcing Complex Work. Tech. Rep. CMU-HCII-11-100, Human-Computer Interaction Institute, School of Computer Science, Carnegie Mellon University, February 2011.Google ScholarGoogle Scholar
  20. A. P. Kulkarni, M. Can, and B. Hartmann. Turkomatic: Auto-matic Recursive Task and Workflow Design for. Mechanical Turk. In CHI, pp. 2053--2058, 2011. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. K. Land, A. Slosar, C. Lintott, D. Andreescu, S. Bamford, P. Murray, R. Nichol, M. J. Raddick, K. Schawinski, A. Szalay, D. Thomas, and J. Vandenberg. Galaxy Zoo: the Large-Scale Spin Statistics of Spiral Galaxies in the Sloan Digital Sky Survey. Monthly Notices of the Royal Astronomical Society, 388:1686--1692, Aug. 2008.Google ScholarGoogle ScholarCross RefCross Ref
  22. G. Little, L. B. Chilton, M. Goldman, and R. C. Miller. TurKit: Human Computation Algorithms onMechanical Turk. In UIST, pp. 57--66, 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. M. Marge, S. Banerjee, and A. Rudnicky. Using the Amazon Mechanical Turk for Transcription of Spoken Language. In ICASSP, pp. 5270--5273, Mar 2010.Google ScholarGoogle ScholarCross RefCross Ref
  24. A. McCallum, K. Schultz, and S. Singh. FACTORIE: Probabilistic Programming via Imperatively Defined Factor Graphs. In NIPS, 2009.Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. M. Odersky and M. Zenger. Scalable Component Abstractions. In OOPSLA, pp. 41--57, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. D. Oleson, V. Hester, A. Sorokin, G. Laughlin, J. Le, and L. Biewald. Programmatic Gold: Targeted and Scalable Quality Assurance in Crowdsourcing. In HCOMP, pp. 43--48, 2011.Google ScholarGoogle Scholar
  27. D. Parikh and L. Zitnick. Human-Debugging of Machines. In NIPS CSS, 2011.Google ScholarGoogle Scholar
  28. D. Shahaf and E. Amir. Towards a Theory of AI Completeness. In Commonsense, 2007.Google ScholarGoogle Scholar
  29. D. Tamir, P. Kanth, and P. Ipeirotis. Mechanical Turk: Now With 40.92% Spam., Dec 2010.Google ScholarGoogle Scholar
  30. L. von Ahn, B. Maurer, C. Mcmillen, D. Abraham, and M. Blum. reCAPTCHA: Human-Based Character Recognition via Web Security Measures. Science, 321(5895):1465--1468, Aug 2008.Google ScholarGoogle ScholarCross RefCross Ref
  31. T. Yan, V. Kumar, and D. Ganesan. CrowdSearch: Exploiting Crowds for Accurate Real-Time Image Search on Mobile Phones. In MobiSys, pp. 77--90, 2010. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. AutoMan: a platform for integrating human-based and digital computation

              Recommendations

              Comments

              Login options

              Check if you have access through your login credentials or your institution to get full access on this article.

              Sign in

              Full Access

              • Published in

                cover image ACM SIGPLAN Notices
                ACM SIGPLAN Notices  Volume 47, Issue 10
                OOPSLA '12
                October 2012
                1011 pages
                ISSN:0362-1340
                EISSN:1558-1160
                DOI:10.1145/2398857
                Issue’s Table of Contents
                • cover image ACM Conferences
                  OOPSLA '12: Proceedings of the ACM international conference on Object oriented programming systems languages and applications
                  October 2012
                  1052 pages
                  ISBN:9781450315616
                  DOI:10.1145/2384616

                Copyright © 2012 ACM

                Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

                Publisher

                Association for Computing Machinery

                New York, NY, United States

                Publication History

                • Published: 19 October 2012

                Check for updates

                Qualifiers

                • research-article

              PDF Format

              View or Download as a PDF file.

              PDF

              eReader

              View online with eReader.

              eReader